[ceph-users] luminous - 12.2.1 - stale RBD locks after client crash

Jason Dillaman jdillama at redhat.com
Wed Nov 22 04:52:31 PST 2017


See previous threads about this subject [1][2] and see step 6 in the
upgrade notes [3].

[1] http://lists.ceph.com/pipermail/ceph-users-ceph.com/2017-September/020722.html
[2] https://www.mail-archive.com/ceph-users@lists.ceph.com/msg41718.html
[3] http://docs.ceph.com/docs/master/release-notes/#upgrade-from-jewel-or-kraken

On Wed, Nov 22, 2017 at 2:50 AM, Nikola Ciprich
<nikola.ciprich at linuxbox.cz> wrote:
> Hello ceph users and developers,
>
> I've stumbled upon a bit strange problem with Luminous.
>
> One of our servers running multiple QEMU clients crashed.
> When we tried restarting those on another cluster node,
> we got lots of fsck errors, disks seemed to return "physical"
> block errors. I figured this out to be stale RBD locks on volumes
> from the crashed machine. Wnen I removed the locks, everything
> started to work. (for some volumes, I was fixing those the another
> day after crash, so it was >10-15hours later)
>
> My question is, it this a bug or feature? I mean, after the client
> crashes, should locks somehow expire, or they need to be removed
> by hand? I don't remember having this issue with older ceph versions,
> but I suppose we didn't  have exclusive locks feature enabled..
>
> I'll be very grateful for any reply
>
> with best regards
>
> nik
> --
> -------------------------------------
> Ing. Nikola CIPRICH
> LinuxBox.cz, s.r.o.
> 28.rijna 168, 709 00 Ostrava
>
> tel.:   +420 591 166 214
> fax:    +420 596 621 273
> mobil:  +420 777 093 799
> www.linuxbox.cz
>
> mobil servis: +420 737 238 656
> email servis: servis at linuxbox.cz
> -------------------------------------
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



-- 
Jason


More information about the ceph-users mailing list