CentOS unrecoverable after Ceph issues
CHANU ROMAIN
romain.chanu at univ-lyon1.fr
Tue Jul 21 13:45:18 UTC 2020
Hello,
I do not use CentOS and XFS but I had a simillar issue after an
outrage. Ceph didnt release the lock on rados block device. You can
check if you are facing the same issue than I did. You have to shutdown
your instance then type this command:
rbd -p your-pool-name lock list instance-volume-id
The command should not return any output if your instance is shut. If
you got an output about 1 exclusive lock just remove it:
rbd -p your-pool-name lock remove instance-volume-id
Best Regards,Romain
On Tue, 2020-07-21 at 14:04 +0100, Grant Morley wrote:
> Hi all,
> We recently had an issue with our ceph cluster which ended up
> going into "Error" status after some drive failures. The system
> stopped allowing writes for a while whilst it recovered. The
> ceph
> cluster is healthy again but we seem to have a few instances
> that
> have corrupt filesystems on them. They are all CentOS 7
> instances.
> We have got them into rescue mode to try and repair the FS with
> "xfs_repair -L" However when we do that we get this:
> 973.026283]
> XFS (vdb1): Mounting V5 Filesystem
>
> [ 973.203261] blk_update_request: I/O error, dev vdb, sector
> 8389693
>
> [ 973.204746] blk_update_request: I/O error, dev vdb, sector
> 8390717
>
> [ 973.206136] blk_update_request: I/O error, dev vdb, sector
> 8391741
>
> [ 973.207608] blk_update_request: I/O error, dev vdb, sector
> 8392765
>
> [ 973.209544] XFS (vdb1): xfs_do_force_shutdown(0x1) called
> from
> line 1236 of file fs/xfs/xfs_buf.c. Return address =
> 0xffffffffc017a50c
>
> [ 973.212137] XFS (vdb1): I/O Error Detected. Shutting down
> filesystem
>
> [ 973.213429] XFS (vdb1): Please umount the filesystem and
> rectify the problem(s)
>
> [ 973.215036] XFS (vdb1): metadata I/O error: block 0x7ffc3d
> ("xlog_bwrite") error 5 numblks 8192
>
> [ 973.217201] XFS (vdb1): failed to locate log tail
>
> [ 973.218239] XFS (vdb1): log mount/recovery failed: error -5
>
> [ 973.219865] XFS (vdb1): log mount failed
>
> [ 973.233792] blk_update_request: I/O error, dev vdb, sector
> 0
> Interestingly
> any debian based instances we could recover. It just seems to
> be
> CentOS and having XFS on CentOS and ceph the instances don't
> seem happy. This seems more low level to me in ceph rather
> than
> a corrupt FS on a guest.
> Does anyone
> know of any "ceph tricks" that we can use to try and at least
> get an "xfs_repair" running?
> Many thanks,
>
>
>
> --
>
>
> Grant Morley
>
> Cloud Lead, Civo Ltd
>
> www.civo.com
>
>
>
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20200721/a3725126/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/x-pkcs7-signature
Size: 4843 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20200721/a3725126/attachment.bin>
More information about the openstack-discuss
mailing list