-
Notifications
You must be signed in to change notification settings - Fork 563
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
NodeStageVolume fails if xfs_repairs returns error after cluster reboot #859
Comments
Can confirm this issue. Simple "rbd map" and a "mount -t xfs .." and unmap/unmount afterwards seems to replay the log and fixes the issue. No xfs_repair needed! Currently I'm not really sure after what action (pod recreation etc.) this error occurs. |
I believe this is not an error in the csi code. The |
we are using. we had hit an issue in xfs_repair as this is fixed in recent kubernetes utils we are updating it for the same reason more info at kubernetes/utils#141 fixes #859 updates rook/rook#4914 Signed-off-by: Humble Chirammal <hchiramm@redhat.com>
NOTE: This PR also updates the kubernetes utils packages we are using. we had hit an issue in xfs_repair as this is fixed in recent kubernetes utils we are updating it for the same reason more info at kubernetes/utils#141 fixes ceph#859 updates rook/rook#4914 Signed-off-by: Humble Chirammal <hchiramm@redhat.com>
NOTE: This PR also updates the kubernetes utils packages we are using. we had hit an issue in xfs_repair as this is fixed in recent kubernetes utils we are updating it for the same reason more info at kubernetes/utils#141 fixes ceph#859 updates rook/rook#4914 Signed-off-by: Humble Chirammal <hchiramm@redhat.com>
I had the same problems, and it is a pain to manually mount the volume on a host node, using rbd map when running kubernetes. |
@whymatter can you try with |
I will give it a try |
fixed in v2.1.0, if not please feel free to reopen it |
Describe the bug
When there are "valuable metadata changes in a log which needs to be replayed" pod creation fails. (See log)
The reason this is a bug and not a feature for me is that this happens after a sudden cluster reboot.
Environment details
Kubernetes Version:
Image/version of Ceph CSI driver
quay.io/cephcsi/cephcsi:v2.0.0
Deployed using rook.io
Logs
Steps to reproduce
Steps to reproduce the behavior:
For me this happens every time I have a simple pod connected to a ceph block PV which uses the xfs file system. After a reboot, the pod can not be recreated.
Actual results
The csi driver tries to run
xfs_repair
but reports an error stating that the volume has to be mounted first.Expected behavior
In my case simply mounting the device (manually) resolved the problem. So I guess there should be a chance to fix this issue automatically by temporarily mounting the volume?
Additional context
Related issues:
The text was updated successfully, but these errors were encountered: