-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
No peers found, but data exists #15
Comments
Ok one more step: I tried a regular start with
And I got those lines of log from mariadb pod:
|
It's quite possible that you need to do manual recovery. Now that some time has passed since I wrote https://github.com/Yolean/kubernetes-mysql-cluster#cluster-un-health it looks like it could need some more links. I haven't experienced a crash due to exceeding max connections. Sounds like a failure mode we can trigger in a test environment, but I guess there's no time for that now. |
You're not getting the crash loop in recovery mode, are you? If so it's a bad init script bug. Maybe, in recovery mode, you need to select the right node? |
@solsson Thanks. I got the right node to be start first. However, the crashloop still happen in recovery mode. |
fyi,
The right node is mariadb-2 which contains |
Sorry for late replies. I'm on vacation. Actually I don't know what to do. Is the error still the same as #15 (comment)? It's quite likely that there are issues with init.sh, in which case you need to try to patch your way around this (edit + re-apply If you want to try to start mariadb without editing the init script you can change the entrypoint of the mariadb container to something like |
Let's say I run a k8s statefulset and it works. However, if the pods are restarted (example case - k8s nodes are restarted), then |
@ngtuna The ambition is that as long as any of the pods is up and running the MariaDB cluster will recover - but I guess self-healing is a strong word there because that assumes that nothing is broken in any pod's volume. If all pods have been down concurrently, my interpretation of MariaDB/Galera docs is that they expect manual intervention. Can you recap the situation. Is the problem now that with I'd also be very interested in a repro case from a new scale=3 cluster. |
@solsson Yes I agree. And sorry for the strong word self-healing. In case all pods are shutdown-ed then we have to do manual recover 👍 I confirm that from the
|
Hi there, I have been experiencing the same issue with our implementation. When there is only the mariadb-0 pod left ( I have added some antiaffinity rules to stop the pods appearing on the same nodes) how do you edit the grastate.dat file manually and set safe_to_bootstrap to 1? Tom |
My mariadb cluster exceeded the max connections and crashed. I did delete the pods to make them recreated but they couldn't start. Checking the init container
init-config
I found those lines:So, I tried three of above options but no luck. The new pods are always
CrashLoopBackOff
Any suggestion would be very appreciated.
The text was updated successfully, but these errors were encountered: