-
Notifications
You must be signed in to change notification settings - Fork 9.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
3.3.21 crashing on every node with "etcdmain: walpb: crc mismatch" #11918
Comments
Once wal file is purged, whether it is an existing cluster or a new cluster, restarting etcd(v3.4.8/v3.3.21) will encounter this issue. Please wait for the new release version to solve this problem. thanks. @Roguelazer |
upgrade from 3.3.19 to 3.3.21 ,i sovle this issue by step,but i don't know why.:
|
@wcollin wal file maybe has not been purged, if you restart etcd now, it will also be crash. v3.3.22,v3.4.9 is not found. @gyuho |
@tangcong thanks a lot ,i will fallback to 3.3.19 and waiting for new version. |
I have verified that 3.3.22 appears to fix this issue. It does now print a bunch of other unexpected errors when restarting a node:
The node comes up successfully, though, and appears to function normally |
@Roguelazer it is normal behavior, it will print warn log when etcd server failed to apply request , it plays a key role in troubleshooting. |
I'm getting the same issue. any news ? |
Hi @karolsteve - etcd 3.3.21 is a three and a half year old release which is no longer supported by the etcd project. Countless bugs and security concerns have been addressed in later releases. Please upgrade to a modern etcd release as a soon as possible. |
okay thanks @jmhbnz |
I am attempting to do a rolling upgrade from 3.3.20 to 3.3.21, and every node has failed with "etcdmain: walpb: crc mismatch". For the first node, I removed it and re-bootstrapped it successfully, and it can stop and start fine on 3.3.21 after re-bootstrapping, but now that the second node in the cluster is crashing on startup with the same error, I'm a little suspicious that this is actually a bug in 3.3.21.
None of these machines have ever suffered any hardware failure or unexpected shutdown, and I have been upgrading them to every 3.3.x patch release without ever seeing this issue before.
These nodes are mostly used with V2-protocol-speaking applications, so almost all the data is in the V2 store. Not sure if that makes a difference.
The text was updated successfully, but these errors were encountered: