-
Notifications
You must be signed in to change notification settings - Fork 154
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Khala v0.1.7 node got occasionally full network stale #551
Comments
how to solve it? |
@coder-chenda No solution yet. Now it's mitigated by restarting every time it got stuck. You can also try to add |
same issue met, @h4x3rotab any progress?
|
@yrong No. According to the tracking issue paritytech/cumulus#607, there's nothing on our end can be done. If you have the control over all the collators, you can add |
@h4x3rotab thks will try and let you know |
Update: the current walkaround is to add We are still waiting for paritytech/cumulus#607 as the final fix. |
This can happen around once per day. Usually the entire network got stuck. A restart (or sometimes two) can have them move forward.
Symptom
The entire network got stuck at a certain block (include full nodes and collators, expect one collator, usually). The nodes got panic when trying to import blocks. When it occurs, each node will fall into one of the following errors, repeat the same message, and never recover until a full restart.
Records:
Type 1: Transaction will be valid in the future
Type 2: Storage root must match that calculated
Minor issue: NotInFinalizedChain
This one can happen after a restart of the node. It can also self recover after a few minutes.
Other background information
We have encountered similar problems in the past when we were running our parachain testnet v1. It's suspicious that we were also running a parachain runtime based on a version of substrate (polkadot) older than the client's. After upgraded the runtime to match the substrate version of the client, we had never seen the problem again, until now.
The text was updated successfully, but these errors were encountered: