-
Notifications
You must be signed in to change notification settings - Fork 975
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
blob.GetAll fails from a certain height #3185
Comments
Does this mean you restarted the node and tried the script again and it still did not work?
Did you reset the store and sync from scratch and then it worked? |
Might be related to #2915 |
Could you please run your node with metrics flags pointing to our otel collecter? We would also need your peerID (show at node start) to filter out your node metrics.
|
In all cases it ended in the same issue. Today I found out that restarting the node or the whole server fixes the issue shortly, but just increases that certain height: I couldn't query |
Got |
Seems related to celestiaorg/go-header#159 |
It is not related to the issue in the syncer. In case the requested height The possible root cause of this issue I've described in #2915. TLDR: by some reason, ipld can stuck during shares fetching. |
I have reproduced the issue and I can confirm your thesis @vgonkivs, because the error occurs when querying one blob more than one time. However, I've appended the logs of my node when reproducing: In another terminal, I have executed the following commands (every time I exited the curl execution with
|
@christopherbrumm, could you also test this behavior on the LN? I can see you using FN and we wonder if that's consistent across node types, particularly light |
I was able to reproduce the issue on the LN. Will provide logs if required. |
For completeness, linking the script to reproduce the issue in this thread: https://github.com/christopherbrumm/scripts/tree/main/celestia |
Hello, @christopherbrumm. Thanks for the script. It helped a lot. Yesterday I had a chance to reproduce this issue. It seems that @renaynay was right: the root cause was in syncer. This issue was fixed at celestiaorg/go-header#159. I tried your script on version v0.13(which already contains the fix) and for now, it works fine. Could you please also try v0.13? |
That's great news, will try and confirm it asap in order to close this issue. Thanks a lot! |
Hi @vgonkivs, after five attempts with different setups, I can confirm that the upgrade to v0.13 did not completely fix the reported issue. The node still stalls at a certain height and makes blob requests not possible. I am still using the provided script to replicate this issue. Also, I've noticed another issue that occurs when requesting share.GetEDS. Essentially, it mirrors the blob issue: the node stalls when requesting data above a certain level. I have added an additional script to reproduce this issue with more context in the README. I deployed five nodes on five different instances. While four of them have the issues listed, one is working fine. After realizing that one node has no problems providing blobs and shares, I tried to duplicate the setup. However, a node with identical setup shows the same problems again. cc @Wondertan |
Currently running 3 nodes on a fresh server and for all 3 the scripts are running through super smoothly. Will keep this thread updated. |
Hello @christopherbrumm. Thanks for your update. Will continue investigating this issue. May I also ask you to attach logs from the failed attempt? |
Update: The It seems that |
I guess we can close this issue then. Regarding the high duration of some requests: the current protocol for getting shares is slow(a known problem) and we are currently working on a better solution. cc @Wondertan |
Yes, lets close then |
Celestia Node version
Semantic version: v0.12.4 Commit: 8e5a717 Build Date: Fri Feb 9 11:20:30 UTC 2024 Golang version: go1.21.1
OS
System version: amd64/linux
Install tools
Followed step-by-step tutorial in the docs to install a celestia DA full node for Mainnet Beta
Others
Running on AWS
m5.4xlarge
:Steps to reproduce it
blob.GetAll
for each height, e.g. usingcurl -X POST -H "Content-Type: application/json" -H "Authorization: Bearer $CELESTIA_NODE_AUTH_TOKEN" -d '{ "id": 1, "jsonrpc": "2.0", "method": "blob.GetAll", "params": [ $HEIGHT, ["AAAAAAAAAAAAAAAAAAAAAAAAAIZiad33fbxA7Z0="] ] }' 127.0.0.1:26658
Expected result
I expect to get a response for every request to store all blobs for a given namespace.
Actual result
Instead, the request stalls starting from a certain height (observed starting from height
740731
,740740
), although requesting all blobs for a lower height works totally fine. The node doesn't throw errors when requesting, CPU and RAM usage limits are not exceeded. A node restart ends in the same issue, although resetting the store withcelestia full unsafe-reset-store
.Relevant log output
Couldn't find any possibly related node logs
Notes
No response
The text was updated successfully, but these errors were encountered: