-
Notifications
You must be signed in to change notification settings - Fork 94
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Node stopped synchronizing after usage of the new docker image #74
Comments
@MirekR anything in Main reason i have seen nodes crash or stop syncing is for those 2 |
Ah looking at your logs you did not clear the data in /db volume. As per the changelog on backend you need to clear and resync. |
Please reopen this if wiping the /db volume does not resolve. |
I did clear the db and I am stuck, should we re-open? |
Sure |
We shouldn't need to wipe /db volume everytime change comes in, it kills the global feed as well. |
@MikeKR We haven't had to - before NFTs mine has not had to resync for 2 months. Also the main issue with resync is not so much the blocks - they happen relatively quickly. Its the TxIndex which is getting replaced by new DB soon that should make it all much faster and easier. |
Performance of re-sync is one question, loosing nodes global feed is another and from the end user perspective more serious issue. |
You can keep the global feed in 2 ways: 1 - dont delete the whole db - eg make sure you keep the one in 2 - use the config option to load global state from a central node
|
Do I read this correctly as "we'll need to re-sync again soon and loose global again"? |
@MirekR Im sure the upgrade can be done in such a way that you keep global state. And note my reply above about ways to not loose global state. |
We're in the process of the backing store to postgres which will make these updates much less painful and more efficient. Global state migration tools will be provided. |
@maebeam may I ask where we are with this? And... has anyone considered a postgres docker image for a faster sync - at least at some set block height? I was having this reorg issue as recently as last weekend (9/4/21) - deso-protocol/core#98 |
Postgres is in beta but not ready to replace badger yet. What errors are you seeing? I haven’t seen any other reports of this issue recently |
Sorry for the delay... It's taking a long time to get anywhere - after a complete rebuild and resync. I get this often - but it's understandable... and at least still moving forward.
It's been 7 hours and still only at 11k - but like I said it's moving. Plus it turns out I was getting the "reorg" error loop at around 9427 - so at least we are beyond that:
I'll keep it going and see what develops |
Do you have a fast and reliable internet connection? I've only seen this type of behavior on low bandwidth / spotty internet. |
I have a fast connection, but also this is running on google cloud platform e2-standard-8 (8 vCPUs, 32 GB memory) ubuntu 20.04 with static IP |
Following up
|
Very weird. You may have faster syncing luck using larger SSD volumes. IOPS are determined by disk size on GCP |
lol. I thought 12 was fast, but even that is a barrier tbf. This latest setup was on a 250 GB SSD - but I hadn't considered the IOPS rate. Good point One day when nodes can go from 0 to sync in 5 minutes, devs will be able to focus on building a business instead of maintaining a node. And I'm sure us early adopters will profit by extension 😺 |
Hi @maebeam, I'm back with server stopped synchronisation.
After yesterday updated of the docker image to "19c446547510f1e0b83d56611e732e3fa6a0b32d" server stopped syncing, I can see last post 12h ago. It has also started ignoring super_admins, at the moment I can see only new post but not server stats ( I was able to for some time yesterday).
Docker compose file:
Full logs attached.
full.log
The text was updated successfully, but these errors were encountered: