-
Notifications
You must be signed in to change notification settings - Fork 895
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Internal error, backward sync completed but failed to import finalized block #5053
Comments
I am seeing the same with besu+nimbus
|
Same on Besu 22.10.3 + Lighthouse 3.4.0 |
i had to wipe out the database and resync... now i get this error: 16:39:41.505+00:00 | nioEventLoopGroup-3-5 | INFO | MainnetBlockValidator | Invalid block 16574659 (0x1705627c83ba63c2f36eebc4a9b6f528fdb8da170da86b05f2d88b1b28222998): Optional[Unable to process block because parent world state 0xaafc4f9ce660df414b92bba48eae0129545714962267a5a383c3aac7342f8137 is not available] |
I also received this error, but I'm trying to sync an archive node so I can't just wipe the DB and start over... The error appears to have happened after the node crashed from OOM After importing all the PoW blocks, it started on the PoS blocks, I got this far:
Then the node OOM. After startup I got this:
The version is 22.10.3 |
Oracle VM
2.5 months |
Logs |
DEBUG Logs: |
More logs as requested. Set logging back to info, then sent this:
This is the resulting logs: Doesn't look super helpful but 🤷♂️ |
Last 1 hour of logs from Besu, Lighthouse, and Teku. Switching to Teku and rebooting seems to have "fixed it" |
@realsnick @smasher816 @estensen if you're still having this issue, based on quickchase's experience above, you could try switching to teku, letting it sync, then restarting EL. Can always switch back to another CL once it's in sync :) |
@realsnick which CL were you using when you got this error? |
Resynced from scratch like @realsnick Lost internet today and changed Wi-Fi while backward syncing was in progress, making it halt. Having to sync another CL seems like a big hacky workaround. |
Is the only option then to do a full resync again? @siladu |
It is indeed a workaround as we investigate this issue. We are treating this as a besu bug, but it could be a bug that is unique to certain combinations of execution + consensus clients.
We are actively looking at this. It's hard to recreate this situation, so any information users can provide is really valuable.
If you can't try another CL then yes, resync will fix it. |
Thanks for looking at it @siladu. I'm just a bit stressed that the db gets corrupted two weeks in a row. I do appreciate the hard work you guys are putting in. These are the logs from when it happened
I switched to my 5 GHz and then it stopped logging for 5 minutes. I then aborted and then I started getting I deleted the log files in the database dir and it managed to resolve the issue without resyncing |
It is unfortunate to happen twice in a row and I'm sure annoying. Be assured we will continue to work on stability issues like this as a top priority :) In addition, we have dramatically improved overall block processing performance and made syncing more robust, the effect being that resyncing from scratch is much less painful when it does have to happen. There is an upcoming release which is looking like it will be announced this week (I'm a bit reticent to say this as it's been delayed in the past!) so please bear with us and look out for announcements 🙏 |
Same problem here. After upgrade to Besu 22.10.3 it spams this on restart:
And after another restart repeating:
After another restart it seems to be back to normal. Not good for my heart health ;P |
@Armaver can I just confirm whether you mean upgrade to 22.10.3, or do you mean the recent release: 23.1.0? Also which CL are you using? |
Same problem here, twice in one week. |
Hi @SgtDinosaur which Besu version, sync-mode and CL are you using? |
CL client is Lighthouse.
|
I was using the second latest and then switched to the latest in the hope of fixing the issue. I now switched to Nethermind, hope to come back quick. My CL was Lighthouse. Thank you guys for responding and looking into it! |
) <!-- Thanks for sending a pull request! Please check out our contribution guidelines: --> <!-- https://github.com/hyperledger/besu/blob/main/CONTRIBUTING.md --> ## PR description There is an issue when restarting Besu when a backward sync session is running, since after the restart it is possible that the Consensus client sends a FcU or a NewPayload for a block that is present in the backward sync storage, but not yet imported, so not on the main chain, but still the backward sync thinks it should not do anything with that block, so it returns like it has completed the sync, but since the sync is not done actually then the internal error that the finalize block is not present. The solution is to persist the backward sync status, so in case of a restart, it can resume from where it was interrupted. ## Fixed Issue(s) <!-- Please link to fixed issue(s) here using format: fixes #<issue number> --> <!-- Example: "fixes #2" --> fixes #5053 ## Documentation - [x] I thought about documentation and added the `doc-change-required` label to this PR if [updates are required](https://wiki.hyperledger.org/display/BESU/Documentation). ## Acceptance Tests (Non Mainnet) - [x] I have considered running `./gradlew acceptanceTestNonMainnet` locally if my PR affects non-mainnet modules. ## Changelog - [x] I thought about the changelog and included a [changelog update if required](https://wiki.hyperledger.org/display/BESU/Changelog). --------- Signed-off-by: Fabio Di Fabio <[email protected]> Co-authored-by: Sally MacFarlane <[email protected]>
This fix has been released in https://github.com/hyperledger/besu/releases/tag/23.1.2 |
…perledger#5182) <!-- Thanks for sending a pull request! Please check out our contribution guidelines: --> <!-- https://github.com/hyperledger/besu/blob/main/CONTRIBUTING.md --> ## PR description There is an issue when restarting Besu when a backward sync session is running, since after the restart it is possible that the Consensus client sends a FcU or a NewPayload for a block that is present in the backward sync storage, but not yet imported, so not on the main chain, but still the backward sync thinks it should not do anything with that block, so it returns like it has completed the sync, but since the sync is not done actually then the internal error that the finalize block is not present. The solution is to persist the backward sync status, so in case of a restart, it can resume from where it was interrupted. ## Fixed Issue(s) <!-- Please link to fixed issue(s) here using format: fixes #<issue number> --> <!-- Example: "fixes #2" --> fixes hyperledger#5053 ## Documentation - [x] I thought about documentation and added the `doc-change-required` label to this PR if [updates are required](https://wiki.hyperledger.org/display/BESU/Documentation). ## Acceptance Tests (Non Mainnet) - [x] I have considered running `./gradlew acceptanceTestNonMainnet` locally if my PR affects non-mainnet modules. ## Changelog - [x] I thought about the changelog and included a [changelog update if required](https://wiki.hyperledger.org/display/BESU/Changelog). --------- Signed-off-by: Fabio Di Fabio <[email protected]> Co-authored-by: Sally MacFarlane <[email protected]>
…perledger#5182) <!-- Thanks for sending a pull request! Please check out our contribution guidelines: --> <!-- https://github.com/hyperledger/besu/blob/main/CONTRIBUTING.md --> ## PR description There is an issue when restarting Besu when a backward sync session is running, since after the restart it is possible that the Consensus client sends a FcU or a NewPayload for a block that is present in the backward sync storage, but not yet imported, so not on the main chain, but still the backward sync thinks it should not do anything with that block, so it returns like it has completed the sync, but since the sync is not done actually then the internal error that the finalize block is not present. The solution is to persist the backward sync status, so in case of a restart, it can resume from where it was interrupted. ## Fixed Issue(s) <!-- Please link to fixed issue(s) here using format: fixes #<issue number> --> <!-- Example: "fixes #2" --> fixes hyperledger#5053 ## Documentation - [x] I thought about documentation and added the `doc-change-required` label to this PR if [updates are required](https://wiki.hyperledger.org/display/BESU/Documentation). ## Acceptance Tests (Non Mainnet) - [x] I have considered running `./gradlew acceptanceTestNonMainnet` locally if my PR affects non-mainnet modules. ## Changelog - [x] I thought about the changelog and included a [changelog update if required](https://wiki.hyperledger.org/display/BESU/Changelog). --------- Signed-off-by: Fabio Di Fabio <[email protected]> Co-authored-by: Sally MacFarlane <[email protected]>
execution_1 | Setting logging level to INFO
execution_1 | 2023-02-04 19:54:23.830+00:00 | main | INFO | Besu | Starting Besu
execution_1 | 2023-02-04 19:54:24.516+00:00 | main | WARN | Besu | --graphql-http-host has been ignored because --graphql-http-enabled was not defined on the command line.
execution_1 | 2023-02-04 19:54:24.693+00:00 | main | INFO | Besu |
execution_1 | ####################################################################################################
execution_1 | # #
execution_1 | # Besu 23.1.0-dev-e7150102 #
execution_1 | # #
execution_1 | # Configuration: #
execution_1 | # Network: Mainnet #
execution_1 | # Network Id: 1 #
execution_1 | # Data storage: Bonsai #
execution_1 | # Sync mode: Snap #
execution_1 | # RPC HTTP APIs: WEB3,ETH,NET #
execution_1 | # RPC HTTP port: 8545 #
execution_1 | # Engine APIs: ENGINE,ETH #
execution_1 | # Engine port: 8551 #
execution_1 | # #
execution_1 | # Host: #
execution_1 | # Java: openjdk-java-17 #
execution_1 | # Maximum heap size: 64.00 GB #
execution_1 | # OS: linux-x86_64 #
execution_1 | # glibc: 2.31 #
execution_1 | # Total memory: 503.87 GB #
execution_1 | # CPU cores: 80 #
execution_1 | # #
execution_1 | ####################################################################################################
execution_1 | 2023-02-04 19:54:24.694+00:00 | main | INFO | Besu | Connecting to 0 static nodes.
execution_1 | 2023-02-04 19:54:24.694+00:00 | main | INFO | Besu | Security Module: localfile
execution_1 | 2023-02-04 19:54:24.756+00:00 | main | INFO | Besu | Using the native implementation of alt bn128
execution_1 | 2023-02-04 19:54:24.808+00:00 | main | INFO | Besu | Using the native implementation of modexp
execution_1 | 2023-02-04 19:54:24.808+00:00 | main | INFO | Besu | Using the native implementation of the signature algorithm
execution_1 | 2023-02-04 19:54:24.814+00:00 | main | INFO | Besu | Using the native implementation of the blake2bf algorithm
execution_1 | 2023-02-04 19:54:24.840+00:00 | main | INFO | DatabaseMetadata | Lookup database metadata file in data directory: /var/lib/besu
execution_1 | 2023-02-04 19:54:24.910+00:00 | main | INFO | RocksDBKeyValueStorageFactory | Existing database detected at /var/lib/besu. Version 2
execution_1 | 2023-02-04 19:54:30.154+00:00 | main | INFO | KeyPairUtil | Loaded public key 0xba5cb5928621f5ef140f8182dde0c6f0c0447562b2b4d7763b72da5c9ce3df41fd872a26687466e7b711ef53d74526bf724b12fe6217504a76268d815c951cd0 from /var/lib/besu/key
execution_1 | 2023-02-04 19:54:30.353+00:00 | main | INFO | AbstractProtocolScheduleBuilder | Protocol schedule created with milestones: [Frontier: 0, Homestead: 1150000, DaoRecoveryInit: 1920000, DaoRecoveryTransition: 1920001, Homestead: 1920010, TangerineWhistle: 2463000, SpuriousDragon: 2675000, Byzantium: 4370000, Petersburg: 7280000, Istanbul: 9069000, MuirGlacier: 9200000, Berlin: 12244000, London: 12965000, ArrowGlacier: 13773000, GrayGlacier: 15050000]
execution_1 | 2023-02-04 19:54:30.392+00:00 | main | INFO | AbstractProtocolScheduleBuilder | Protocol schedule created with milestones: [Frontier: 0, Homestead: 1150000, DaoRecoveryInit: 1920000, DaoRecoveryTransition: 1920001, Homestead: 1920010, TangerineWhistle: 2463000, SpuriousDragon: 2675000, Byzantium: 4370000, Petersburg: 7280000, Istanbul: 9069000, MuirGlacier: 9200000, Berlin: 12244000, London: 12965000, ArrowGlacier: 13773000, GrayGlacier: 15050000]
execution_1 | 2023-02-04 19:54:30.401+00:00 | main | INFO | AbstractProtocolScheduleBuilder | Protocol schedule created with milestones: []
execution_1 | 2023-02-04 19:54:31.232+00:00 | main | INFO | TransactionPoolFactory | Enabling transaction pool
execution_1 | 2023-02-04 19:54:31.233+00:00 | main | INFO | TransactionPoolFactory | Disabling transaction handling during initial sync
execution_1 | 2023-02-04 19:54:31.236+00:00 | main | INFO | EthPeers | Updating the default best peer comparator
execution_1 | 2023-02-04 19:54:31.260+00:00 | main | INFO | BesuControllerBuilder | TTD difficulty is present, creating initial sync for PoS
execution_1 | 2023-02-04 19:54:31.276+00:00 | main | INFO | SnapDownloaderFactory | Snap sync was requested, but cannot be enabled because the local blockchain is not empty.
execution_1 | 2023-02-04 19:54:31.278+00:00 | main | INFO | TransitionBesuControllerBuilder | TTD present, creating DefaultSynchronizer that stops propagating after finalization
execution_1 | 2023-02-04 19:54:31.308+00:00 | main | INFO | RunnerBuilder | Detecting NAT service.
execution_1 | 2023-02-04 19:54:31.474+00:00 | main | INFO | AbstractProtocolScheduleBuilder | Protocol schedule created with milestones: []
execution_1 | 2023-02-04 19:54:31.537+00:00 | main | INFO | AbstractProtocolScheduleBuilder | Protocol schedule created with milestones: []
execution_1 | 2023-02-04 19:54:31.768+00:00 | main | INFO | AbstractProtocolScheduleBuilder | Protocol schedule created with milestones: []
execution_1 | 2023-02-04 19:54:31.785+00:00 | main | INFO | Runner | Starting external services ...
execution_1 | 2023-02-04 19:54:31.786+00:00 | main | INFO | MetricsHttpService | Starting metrics http service on 0.0.0.0:6060
execution_1 | 2023-02-04 19:54:31.889+00:00 | vert.x-eventloop-thread-1 | INFO | MetricsHttpService | Metrics service started and listening on 0.0.0.0:6060
execution_1 | 2023-02-04 19:54:31.889+00:00 | main | INFO | JsonRpcHttpService | Starting JSON-RPC service on 0.0.0.0:8545
execution_1 | 2023-02-04 19:54:31.905+00:00 | vert.x-eventloop-thread-1 | INFO | JsonRpcHttpService | JSON-RPC service started and listening on 0.0.0.0:8545
execution_1 | 2023-02-04 19:54:31.905+00:00 | main | INFO | EngineJsonRpcService | Starting JSON-RPC service on 0.0.0.0:8551
execution_1 | 2023-02-04 19:54:31.911+00:00 | vert.x-eventloop-thread-1 | INFO | EngineJsonRpcService | JSON-RPC service started and listening on 0.0.0.0:8551
execution_1 | 2023-02-04 19:54:31.912+00:00 | main | INFO | WebSocketService | Starting Websocket service on 0.0.0.0:8546
execution_1 | 2023-02-04 19:54:31.915+00:00 | vert.x-eventloop-thread-1 | INFO | WebSocketService | Websocket service started and listening on 0.0.0.0:8546
execution_1 | 2023-02-04 19:54:31.918+00:00 | main | INFO | AutoTransactionLogBloomCachingService | Starting auto transaction log bloom caching service.
execution_1 | 2023-02-04 19:54:31.919+00:00 | main | INFO | LogBloomCacheMetadata | Lookup cache metadata file in data directory: /var/lib/besu/caches
execution_1 | 2023-02-04 19:54:31.928+00:00 | main | INFO | Runner | Starting Ethereum main loop ...
execution_1 | 2023-02-04 19:54:31.928+00:00 | main | INFO | DockerNatManager | Starting docker NAT manager.
execution_1 | 2023-02-04 19:54:31.937+00:00 | main | INFO | NetworkRunner | Starting Network.
execution_1 | 2023-02-04 19:54:31.940+00:00 | main | INFO | DefaultP2PNetwork | Starting DNS discovery with URL enrtree://AKA3AM6LPBYEUDMVNU3BSVQJ5AD45Y7YPOHJLEF6W26QOE4VTUDPE@all.mainnet.ethdisco.net
execution_1 | 2023-02-04 19:54:32.010+00:00 | nioEventLoopGroup-2-1 | INFO | RlpxAgent | P2P RLPx agent started and listening on /0.0.0.0:30303.
execution_1 | 2023-02-04 19:54:32.011+00:00 | main | INFO | PeerDiscoveryAgent | Starting peer discovery agent on host=0.0.0.0, port=30303
execution_1 | 2023-02-04 19:54:32.054+00:00 | vert.x-eventloop-thread-1 | INFO | VertxPeerDiscoveryAgent | Started peer discovery agent successfully, on effective host=0.0.0.0 and port=30303
execution_1 | 2023-02-04 19:54:32.055+00:00 | vert.x-eventloop-thread-1 | INFO | PeerDiscoveryAgent | P2P peer discovery agent started and listening on /0.0.0.0:30303
execution_1 | 2023-02-04 19:54:32.108+00:00 | vert.x-eventloop-thread-1 | INFO | PeerDiscoveryAgent | Writing node record to disk. NodeRecord{seq=83, publicKey=0x02ba5cb5928621f5ef140f8182dde0c6f0c0447562b2b4d7763b72da5c9ce3df41, udpAddress=Optional[/127.0.0.1:30303], tcpAddress=Optional[/127.0.0.1:30303], asBase64=-Je4QFFXzeqiZb2EpXqAagdzEjnpXxq90PzI95N7LoRAxSRWULLBJzCv-OzLzUZWApHzyrurQLxpqvSHPyh4a_b9s0hTg2V0aMfGhPCv0OOAgmlkgnY0gmlwhH8AAAGJc2VjcDI1NmsxoQK6XLWShiH17xQPgYLd4MbwwER1YrK013Y7ctpcnOPfQYN0Y3CCdl-DdWRwgnZf, nodeId=0x191631e5043602531ea9c04190dd41bbc6f07a2a368438865f39dea4bdce1e33, customFields={tcp=30303, udp=30303, ip=0x7f000001, eth=[[0xf0afd0e3, 0x]], id=V4, secp256k1=0x02ba5cb5928621f5ef140f8182dde0c6f0c0447562b2b4d7763b72da5c9ce3df41}}
execution_1 | 2023-02-04 19:54:32.181+00:00 | main | INFO | DefaultP2PNetwork | Enode URL enode://ba5cb5928621f5ef140f8182dde0c6f0c0447562b2b4d7763b72da5c9ce3df41fd872a26687466e7b711ef53d74526bf724b12fe6217504a76268d815c951cd0@127.0.0.1:30303
execution_1 | 2023-02-04 19:54:32.183+00:00 | main | INFO | DefaultP2PNetwork | Node address 0x90dd41bbc6f07a2a368438865f39dea4bdce1e33
execution_1 | 2023-02-04 19:54:32.189+00:00 | main | INFO | NetworkRunner | Supported capabilities: [eth/63, eth/64, eth/65, eth/66, eth/67, eth/68], [snap/1]
execution_1 | 2023-02-04 19:54:32.189+00:00 | main | INFO | DefaultSynchronizer | Starting synchronizer.
execution_1 | 2023-02-04 19:54:32.192+00:00 | main | INFO | TransactionPoolFactory | Enabling transaction handling following initial sync
execution_1 | 2023-02-04 19:54:32.196+00:00 | main | INFO | DefaultSynchronizer | Stopping block propagation.
execution_1 | 2023-02-04 19:54:32.197+00:00 | main | INFO | DefaultSynchronizer | Stopping the pruner.
execution_1 | 2023-02-04 19:54:32.198+00:00 | main | INFO | Runner | Ethereum main loop is up.
execution_1 | 2023-02-04 19:54:54.203+00:00 | Timer-0 | INFO | DNSResolver | Resolved 2409 nodes
execution_1 | 2023-02-04 19:56:06.947+00:00 | vert.x-worker-thread-0 | INFO | BackwardSyncContext | Starting a new backward sync session
execution_1 | 2023-02-04 19:56:06.955+00:00 | vert.x-worker-thread-0 | INFO | BackwardsSyncAlgorithm | Current backward sync session is done
execution_1 | 2023-02-04 19:56:08.179+00:00 | vert.x-worker-thread-0 | WARN | MergeCoordinator | Internal error, backward sync completed but failed to import finalized block 0xe8f33c4587ca5e19d441139e1d45ba19bb58450b89e1e4e5e21d179b858ab964
execution_1 | 2023-02-04 19:56:08.196+00:00 | vert.x-worker-thread-0 | WARN | MergeCoordinator | Internal error, backward sync completed but failed to import finalized block 0xe8f33c4587ca5e19d441139e1d45ba19bb58450b89e1e4e5e21d179b858ab964
execution_1 | 2023-02-04 19:56:36.760+00:00 | vert.x-worker-thread-0 | WARN | MergeCoordinator | Internal error, backward sync completed but failed to import finalized block 0xe8f33c4587ca5e19d441139e1d45ba19bb58450b89e1e4e5e21d179b858ab964
execution_1 | 2023-02-04 19:56:38.094+00:00 | vert.x-worker-thread-0 | WARN | MergeCoordinator | Internal error, backward sync completed but failed to import finalized block 0xe8f33c4587ca5e19d441139e1d45ba19bb58450b89e1e4e5e21d179b858ab964
execution_1 | 2023-02-04 19:56:38.320+00:00 | vert.x-worker-thread-0 | WARN | MergeCoordinator | Internal error, backward sync completed but failed to import finalized block 0xe8f33c4587ca5e19d441139e1d45ba19bb58450b89e1e4e5e21d179b858ab964
execution_1 | 2023-02-04 19:56:43.985+00:00 | vert.x-worker-thread-0 | WARN | MergeCoordinator | Internal error, backward sync completed but failed to import finalized block 0xe8f33c4587ca5e19d441139e1d45ba19bb58450b89e1e4e5e21d179b858ab964
execution_1 | 2023-02-04 19:56:49.107+00:00 | vert.x-worker-thread-0 | WARN | MergeCoordinator | Internal error, backward sync completed but failed to import finalized block 0xe8f33c4587ca5e19d441139e1d45ba19bb58450b89e1e4e5e21d179b858ab964
execution_1 | 2023-02-04 19:57:00.682+00:00 | vert.x-worker-thread-0 | WARN | MergeCoordinator | Internal error, backward sync completed but failed to import finalized block 0xe8f33c4587ca5e19d441139e1d45ba19bb58450b89e1e4e5e21d179b858ab964
The text was updated successfully, but these errors were encountered: