-
-
Notifications
You must be signed in to change notification settings - Fork 416
feat: make aggregate with randomness async again #7761
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: make aggregate with randomness async again #7761
Conversation
Codecov ReportAttention: Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## unstable #7761 +/- ##
=========================================
Coverage 56.08% 56.08%
=========================================
Files 823 823
Lines 58031 58033 +2
Branches 4461 4467 +6
=========================================
+ Hits 32545 32550 +5
+ Misses 25419 25416 -3
Partials 67 67 🚀 New features to boost your workflow:
|
Performance Report✔️ no performance regression detected Full benchmark results
|
|
Hey Matthew and Lodestar team, thanks for giving this another try so quickly. I noticed a big change in event loop lag after upgrading our Gnosis Chain nodes from Lodestar 1.27.0-rc.0 (which included this change) to 1.29.0 (which didn't anymore): Even in 1.27.0-rc.0, event loop lag was kind of high but 1.29 made it worse, leading to very erratic API response times. I am now running this branch on one of our Gnosis nodes and so far it is looking very good. On the main thread, event loop lag is much better than on 1.29.0: In terms of the issues seen in #7218 . Event loop lag is worse on the network thread than on 1.29.0: GC on the network thread is only slightly worse than on 1.29.0:
And peers are looking just fine on this node: I'll keep an eye on this over the coming days but those API response times are much more stable now: |
|
need to make sure metrics on network thread is not as worse as monitored here |
|
shared some metrics on discord |
|
the bad peer count metrics I have last time does not happen, it just look like #7359 (comment) I do see this PR improved the main thread a lot that caused performance issue on network thread, I track the issue in #7801 so looks good to me overal, that's something for us to improve in the future |
nflaig
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
wemeetagain
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
metrics look good on this, excited to finally get it back
|
🎉 This PR is included in v1.30.0 🎉 |






Motivation
Gnosis analysis shows that there is a lot of time spent on main thread for running aggregateWithRandomness on main thread. Now that libp2p@v2 is more stable and so are some of the other conditions that triggered our revert last time, looking to check this again to see what the metrics look like for running not on main thread.