- We proposed a fix for the performance degradation observed when running
distributed multi-node benchmarks in the UTxO HD feature branch. While this
fixed the problems observed when running local benchmarks, it broke the
ThreadNettests due to concurrency issues. Therefore, we think it is wise to start redesigning the UTxO HD mempool integration.
- We did several rounds of code review on the alternative implementation of diff-sequences required by the UTxO HD feature based on the idea of anti-diffs. This alternative implementation is close to being merged, and the next step is to integrate this to the UTxO HD branch, so that we can run ad-hoc replaying and syncing from scratch benchmarks and compare these with the baseline. The micro-benchmarks we elaborated for the alternative implementation show speedups of up to 4x, so we are optimistic about the performance of replaying and syncing from scratch benchmarks, however it is important to notice that due to the nature of UTxO HD we will still be slower than the baseline.
- The final draft of the Genesis implementation specification is ready for review.
- We implemented a prototype for the happy path of Genesis' ChainSync Jumping (CSJ). The prototype is slower than the baseline, however it is not the latest version of the prototype and the jump interval is very small.
- Work on integrating Conway has stopped since priorities have changed.
- We started work on benchmarking epoch-boundaries and epoch overhead
pr-4014. To this end, we made use of a modified version of our
db-analysertool. We ran the new benchmarking setup using the Cardano mainnet chain, and we can see that block tick and application take substantially longer at epoch boundaries, although there are a couple of slots during an epoch in which these computations take more than normal. We notified the ledger team about these findings. We will use this modified version of
db-analyserto investigate the epoch overhead.
Spent quite some time investigating the root cause of the degradation in performance observed in the benchmarks. We run the
make forge-stressbenchmarks locally in order to debug this behavior.
Transaction batching doesn't make a notable difference in the outcome (considering we are using the in-memory backend).
The mempool batching implementation required asynchronous transaction validation which is a violation of the
LocalTxSubmissionprotocol contract and therefore if we continued on that route, the impact would have been quite big.
The STM logic we implemented by using a
TMVarfor the mempool internal state was buggy and under certain circumstances it seemed to lock. Reverting the mempool internal state to be stored in a
TVarseems to solve this problem.
The results we get after this change look almost identical to the ones from the baseline.
The anti-diff prototype (PR #3997) has been reviewed and is close to being merged.
- A follow-up issue (issue
to integrate the anti-diff prototype in the various
consensuspackages was created. A first version of the integration exists, and all tests pass. A next step is to get some indication of the "real" performance gain by profiling
- A follow-up issue (issue #4010) to integrate the anti-diff prototype in the various
Final draft of the Genesis implementation specification, now up for review.
Local benchmark setup for parameter tuning via the happy path ChainSync Jumping (CSJ) prototype (Issue 3987).
Context: Our Genesis design requires us to check in with a large (~20) number of servers periodically while syncing. These servers are offered jump requests via the ChainSync protocol (hence the name), which they can accept or decline. If a peer declines, the Genesis rule allows us to determine whether a node actually has a better chain.
The "happy path" is when no peer declines a jump. We want this to have close to no overhead compared to status quo, i.e. syncing without Genesis.
We implemented a prototype for this happy path, and are now starting to test in various configurations (number of peers, latency, bandwidth) to tune the performance of ChainSync jumping, i.e. how complicated our logic of choosing when to jump needs to be.
Simulated connection: 50 MBit/s, 50ms latency
Jump interval: 3000 slots (on the low end, could be increased to up to
Red: baseline (1.35.3), one peer in topology file
Blue: Preliminary version of our prototype, with 10 peers.
It is slower by about ~30%, but it is not the latest version of the prototype, and the jump interval is very small, making CSJ more of a bottleneck.
- Fix flakiness in ChainDB QSM tests (PR 3990).