sipa
commented at 7:06 pm on December 22, 2024:
member
Part of cluster mempool (#30289). Builds on top of #31444.
During reorganisations, it is possible that dependencies get added which would result in clusters that violate policy limits (cluster count, cluster weight), when linking the new from-block transactions to the old from-mempool transactions. Unlike RBF scenarios, we cannot simply reject the changes when they are due to received blocks. To accommodate this, add a TxGraph::Trim(), which removes some subset of transactions (including descendants) in order to make all resulting clusters satisfy the limits.
Conceptually, the way this is done is by defining a rudimentary linearization for the entire would-be too-large cluster, iterating it from beginning to end, and reasoning about the counts and weights of the clusters that would be reached using transactions up to that point. If a transaction is encountered whose addition would violate the limit, it is removed, together with all its descendants.
This rudimentary linearization is like a merge sort of the chunks of the clusters being combined, but respecting topology. More specifically, it is continuously picking the highest-chunk-feerate remaining transaction among those which have no unmet dependencies left. For efficiency, this rudimentary linearization is computed lazily, by putting all viable transactions in a heap, sorted by chunk feerate, and adding new transactions to it as they become viable.
The Trim() function is rather unusual compared to the TxGraph functionality added in previous PRs, in that Trim() makes it own decisions about what the resulting graph contents will be, without good specification of how it makes that decision - it is just a best-effort attempt (which is improved in the last commit). All other TxGraph mutators are simply to inform the graph about changes the calling mempool code decided on; this one lets the decision be made by txgraph.
As part of this, the “oversized” property is expanded to also encompass a configurable cluster weight limit (in addition to cluster count limit).
DrahtBot
commented at 7:06 pm on December 22, 2024:
contributor
The following sections might be updated with supplementary metadata relevant to reviewers and maintainers.
Try to run the tests locally, according to the documentation. However, a CI failure may still
happen due to a number of reasons, for example:
Possibly due to a silent merge conflict (the changes in this pull request being
incompatible with the current code in the target branch). If so, make sure to rebase on the latest
commit of the target branch.
A sanitizer issue, which can only be found by compiling with the sanitizer and running the
affected test.
An intermittent issue.
Leave a comment here, if you need help tracking down a confusing failure.
sipa force-pushed
on Jan 9, 2025
DrahtBot removed the label
CI failed
on Jan 9, 2025
sipa force-pushed
on Jan 9, 2025
in
src/txgraph.cpp:1704
in
0c8dc2323eoutdated
948+ Ref ret;
949+ // Construct a new Entry, and link it with the Ref.
950+ auto idx = m_entries.size();
951+ m_entries.emplace_back();
952+ auto& entry = m_entries.back();
953+ entry.m_ref = &ret;
I’m not sure how this is intended to be used, but storing a stack address seems like a problem? RVO may help but that seems brittle. I imagine the caller should be passing in their own Ref instead?
Ah, right, I missed that the move ctor would handle the update. Thanks for explaining.
in
src/txgraph.cpp:1172
in
0c8dc2323eoutdated
1167+ }
1168+}
1169+
1170+TxGraph::Ref& TxGraph::Ref::operator=(Ref&& other) noexcept
1171+{
1172+ // Inform both TxGraphs about the Refs being swapped.
Why is this doing an effective swap? I would expect this to call UnlinkRef on the moved-from value and reset its m_graph and m_index. Otherwise it wouldn’t be unlinked until the moved-from variable goes out of scope, no?
The vector now holds the old ref and UnlinkRef will not be called until that element is removed. I realize it’s allowed to be a “valid but unspecified state”, but I wouldn’t expect a ref to be hanging around.
in
src/txgraph.cpp:188
in
0c8dc2323eoutdated
183+ /** A class of objects held internally in TxGraphImpl, with information about a single
184+ * transaction. */
185+ struct Entry
186+ {
187+ /** Pointer to the corresponding Ref object, if any. */
188+ Ref* m_ref;
292+ LinearizationIndex lin_idx{0};
293+ // Iterate over the chunks.
294+ for (unsigned chunk_idx = 0; chunk_idx < chunking.NumChunksLeft(); ++chunk_idx) {
295+ auto chunk = chunking.GetChunk(chunk_idx);
296+ // Iterate over the transactions in the linearization, which must match those in chunk.
297+ while (true) {
Trying to convince myself this is guaranteed to terminate…
do{} while (!chunk.transactions.None()) rather than the break for readability? Or just while() if we need to guard against an empty linearization (presumably not?)
Every chunk contains at least one element (added an Assume for that)
In the inner loop, one element from that chunk is Reset() (added an Assume that it indeed resets a bit that was previously set).
I’ve changed it to a do {} while(chunk.transactions.Any()); loop in the first commits, though it reverts back to a while (true) { ... } loop later, when the loop becomes a bit more complex.
in
src/txgraph.cpp:822
in
0c8dc2323eoutdated
306+}
307+
308+void Cluster::ApplyRemovals(TxGraphImpl& graph, std::span<GraphIndex>& to_remove) noexcept
309+{
310+ // Iterate over the prefix of to_remove that applies to this cluster.
311+ SetType todo;
Done. I’ve also added a comment to the Cluster::ApplyRemovals() function definition stating that at least one element from the front of to_remove must belong to this Cluster (which is really why that requirement exists).
in
src/txgraph.cpp:1662
in
0c8dc2323eoutdated
1002+ // Make sure the transaction isn't scheduled for removal.
1003+ ApplyRemovals();
1004+ return m_entries[GetRefIndex(arg)].m_locator.IsPresent();
1005+}
1006+
1007+std::vector<TxGraph::Ref*> Cluster::GetAncestorRefs(const TxGraphImpl& graph, ClusterIndex idx) noexcept
Try to run the tests locally, according to the documentation. However, a CI failure may still
happen due to a number of reasons, for example:
Possibly due to a silent merge conflict (the changes in this pull request being
incompatible with the current code in the target branch). If so, make sure to rebase on the latest
commit of the target branch.
A sanitizer issue, which can only be found by compiling with the sanitizer and running the
affected test.
An intermittent issue.
Leave a comment here, if you need help tracking down a confusing failure.
sipa force-pushed
on Jan 10, 2025
DrahtBot removed the label
CI failed
on Jan 10, 2025
sdaftuar
commented at 2:10 pm on January 24, 2025:
FYI – in my rebase of #28676, I’m seeing tx_pool fuzz test failures due to this line. Not clear to me whether we should require the caller to enforce the policy requirement that a single tx be below the cluster size limit, or just let the caller discover a changeset is oversized and then reject?
Right. That rule exists because the alternative requires existing clusters to be oversized as AddTransaction constructs a singleton cluster instantly. All other forms of oversizedness happen as a result of applying dependencies, which are done lazily.
Done, it is now allowed to have individually oversized transactions.
sipa force-pushed
on Jan 24, 2025
sipa
commented at 10:14 pm on January 24, 2025:
member
Some changes:
As a result of dropping Cleanup in the base PR, Trim now reports which transactions it removed, as it becomes the caller’s responsibility of destroying Refs.
DrahtBot added the label
CI failed
on Jan 24, 2025
DrahtBot
commented at 11:23 pm on January 24, 2025:
contributor
Try to run the tests locally, according to the documentation. However, a CI failure may still
happen due to a number of reasons, for example:
Possibly due to a silent merge conflict (the changes in this pull request being
incompatible with the current code in the target branch). If so, make sure to rebase on the latest
commit of the target branch.
A sanitizer issue, which can only be found by compiling with the sanitizer and running the
affected test.
An intermittent issue.
Leave a comment here, if you need help tracking down a confusing failure.
sipa force-pushed
on Jan 26, 2025
sipa
commented at 4:42 am on January 26, 2025:
member
Add support for calling AddTransaction with a feerate whose size already violates the cluster size limit.
DrahtBot removed the label
CI failed
on Jan 26, 2025
sipa force-pushed
on Jan 26, 2025
sipa force-pushed
on Jan 30, 2025
DrahtBot
commented at 1:01 am on January 31, 2025:
contributor
Try to run the tests locally, according to the documentation. However, a CI failure may still
happen due to a number of reasons, for example:
Possibly due to a silent merge conflict (the changes in this pull request being
incompatible with the current code in the target branch). If so, make sure to rebase on the latest
commit of the target branch.
A sanitizer issue, which can only be found by compiling with the sanitizer and running the
affected test.
An intermittent issue.
Leave a comment here, if you need help tracking down a confusing failure.
DrahtBot added the label
CI failed
on Jan 31, 2025
sipa force-pushed
on Jan 31, 2025
DrahtBot removed the label
CI failed
on Jan 31, 2025
sipa force-pushed
on Jan 31, 2025
sipa force-pushed
on Feb 1, 2025
sipa force-pushed
on Feb 4, 2025
DrahtBot added the label
CI failed
on Feb 4, 2025
DrahtBot removed the label
CI failed
on Feb 4, 2025
sipa force-pushed
on Feb 6, 2025
sipa force-pushed
on Feb 11, 2025
sipa force-pushed
on Feb 12, 2025
DrahtBot
commented at 11:49 pm on February 12, 2025:
contributor
Try to run the tests locally, according to the documentation. However, a CI failure may still
happen due to a number of reasons, for example:
Possibly due to a silent merge conflict (the changes in this pull request being
incompatible with the current code in the target branch). If so, make sure to rebase on the latest
commit of the target branch.
A sanitizer issue, which can only be found by compiling with the sanitizer and running the
affected test.
An intermittent issue.
Leave a comment here, if you need help tracking down a confusing failure.
DrahtBot added the label
CI failed
on Feb 12, 2025
sipa force-pushed
on Feb 13, 2025
DrahtBot removed the label
CI failed
on Feb 13, 2025
sipa force-pushed
on Feb 14, 2025
sipa force-pushed
on Feb 20, 2025
sipa force-pushed
on Feb 21, 2025
in
src/cluster_linearize.h:1381
in
e43f6ca3b8outdated
1364+ // Figure out which elements need to be moved before elem.
1365+ SetType place_before = done & depgraph.Ancestors(elem);
1366+ // Find which position to place elem in (updating j), continuously moving the elements
1367+ // in between forward.
1368+ while (place_before.Any()) {
1369+ // j cannot be 0 here; if it was, then there was necessarily nothing earlier which
yancyribbens
commented at 10:20 pm on February 24, 2025:
0 // j cannot be 0 or less; if it was, then there was necessarily nothing earlier which
yancyribbens
commented at 10:23 pm on February 24, 2025:
0if it was, then there was necessarily nothing earlier which elem needs to be place before anymore, and place_before would be empty.
This comment seems jumbled and hard to understand. Is it possible to word this better?
Try to run the tests locally, according to the documentation. However, a CI failure may still
happen due to a number of reasons, for example:
Possibly due to a silent merge conflict (the changes in this pull request being
incompatible with the current code in the target branch). If so, make sure to rebase on the latest
commit of the target branch.
A sanitizer issue, which can only be found by compiling with the sanitizer and running the
affected test.
An intermittent issue.
Leave a comment here, if you need help tracking down a confusing failure.
sipa force-pushed
on Mar 19, 2025
DrahtBot removed the label
CI failed
on Mar 20, 2025
DrahtBot added the label
Needs rebase
on Mar 20, 2025
sipa force-pushed
on Mar 20, 2025
DrahtBot added the label
CI failed
on Mar 20, 2025
DrahtBot
commented at 1:42 pm on March 20, 2025:
contributor
Try to run the tests locally, according to the documentation. However, a CI failure may still
happen due to a number of reasons, for example:
Possibly due to a silent merge conflict (the changes in this pull request being
incompatible with the current code in the target branch). If so, make sure to rebase on the latest
commit of the target branch.
A sanitizer issue, which can only be found by compiling with the sanitizer and running the
affected test.
An intermittent issue.
Leave a comment here, if you need help tracking down a confusing failure.
sipa force-pushed
on Mar 20, 2025
DrahtBot removed the label
Needs rebase
on Mar 20, 2025
DrahtBot removed the label
CI failed
on Mar 20, 2025
sipa force-pushed
on Mar 22, 2025
DrahtBot added the label
CI failed
on Mar 22, 2025
DrahtBot
commented at 1:23 am on March 22, 2025:
contributor
Try to run the tests locally, according to the documentation. However, a CI failure may still
happen due to a number of reasons, for example:
Possibly due to a silent merge conflict (the changes in this pull request being
incompatible with the current code in the target branch). If so, make sure to rebase on the latest
commit of the target branch.
A sanitizer issue, which can only be found by compiling with the sanitizer and running the
affected test.
An intermittent issue.
Leave a comment here, if you need help tracking down a confusing failure.
sipa force-pushed
on Mar 22, 2025
DrahtBot removed the label
CI failed
on Mar 22, 2025
sipa force-pushed
on Mar 23, 2025
sipa force-pushed
on Mar 24, 2025
glozow referenced this in commit
f1d129d963
on Mar 26, 2025
sipa force-pushed
on Mar 27, 2025
sipa
commented at 3:49 am on March 27, 2025:
member
Rebased after merge of #31363, and on top of #32151.
sipa force-pushed
on Mar 27, 2025
sipa force-pushed
on Mar 27, 2025
sipa force-pushed
on Mar 28, 2025
sipa force-pushed
on Mar 28, 2025
sipa force-pushed
on Mar 28, 2025
sipa force-pushed
on Apr 4, 2025
sipa force-pushed
on Apr 4, 2025
sipa force-pushed
on Apr 7, 2025
sipa force-pushed
on Apr 7, 2025
DrahtBot added the label
Needs rebase
on Apr 17, 2025
sipa force-pushed
on Apr 17, 2025
DrahtBot removed the label
Needs rebase
on Apr 17, 2025
txgraph: Add GetMainStagingDiagrams function (feature)
This allows determining whether the changes in a staging diagram unambiguously improve
the graph, through CompareChunks().
e1cb50a957
txgraph: Maintain chunk index (preparation)
This is preparation for exposing mining and eviction functionality in
TxGraph.
This is preparation for a next commit which will introduce a class whose
objects hold references to internals in TxGraphImpl, which disallows
modifications to the graph while such objects exist.
9c2af16cd7
txgraph: Generalize GetClusterRefs to support subsections (preparation)
This is preparation for a next commit which will need a way to extract Refs
for just individual chunks from a cluster.
This interface lets one iterate efficiently over the chunks of the main
graph in a TxGraph, in the same order as CompareMainOrder. Each chunk
can be marked as "included" or "skipped" (and in the latter case,
dependent chunks will be skipped).
txgraph: Skipping end of cluster has no impact (optimization)df4578d2aa
txgraph: Special-case singletons in chunk index (optimization)a0becaaa9c
txgraph: Add ability to configure maximum cluster size/weight (feature)
This is integrated with the oversized property: the graph is oversized when
any connected component within it contains more than the cluster count limit
many transactions, or when their combined size/weight exceeds the cluster size
limit.
It becomes disallowed to call AddTransaction with a size larger than this limit.
In addition, SetTransactionFeeRate becomes SetTransactionFee, so that we do not
need to deal with the case that a call to this function might affect the
oversizedness.
7e869040fe
txgraph: Permit transactions that exceed cluster size limit (feature)f3ee8d75f2
txgraph: Add ability to trim oversized clusters (feature)
During reorganisations, it is possible that dependencies get add which
result in clusters that violate limits (count, size), when linking the
new from-block transactions to the old from-mempool transactions.
Unlike RBF scenarios, we cannot simply reject these policy violations
when they are due to received blocks. To accomodate this, add a Trim()
function to TxGraph, which removes transactions (including descendants)
in order to make all resulting clusters satisfy the limits.
c4c96fb3e3
txgraph: Track multiple potential would-be clusters in Trim (improvement)
In a Trim function, for any given would-be group of clusters, a (rudimentary)
linearization for the would-be cluster is constructed on the fly by adding
eligible transactions to a heap. This continues until the total count or
size of the transaction exists a configured limit. Any transactions which
appear later in this linearization are discarded.
However, given that transactions at the end are discarded, it is possible that
the would-be cluster splits apart into multiple clusters. And those clusters
may well permit far more transactions before their limits are reached.
Take this into account by using a union-find structure inside TrimTxData to
keep track of the count/size of all would-be clusters that would be formed
at any point.
This is not an optimization in terms of CPU usage or memory; it just
improves the quality of the transactions removed by Trim().
This is a metadata mirror of the GitHub repository
bitcoin/bitcoin.
This site is not affiliated with GitHub.
Content is generated from a GitHub metadata backup.
generated: 2025-04-28 12:13 UTC
This site is hosted by @0xB10C More mirrored repositories can be found on mirror.b10c.me