-
Notifications
You must be signed in to change notification settings - Fork 1.8k
feat: added parallelization on key partitioned data #18919
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: added parallelization on key partitioned data #18919
Conversation
| pub preserve_partition_values: bool, | ||
| /// Cached result of key_partition_exprs computation to avoid repeated work | ||
| #[allow(clippy::type_complexity)] | ||
| key_partition_exprs_cache: OnceLock<Option<Vec<Arc<dyn PhysicalExpr>>>>, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Caches results of compute_key_partition_exprs() which is expensive:
- loops through file groups and does hash set operations
- called multiple times (output_partitioning() and eq_properties())
| } | ||
| Distribution::KeyPartitioned(_) => { | ||
| // Nothing to do: treated as satisfied upstream | ||
| } |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No-op because we can guarantee that our data is correctly distributed
| 02)--AggregateExec: mode=FinalPartitioned, gby=[a@0 as a], aggr=[nth_value(multiple_ordered_table.c,Int64(1)) ORDER BY [multiple_ordered_table.c ASC NULLS LAST]], ordering_mode=Sorted | ||
| 03)----SortExec: expr=[a@0 ASC NULLS LAST], preserve_partitioning=[true] | ||
| 04)------CoalesceBatchesExec: target_batch_size=8192 | ||
| 05)--------RepartitionExec: partitioning=Hash([a@0], 4), input_partitions=4 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Eliminates this hash because it would break ordering guarantees
|
Thanks a lot for the description and companion doc, they are super useful. This work is super nice and is even crucial for distributed DataFusion. Reusing partitioning and avoiding repartitions can make a huge difference when the repartition is done on the network. The plans you posted as examples are exactly what we should be aiming for. I think I am still missing part of the point of
So my current understanding is: Sorry for the wall of text, I am mostly trying to wrap my head around this, please correct anything I missed in here. |
| /// Allocate rows based on a hash of one of more expressions and the specified number of | ||
| /// partitions | ||
| Hash(Vec<Arc<dyn PhysicalExpr>>, usize), | ||
| /// Partitions that are already organized by disjoint key values for the provided expressions. | ||
| /// Rows that have the same values for these expressions are guaranteed to be in the same partition. | ||
| KeyPartitioned(Vec<Arc<dyn PhysicalExpr>>, usize), |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
My impression over all is that KeyPartitioned should not be adding anything that is not already representable with Hash. I was planning on doing a longer reasoning on this, but @fmonjalet is right on point in his comment here #18919 (comment), so I'd just +1 his comment, grab some 🍿, and see what comes out of it.
Yes, key partitioning guarantees that each distinct value of the key is fully contained within a single partition which is pretty much a stronger hash partitioning. Another thing to note is that key partitioning can only root from the file scan level as of now compared to hash which of course has a repartitioning operator.
Yes, I believe you have the right idea but to be sure,
Yes, this is a noted limit to the original design. I added the comment: "best with moderate partition counts (10-100 partitions)." in the config. This is rooting from splitting distinct keys into their own partitions as of now. I did this to keep the first iteration relatively simple as the PR is large. In a follow up issue, some gerat work would be to merge group to
It depends what "group" means. If we simply merge key partitioned data into a single partition, no, this is still key partitioned as each key is still fully in one place. If we are repartitioning or shuffling data, we lose key partitioning and fallback to hash
For this first PR, yes, but I think this isn't a one PR fix all scenario. I think this comes down to how intentional the user is. Yes, key partitioned data is rarer than say hash, but it is powerful enough for people to consider it. The use cases will also increse as follow up issues are resolved: higher cardinality, propagation through joins, etc.
BEFORE: DataSourceExec -> Aggregate Partial (gby: a) -> Repartition Hash(a) -> Aggregate Final (gby: a) In some cases we also eliminate bottlenecks due to SPMs between aggregations:
I am in favor of keeping Hash and KeyPartitioned separate as I see them as two distinct methods of partitioning. I also don't knowif adding more information into Hash partitioning will eliminate cimplexity and raher just cause more indirection. I do like the idea of merging file groups for higher cardinality as this was my main concern with this v1 (as noted in the comments) but chose to refrain due to complexity.
Do not apologize, this is a lot of the internal debates I was / am having and am glad to talk about the trade offs. Let me know what you think 😄 CC: @gabotechs |
|
I suggest, that this PR remain the limited scope, not meant for high cardinality queries. This was my motivation for having this option set to false by default. Then submit follow up issues to address grouping files into partitions to help with higher cardinality. I just do not want to introduce to many things in this PR and adding this seems like another substantial PR in itself. |
I think @fmonjalet's suggestion (and mine) is to not introduce I think avoiding extra repartitions by reusing an existing hash partitioning in an operator that requires some other hash partitioning that just partially matches the incoming one should be achievable without introducing new partitioning methods. Also note that ideally the amount of partitions in a data stream should not be given by the nature of the data, but by the amount of CPUs a machine has, that's what allows us to optimize for resource usage regardless of how the data happened to be laid out. |
|
Thanks a lot for the explanations @gene-bordegaray, I think I actually start to understand 💡 When the partitioning is by one field, Technically this layout satisfies all the following partitioning:
Now you want to compute: SELECT customer_id, SUM(amount) FROM orders GROUP BY customer_id
The following query (a bit artificial ,sorry): WITH max_order_per_provider AS (
SELECT customer_id, provider_id, MAX(amount) AS max_amount FROM orders GROUP BY customer_id, provider_id
)
SELECT customer_id, MIN(max_amount) as min_max FROM max_order_per_provider GROUP BY customer_id
From there, I see I am now wondering about whether we should have |
|
I will review this PR this week, too |
|
hey @fmonjalet thank you for the thoughtful response. here are some of my thoughts following this. Let me know what you think.
This is correct but with a caveat. Say data is partitioned on column "a", this is not necessarily the same for hash vs key partitioned. Imagine we hash partition with our hash function: "Hash(a) = a % 3" this would put the values 1, 4, 7 in the same partition. Now say we partition the same data but use key partitioning. Now, 1, 4, 7 will initially be put into separate groups and thus different partitions, then with the follow up work to merge groups to improve high cardinality (as described in above comments) these could merge into the same or different partitions. With the key partitioned approach it is important to note that we can initially separate by the key values then merge them based on size to improve cardinality without breaking the key partitioning. If we tried to merge hash or reshuffle based on size we would not truly be hash partitioned by our hash function. I do not know what the implications of this is throughout Datafusion but nevertheless this is not true hash partitioning. Another important difference between hash and key partition is there place of origin. Key partitioning can only originate from the data scan itself, making it explicit and a guarantee. Hash partitioning has an associated repartitioning operator which can be introduced any where in the plan, this makes hash an implicit guarantee. This is just another difference I wanted to point out about the partition styles.
Just a comment to ensure we are on the same page. This layout has the capability to achieve all of these partitioning types. If you key or hash partition on
Yes, spot on.
This is not true for the current implementation, but can be true as an option in follow up work. I noted that yes if our data is set up in a hierarchical format we could implicitly key partition by a superset of the data if it benefitted some parallelization, but decided to not implement this in the first PR as it would require some heuristic or additional user option (I don't know if this is too many knobs for the user to be turning). Due to the complexity and ambiguity in the implementation I decided against it. This is a good thing to point out though, that yes this would be another differentiating factor between hash and key partitioned.
This goes hand-in-hand with the last statement I made. Say your data is organized to actually be partitioned hierarchically by Using hash partitioning say you declare the data partitioned by Hash(customer_id, provider_id). The optimizer will think when it needs to do an aggregation on the group by clause Now, using key partitioning on the same columns: KeyPartitioned(customer_id, provider_id), and using some heuristic or options that determines if it is worth it to repartition by a superset of the passed in partitioning, the optimizer can recognize that with hierarchical organization, this means that all rows with the same
Yes, this is correct but taking into account my past statements about implementing the heuristic or option to partition by key partition super set when beneficial.
Yes this is the main idea, I am just seeing KeyPartitioned as the mechanism to do this behavior. The reason so is because although it holds similar properties to hash partitioning, they are based on fundamentally different concepts. My two biggest strifes with this are:
This approach could work but seems like we may be trying to stretch the functionality of hash partitioning too far, turning into something it was not designed to do. I do think that think that an implementation of checking to see if a plan would benefit from hash repartitioning on a superset would be able to achieve similar results but I think that having a different type of partitioning would be a clearer implementation. A rule like this would get pretty tricky as you would have to take into account all partitioning requirements throughout the plan when determining if you would rather partition by a superset hash. With the key partitioning approach the fact that you data is partitioned is this way is apparent from the data source and operators can more naturally decide what to do with this property. With this said I can see an optimization rule like this being beneficial regardless of if we decide to move forward with the key partitioned of hash partitioned approach. In another scenario say that earlier in the plan we are forces to repartition our data by also ccing @gabotechs in these to keep everyone in the loop. :) |
|
I am making a note here that is follow up work to this PR (listed in order of priority):
These issues are dependent on this PR so I will hold off on making these for now. Chime in if I missed anything or if some of this is seen as not needed. |
|
I am actively reviewing this PR |
|
Thanks again for the thorough response 🙇
I think this is a good summary, and it will be for maintainers to give a more formal answer on what they prefer seeing maintained. Both solutions have different trade offs, I agree that the alternative to In any case, in the end I think the plan simplifications you are achieving are worth the effort either way! 🚀 |
| // Here we do not check the partition count for hash partitioning and assumes the partition count | ||
| // and hash functions in the system are the same. In future if we plan to support storage partition-wise joins, | ||
| // then we need to have the partition count and hash functions validation. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looking at this comment I am also thinking KeyPartitioned addresses this concern, in the sense that KeyPartitioned(my_hash_func(key)) expresses that the set is partitioned according to a precise hash function (same for range partitioning). It removes the "unknown hash" from the equation.
So for a join, if the two sides are partitioned by KeyPartitioned(f(key)) with the same f, then we can use a partitioned HashJoin without repartition.
|
My main two concerns with this approach are:
I think more experience people in DataFusion should chime in and give their opinion, so in the mean time, one thing that comes to mind, is that we can ship first a benchmark or test that would benefit from this so that we can potentially compare different approaches. WDYT? In the mean time, some people that come to mind whose opinion could be useful is @crepererum as a core contributor to the repartitioning mechanism, and @adriangb as potential interested in a feature like this. |
|
Thanks @fmonjalet and @gabotechs for the great comments. Gene and I chatted, and he is currently working on:
He’ll be posting the details here soon. |
|
I will open a new PR with a modified solution after discussing with @gabotechs @fmonjalet and @NGA-TRAN in person. |
Full Report
Issue 18777 Parallelize Key Partitioned Data.pdf
Which issue does this PR close?
Rationale for this change
Optimize aggregations on Hive-partitioned tables by eliminating unnecessary repartitioning/coalescing when grouping by partition columns. This enables parallel computation of complete results without a merge bottleneck.
What changes are included in this PR?
KeyPartitionedAre these changes tested?
Benchmarking
For tpch it was unaffected as expected (not partitioned):
I create my own benchmark and saw these results:
These are not huge improvements as in memory hashing is pretty efficient but these are consistent gain (ran many times).
These improvements will be crucial for distributed datafusion as network shuffles are much less efficient than in memory repartitioning.
Are there any user-facing changes?
listing_table_preserve_partition_values