You are viewing a plain text version of this content. The canonical link for it is here.
Posted to github@arrow.apache.org by GitBox <gi...@apache.org> on 2021/05/11 20:33:08 UTC

[GitHub] [arrow-datafusion] Dandandan commented on issue #27: Implement hash-partitioned hash aggregate

Dandandan commented on issue #27:
URL: https://github.com/apache/arrow-datafusion/issues/27#issuecomment-839140660


   I also realize I may have implemented this in an inefficient way. Looking at Spark it does first a partial aggregate followed by a hash repartition followed by another aggregation. I think this will not slow down the low cardinality query that much (as the result set will be small already) while being fast - maybe even faster - for high cardinality aggregates.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org