-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reduce duplication between BoundedAggregateStream
and GroupedHashAggregateStream
#6798
Comments
I wonder if you have any thoughts on this proposal @metesynnada and @mustafasrepo ? |
During streaming support for aggregation I tried to reduce duplication between However, I think to benefit both implementations with the following major change. It is better to reduce duplication as much as possible even if it introduces minor overhead during benchmarks. |
Thanks for the comments @mustafasrepo -- I have been studying the code and I am confident I can reduce the duplication. I hope to have a PR ready for review that brings us in this direction in a few days. |
I plan to do this |
I have completed my study of the code -- and I added some comments here #6881 Rather than try and consolidate the BoundedAggregateStream on |
I have a PR ready that consolidates the logic which I am quite pleased with: #6932 |
Is your feature request related to a problem or challenge?
We are trying to make hash based aggregation significantly faster -- see #4973
This will require some non trivial changes to the organization of how hash aggregation works. At the moment
BoundedAggregateStream
andGroupedHashAggregateStream
both share significant amounts of code and so either we will have to duplicate the work to make hashing aggregation faster or elseBoundedAggregateStream
will not get the benefits.Here is a visual depiction of the common code:
Describe the solution you'd like
Reduce duplication between
BoundedAggregateStream
andGroupedHashAggregateStream
The major differences are:
Describe alternatives you've considered
No response
Additional context
No response
The text was updated successfully, but these errors were encountered: