Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

group by / create batch of max x rows function #46

Open
chris-twiner opened this issue Aug 14, 2023 · 1 comment
Open

group by / create batch of max x rows function #46

chris-twiner opened this issue Aug 14, 2023 · 1 comment

Comments

@chris-twiner
Copy link
Contributor

per pyspark group and batch of x rows and another I can't find (probably deleted) that wanted api calls bucketed it seems there is a gap for partition id with a maxed counter, so stateful with an increment every x rows, and reset count on new partition.

This would allow chunking but ideally the chunks would be mappable, so perhaps combined with a collect_set and a custom udf.

@chris-twiner
Copy link
Contributor Author

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant