-
Notifications
You must be signed in to change notification settings - Fork 31
Open
Labels
pr-size: mediumModerate update sizeModerate update sizetriage doneIssues that are triaged by dev team and are in investigation.Issues that are triaged by dev team and are in investigation.under development
Description
Is your feature request related to a problem? Please describe.
Python lists / arrays are pretty slow and memory-intense, therefore many libraries allow reading things as Arrow Data. Examples are fastexcel or arrow-odbc. Many Data libraries support Arrow out of the box, like DuckDB/Polars/Spark/Pandas/DeltaLake. Would also be a great source for Bulk Import.
Describe the solution you'd like
Best would be an API similar to duckdb which has methods for Arrow Tables and RecordBatchReader: https://duckdb.org/docs/stable/guides/python/export_arrow.html#export-as-a-recordbatchreader This can also be done without a Dependency on PyArrow, since arrow supports PyCapsules (eg, see arro3 )
Describe alternatives you've considered
Well, I can also live with not-so fast queries :)
keen85, zzmark, harryatgem, agrueneberg, dhirschfeld and 5 more
Metadata
Metadata
Assignees
Labels
pr-size: mediumModerate update sizeModerate update sizetriage doneIssues that are triaged by dev team and are in investigation.Issues that are triaged by dev team and are in investigation.under development