You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have a dataset of ~1M rows that has a column of np.str_ in the metadata. It takes 4 seconds / 1000 records to load this column, while it takes seconds for 1,000,000 records for integer columns.
Steps to Reproduce
Create a dataset of 1,000,000 rows with a metadata of a mixture of strings and integers.
Expected/Desired Behavior
Strings should load approximately as fast as integers, otherwise are there other recommendations? Trying to understand the nature of the problem
Python Version
No response
OS
No response
IDE
No response
Packages
No response
Additional Context
No response
Possible Solution
No response
Are you willing to submit a PR?
I'm willing to submit a PR (Thank you!)
The text was updated successfully, but these errors were encountered:
Seems the issue is not generally reproducible and depends on the specific version of deeplake, python or numpy. Can you please provide more details about the versions you used. If there’s a reproducible script that’d be better.
Severity
P1 - Urgent, but non-breaking
Current Behavior
I have a dataset of ~1M rows that has a column of np.str_ in the metadata. It takes 4 seconds / 1000 records to load this column, while it takes seconds for 1,000,000 records for integer columns.
Steps to Reproduce
Create a dataset of 1,000,000 rows with a metadata of a mixture of strings and integers.
Expected/Desired Behavior
Strings should load approximately as fast as integers, otherwise are there other recommendations? Trying to understand the nature of the problem
Python Version
No response
OS
No response
IDE
No response
Packages
No response
Additional Context
No response
Possible Solution
No response
Are you willing to submit a PR?
The text was updated successfully, but these errors were encountered: