-
Notifications
You must be signed in to change notification settings - Fork 234
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Slow speed #560
Comments
Issue-Label Bot is automatically applying the label Links: app homepage, dashboard and code for this bot. |
Hi, Definitely this a slow operation. How much time is taking? Maybe we could explorer how to speed this up? |
It took 2-3 hours approximately for an individual operation. |
Looking at the function https://github.com/ironmussa/Optimus/blob/master/optimus/functions.py#L398 we use fastnumbers library to speed things up when handling ints and floats. The function is super simple I am not sure if we have room for improvements here. @FavioVazquez can you take a look on this? |
@argenisleon @FavioVazquez - did you guys made some changes to address this issue in the new release? Just curious! |
Not at the moment. We are thinking about other approaches. Any thought about this? |
We could improve performance If we implement #322 |
Hi @IshanDindorkar, We rebuild the profiler to improve parallelism. In local mode, we cut the time by 1/3. |
Hello @argenisleon / @FavioVazquez -
I am running couple of operations like mentioned below -
df.rows.select(fbdt("id", "integer")).table()
df.rows.select(fbdt("id", "float")).table()
I am using notebook for performing these operations. As per console, the operation is in progress but it is slow in spitting out results. The dataset on which I am running operations has close to 90 million records. My question is there any configuration which I can use to speed up computation of spark job.
FYI, I am using an Ubuntu machine with 4 chores and 32 GB RAM.
Could you please advise.
Thank you for your support.
Best-
Ishan
The text was updated successfully, but these errors were encountered: