-
Notifications
You must be signed in to change notification settings - Fork 3.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
schema: latency spike with new index backfill on tpc-c 10k (with partitioning) #36925
Comments
I dropped the index and set the cluster back up to run tpc-c. I confirmed that it passed again. Then, I ran tpc-c and killed one node: cc @bdarnell what do you think about this? The cluster is definitely worse with a schema change as opposed to with a dead node. |
So the schema change was running from about 20:20 to 21:10, and then from 21:30 to 22:00 the schema change was no longer running and you were just seeing the impact of the new index? This seems to show that the schema change isn't at fault, but the new schema is. You're adding a new unpartitioned, time-ordered index ( |
Ahh the old partitioning the index ux surprise. I want to run an experiment with the schema change done with no load and then running the load to see what happens. But the index is hung for some reason. Next up, if the index completes, I'll partition the index and run. But, I think the schema change would have to be run first and then partitioned after it's run. So it would have this large initial latency spike regardless of if partitioning solves the sustained latency problem. |
Debug.zip https://drive.google.com/open?id=1Ql9bhbQ-PrXKOvp9JMauBOoLXtJn1pao |
We have marked this issue as stale because it has been inactive for |
Describe the problem
I ran tpc-c 10k (with partitioning set) and passed with 97% efficiency:
Then I ran a schema change:
CREATE UNIQUE INDEX ON tpcc.order (o_entry_d, o_w_id, o_d_id, o_carrier_id, o_id);
I immediately saw a latency spike + QPS drop:
![image](https://user-images.githubusercontent.com/22278911/56325254-8bee6080-613f-11e9-8a65-fe85bdc3f444.png)
I tried running again after the schema change finished:
![image](https://user-images.githubusercontent.com/22278911/56325320-d079fc00-613f-11e9-8598-b73f389792d1.png)
To Reproduce
Environment:
v19.1.0-rc.3-11-gc65b71a
Debug.zip
https://drive.google.com/open?id=17ZYGOYU7Dpo7Yta6KhHTAf0nUFyRZSgg
Extract from the latency spike till now.
extract.txt
Epic CRDB-8816
Jira issue: CRDB-4469
The text was updated successfully, but these errors were encountered: