-
Notifications
You must be signed in to change notification settings - Fork 3.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
sql: fix cluster setting propagation flake take 2 #95583
Conversation
Previously we tried to fix this with one retry but that was insufficient. Extend it to all queries in this section of the test. Release note: None Epic: CRDB-20535
This test runs 10k times w/ stress w/o flaking now. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Re: can we ensure that all nodes in the cluster see the updated cluster setting value? Perhaps we could adjust the logic test to explicitly read the cluster setting from each node (with nodeidx
directive) and with a retry
option to wait until the setting is propagated, but we run this tests in 1- and 3-node configs, so we'd also need to skip some of those reads. In short, adding the retries seems like the easiest option.
Reviewed 1 of 1 files at r1, all commit messages.
Reviewable status: complete! 1 of 0 LGTMs obtained (waiting on @cucaroach)
So basically the read side is a problem, there's a rangefeed on everynode that updates the in memory settings cache. What I'd like to see is a primitive to wait for range feed invocations to be done or something, ie:
But I have no idea how that would work, presumably we'd read the latest timestamp from that table and then check that all the other nodes range feeds have advanced to that timestamp? There's some discussion here: Basically I agreed with Yahor retries are the easiest option but this feels like a problem that could come up again and could use a better solution. |
bors r+ |
Build succeeded: |
Previously we tried to fix this with one retry but that was
insufficient. Extend it to all queries in this section of the test.
Release note: None
Epic: CRDB-20535