Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug] too many "send fragment timeout. backend id" error #8942

Closed
3 tasks done
morningman opened this issue Apr 9, 2022 · 1 comment · Fixed by #9720
Closed
3 tasks done

[Bug] too many "send fragment timeout. backend id" error #8942

morningman opened this issue Apr 9, 2022 · 1 comment · Fixed by #9720

Comments

@morningman
Copy link
Contributor

Search before asking

  • I had searched in the issues and found no similar issues.

Version

trunk

What's Wrong?

In some high concurrency scenarios, a large number of "send fragment timeout. backend id xxx" errors may appear in the fe.log. And all subsequent requests keep reporting this error and cannot be recovered.

What You Expected?

This is because in high load scenarios, the execution thread pool on the BE side is full, and new requests will enter the waiting queue of the thread pool.
However, the timeout period of the plan fragment request sent by FE to BE is only 5 seconds. The waiting time for requests entering the waiting queue may be very long, resulting in a large number of rpc errors send fragment timeout.

And subsequent requests will continue to enter the waiting queue, causing all subsequent requests to time out.

How to Reproduce?

Run a complex queries with high concurrency.

Anything Else?

No response

Are you willing to submit PR?

  • Yes I am willing to submit a PR!

Code of Conduct

@ChPi
Copy link
Contributor

ChPi commented May 12, 2022

version 0.15。
modified remote_fragment_exec_timeout_ms=600000, this problem still exists

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants