-
Notifications
You must be signed in to change notification settings - Fork 25k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Duplicate Watcher events firing #38482
Comments
Pinging @elastic/es-core-features |
can you stop/start watcher and see if the problem goes away? There is a state issue hidden in watcher when shards are relocating. Can you to the latest version and see if that happens again? For more info, see #33167 |
Thanks for the quick reply @spinscale! What do you mean by stopping/starting watcher? Is there a way to do that without restarting nodes/disabling X-Pack? And thanks for the linked issue, we'll try updating up to the latest v6 version and report back if that fixes anything! |
yes there is. Use the dedicated APIs mentioned at https://www.elastic.co/guide/en/elasticsearch/reference/6.5/watcher-api-stop.html and https://www.elastic.co/guide/en/elasticsearch/reference/6.5/watcher-api-start.html |
As an update, things are looking good after stopping/starting watcher! No duplicate events have fired since then. |
this still means you should upgrade to the latest release as this can pop up anytime again shards of the |
Elasticsearch version (
bin/elasticsearch --version
):v6.3.1
Plugins installed:
x-pack, discovery-ec2, repository-s3
JVM version (
java -version
):OS version (
uname -a
if on a Unix-like system):Description of the problem including expected versus actual behavior:
Occasionally, our configured watchers will fire twice, resulting in duplicate notifications. We have Slack as an output and we'll see two messages posted in the channel. But when we go and check the actual index, only one copy of the documents exist. From a recommendation I saw in the discussion forum, I also outputted the watcher ID in the message to see if there is another watcher that is triggering it that we aren't aware of, but every duplicate has the same watcher ID.
I have a feeling it might be because our cluster is a bit overloaded at times. A minute or so before the duplicate event fires, we'll see some thread pool rejections. However, despite being overloaded, it is working just fine for our use-cases, so it would be ideal if we didn't have to add more power to the cluster because of this. It would also be helpful to know why this would cause a problem for the watcher.
Steps to reproduce:
Unfortunately, since this is non-deterministic, I don't know of a way to reliably reproduce the issue.
Provide logs (if relevant):
A log line of the thread pool rejection:
The text was updated successfully, but these errors were encountered: