Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

otelconsumer: handle entity too large errors #41523

Merged
merged 2 commits into from
Nov 8, 2024

Conversation

mauri870
Copy link
Member

@mauri870 mauri870 commented Nov 5, 2024

Proposed commit message

The elasticsearchexporter does not handle a 413 Request Entity too Large from Elasticsearch, only forwarding the error back to the client.

When using the batcher config in the ES exporter, it runs synchronously, any error reported can be intercepted in the otelconsumer. When using the batch processor this happens asynchronously and there is no way to handle the error. If we can intercept an entity too large error, split the batch and retry.

See open-telemetry/opentelemetry-collector-contrib#36163

Checklist

  • My code follows the style guidelines of this project
  • I have commented my code, particularly in hard-to-understand areas
  • I have made corresponding changes to the documentation
  • I have made corresponding change to the default configuration files
  • I have added tests that prove my fix is effective or that my feature works
  • I have added an entry in CHANGELOG.next.asciidoc or CHANGELOG-developer.next.asciidoc.

Author's Checklist

  • [ ]

Related issues

The elasticsearchexporter does not handle a 413 Request Entity too Large from
Elasticsearch, only forwarding the error back to the client.

When using the batcher config in the ES exporter, it runs synchronously, any
error reported can be intercepted in the otelconsumer. When using the batch
processor this happens asynchronously and there is no way to handle the
error. If we can intercept an entity too large error, split the batch and
retry.
@mauri870 mauri870 added this to the otel milestone Nov 5, 2024
@mauri870 mauri870 self-assigned this Nov 5, 2024
@botelastic botelastic bot added the needs_team Indicates that the issue/PR needs a Team:* label label Nov 5, 2024
Copy link
Contributor

mergify bot commented Nov 5, 2024

This pull request does not have a backport label.
If this is a bug or security fix, could you label this PR @mauri870? 🙏.
For such, you'll need to label your PR with:

  • The upcoming major version of the Elastic Stack
  • The upcoming minor version of the Elastic Stack (if you're not pushing a breaking change)

To fixup this pull request, you need to add the backport labels for the needed
branches, such as:

  • backport-8./d is the label to automatically backport to the 8./d branch. /d is the digit

Copy link
Contributor

mergify bot commented Nov 5, 2024

backport-8.x has been added to help with the transition to the new branch 8.x.
If you don't need it please use backport-skip label and remove the backport-8.x label.

@mergify mergify bot added the backport-8.x Automated backport to the 8.x branch with mergify label Nov 5, 2024
@mauri870 mauri870 marked this pull request as ready for review November 5, 2024 15:26
@mauri870 mauri870 requested a review from a team as a code owner November 5, 2024 15:26
@pierrehilbert pierrehilbert added the Team:Elastic-Agent-Data-Plane Label for the Agent Data Plane team label Nov 5, 2024
@elasticmachine
Copy link
Collaborator

Pinging @elastic/elastic-agent-data-plane (Team:Elastic-Agent-Data-Plane)

@botelastic botelastic bot removed the needs_team Indicates that the issue/PR needs a Team:* label label Nov 5, 2024
@pierrehilbert pierrehilbert requested review from rdner and leehinman and removed request for khushijain21 and VihasMakwana November 5, 2024 16:09
Copy link
Contributor

@leehinman leehinman left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM. Thanks.

@mauri870 mauri870 merged commit 9c34e4e into elastic:main Nov 8, 2024
141 checks passed
mergify bot pushed a commit that referenced this pull request Nov 8, 2024
* otelconsumer: handle entity too large errors

The elasticsearchexporter does not handle a 413 Request Entity too Large from
Elasticsearch, only forwarding the error back to the client.

When using the batcher config in the ES exporter, it runs synchronously, any
error reported can be intercepted in the otelconsumer. When using the batch
processor this happens asynchronously and there is no way to handle the
error. If we can intercept an entity too large error, split the batch and
retry.

* use logp.TestingSetup

(cherry picked from commit 9c34e4e)
mauri870 added a commit that referenced this pull request Nov 8, 2024
* otelconsumer: handle entity too large errors

The elasticsearchexporter does not handle a 413 Request Entity too Large from
Elasticsearch, only forwarding the error back to the client.

When using the batcher config in the ES exporter, it runs synchronously, any
error reported can be intercepted in the otelconsumer. When using the batch
processor this happens asynchronously and there is no way to handle the
error. If we can intercept an entity too large error, split the batch and
retry.

* use logp.TestingSetup

(cherry picked from commit 9c34e4e)

Co-authored-by: Mauri de Souza Meneguzzo <[email protected]>
mauri870 added a commit to mauri870/beats that referenced this pull request Dec 5, 2024
This code was initially added in elastic#41523 because of a limitation from the
elasticsearch exporter.

The elasticsearch exporter has been updated to enforce flush::max_bytes
for the batcher extension and will automatically split the batch if it
exceeds the limit.

This error is now fixed in the collector v0.115.0.

See open-telemetry/opentelemetry-collector-contrib#36396.
mauri870 added a commit that referenced this pull request Dec 10, 2024
This code was initially added in #41523 because of a limitation from the
elasticsearch exporter.

The elasticsearch exporter has been updated to enforce flush::max_bytes
for the batcher extension and will automatically split the batch if it
exceeds the limit.

This error is now fixed in the collector v0.115.0.

See open-telemetry/opentelemetry-collector-contrib#36396.
mergify bot pushed a commit that referenced this pull request Dec 10, 2024
This code was initially added in #41523 because of a limitation from the
elasticsearch exporter.

The elasticsearch exporter has been updated to enforce flush::max_bytes
for the batcher extension and will automatically split the batch if it
exceeds the limit.

This error is now fixed in the collector v0.115.0.

See open-telemetry/opentelemetry-collector-contrib#36396.

(cherry picked from commit dbeb9cd)
mauri870 added a commit that referenced this pull request Dec 10, 2024
…41971)

This code was initially added in #41523 because of a limitation from the
elasticsearch exporter.

The elasticsearch exporter has been updated to enforce flush::max_bytes
for the batcher extension and will automatically split the batch if it
exceeds the limit.

This error is now fixed in the collector v0.115.0.

See open-telemetry/opentelemetry-collector-contrib#36396.

(cherry picked from commit dbeb9cd)

Co-authored-by: Mauri de Souza Meneguzzo <[email protected]>
michalpristas pushed a commit to michalpristas/beats that referenced this pull request Dec 13, 2024
…41911)

This code was initially added in elastic#41523 because of a limitation from the
elasticsearch exporter.

The elasticsearch exporter has been updated to enforce flush::max_bytes
for the batcher extension and will automatically split the batch if it
exceeds the limit.

This error is now fixed in the collector v0.115.0.

See open-telemetry/opentelemetry-collector-contrib#36396.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
backport-8.x Automated backport to the 8.x branch with mergify enhancement Team:Elastic-Agent-Data-Plane Label for the Agent Data Plane team
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants