Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Beyla cache: prevent OOMing during start in big clusters #1354

Open
mariomac opened this issue Nov 13, 2024 · 1 comment
Open

Beyla cache: prevent OOMing during start in big clusters #1354

mariomac opened this issue Nov 13, 2024 · 1 comment
Labels
documentation Improvements or additions to documentation k8s-cache roadmap

Comments

@mariomac
Copy link
Contributor

mariomac commented Nov 13, 2024

In big clusters (800 nodes), when the cache pod starts, it receives a huge amount of K8s metadata, and it is enqueued in the memory faster than the Beyla cache service it is really able to forward to the Beyla client instances.

This causes that, when a beyla-k8s-cache pod starts, it accumulates GBs of memory until it is OOMed, then connected Beyla clients might move to another instances, and when the pod is restarted, it is idle enough to process all the information faster than it accumulates in main memory.

We need to find a way to desynchronize the informers receive-transform-store thread from the client message submission.

Image

Image

@mariomac
Copy link
Contributor Author

This has been fixed by properly setting GOMEMLIMIT. We just need to document this.

@marctc marctc added the documentation Improvements or additions to documentation label Nov 26, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation k8s-cache roadmap
Projects
None yet
Development

No branches or pull requests

2 participants