-
Notifications
You must be signed in to change notification settings - Fork 64
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
oomkill in aks-based openshift 4.10 cluster #984
Comments
If you can avoid it, it's best not to inform with a cache on Secrets and instead look them up as needed. ServiceAccount are probably still ok as each resource is very small. |
What are the goals for this issue? Is raising the requirements acceptable? If the goal is reducing our memory footprint, what are the parameters for that, considering we currently know we have a cache that grows linearly and unbounded with the number and size of objects? Are we looking at reducing the overhead per object, or are we looking to constrain memory consumption to a ceiling? The former is easier, but without the latter, operators will potentially always have to manage resource limits for Cartographer, which has always felt sub-par to me. Some activities that might help us understand where we could have impact reducing memory footprint:
|
Bug description:
In an Azure-based openshift 4.10 cluster we observe that Cartographer's default
limits are not enough in the presence of 15k+ secrets and thousands of
serviceaccounts (carto v0.4.3).
Steps to reproduce:
TODO (/cc todd)
Expected behavior:
Have cartographer up and running despite the great number of other objects
unrelated to the necessary functionality of cartographer.
Actual behavior:
OOMKill
Logs, k8s object dumps:
TODO
Versions:
TODO
Deployment info:
TODO
Additional context:
The text was updated successfully, but these errors were encountered: