-
Notifications
You must be signed in to change notification settings - Fork 525
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Imbalance resource usage of compactor instances #4067
Comments
As the rate of block creation increases I would recommend lowering the following setting to allow more compactors to participate in reducing the length of the blocklist:
Lowering this value too much will prevent compactors from finding blocks to attempt to compact so perhaps try 30m and see the impact? |
This issue has been automatically marked as stale because it has not had any activity in the past 60 days. |
@joe-elliott are there any other measures one could try? Changing the compaction_window to 30min did not change the situation For us the situation is the same, most of the time there is an idle pod even though the active pods might go over the resources they have and get oomkilled |
What's the value of |
We've also experienced the same issue and I've noticed some log activity:
And then suddenly this instance's activity goes down to zero and never recovers. Our blocklist length is ~75K and outstanding blocks are steady around ~60K across 23 pods. |
Describe the bug
We setup 8 replicas of compactor, and observed that the resource usage of them is imbalance, some instances' cpu and memory usage decreased suddenly to a very low level:
Seems this instance didn't send request to backend from a time point:
The normal instance:
To Reproduce
Steps to reproduce the behavior:
Expected behavior
All compactor instances should have balanced resource usage
Environment:
Additional Context
The TempoCompactorsTooManyOutstandingBlocks is triggered, and
tempodb_compaction_outstanding_blocks
is increasing.The text was updated successfully, but these errors were encountered: