Scaling / Configuration recommendations for compactor #3369
Replies: 2 comments 3 replies
-
Hi, I have some suggestions:
What does your longer-term graph (few days) of |
Beta Was this translation helpful? Give feedback.
-
Just so others could find it when running into this For us the issue was that the default limits were too low, and compactor blocks/ tenant for our biggest tenancy just kept growing and growing. Our biggest tenant had ~350mil active series. Essentially, setting |
Beta Was this translation helpful? Give feedback.
-
I need some advice on proper configuration / scaling for the compactor in our environment.
Current setup
Non-default compactor config (note, I was guessing at most of these):
- compaction_concurrency: 4
- compactor_split_groups: 2
- compactor_split_and_merge_shards: 4
This config seemed to be working okay at the beginning, but now I'm seeing compactor-0 showing 8.75 hours since last successful, and compactor-1 showing 3.5 hours. That seems like a lot.
Compactor Resources dashboard shows both instances using lots of CPU and RAM (but not bottlenecking on either), and 1-200mb/sec of disk read/write activity. On gp2 that's pretty close to the max throughput, but it's not looking like it's bottlenecked from the graphs.
Clearly, there's something I could do here to improve the configuration and/or scaling of my compactors. I would be greatful for some suggested instance size / count based on our environment listed above. I have EKS resources to throw at the problem, but I don't want to do that if there's a config option that's going to magically fix everything.
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions