How-to adjust Checkmk performance
The following article explains how to monitor and adjust the performance in Checkmk.
LAST TESTED ON CHECKMK 2.2.0P1
First, there is no difference in the requirements for the virtual and hardware appliances.
The needed resources are based on the number of services, active checks, and types of hosts. If you have a lot of SNMP hosts, you'll need more CPU cores for executing the SNMP walks on those.
The only overview regarding needed resources we have is, and it is just a rough approximation: https://checkmk.com/product/appliances
We always recommend customers to orientate on the specifications for the HW Appliance.
When importing the virtual appliance, we have some default values preconfigured. Please check out this page: Installation of the virtual appliance
As this is a virtual machine, you can adjust these values anytime.
Configuration of Fetcher/Checker settings
Hands-On
Required services to monitor
To configure the right resources, we recommend checking the following graphs:
- PDF report with graphs of
- CPU
- Memory
- OMD <SITENAME> Performance
- activate the "Core statistics" snap in
- Check_MK
- Disk I/O Summary
- The local structure
- find -L ~/local > local.txt (as site user)
Let's give you an example:
With Core Statistics snap-in, you can check the load of the fetcher and helper. At 70%, we recommend increasing these values in the Global Settings. The CPU load and memory consumption will grow while you increase these values.
That's why we also recommend checking these graphs:
You will find more information about the fetcher and checker architecture here:
https://checkmk.com/blog/checkmk-2-0-cmc
https://checkmk.com/werk/11500
Important information about the Checkers: The checkers should not exceed your CPU core count!
Adjust the helper settings
If you decide to adjust the helper settings, please be aware of these settings:
Setup → General → Global Settings → Monitoring Core →
- Maximum concurrent active checks
- The usage should stay under 80% on average.
Maximum concurrent Checkmk fetchers
- With increasing the number of fetchers, your RAM usage will rise, so make sure to adjust this setting carefully and keep an eye on the memory consumption of your server.
- The usage should stay under 80% on average.
Maximum concurrent Checkmk checkers
- The number of checkers should not be higher than your CPU core count! If you have more than two cores, the general rule is:
Maximum checkers = number of cores - 1
. - The usage should stay under 80% on average.
- Maximum concurrent Livestatus connections
- In a distributed monitoring setup, having different values for the remote sites may be helpful. You will find the guidance on how to do that here!
Check the Livestatus Performance
If you face issues like this:
Please see this manual to check the Livestatus Performance
Required log files
Please see this manual to enable debug log of the helpers. The required settings are:
- Core
- Debugging of Checkmk helpers
High Fetcher Usage Although the fetcher helper count is already high
Also, please check out our article on Troubleshooting high CPU usage of the Checkmk micro core (cmc)
If you face the following problems:
- Fetcher helper usage is permanently above 96%, and fetcher count is already high (i.e., >50 or 100 or more) and
- the service "Check_MK" runs into constant "CRIT with fetcher timeouts
You can also use this command as site user to narrow down and find slow-running active checks.
lq "GET services\nColumns: execution_time host_name display_name" | awk -F';' '{ printf("%.2f %s %s\n", $1, $2, $3)}' | sort -rn | head
This can have several reasons:
- Firewalls are dropping traffic from Checkmk to the monitored systems. If the packets are dropped rather than blocked, Checkmk must wait for a timeout instead of instantly terminating the fetching process.
- You might have too many DOWN hosts, which are still being checked. Checkmk still tries to query those hosts, and the fetchers need to wait for a timeout every time. This can bind a lot of fetcher helpers, which are blocked for that time. Remove hosts which are in a DOWN state from your monitoring. Either permanently or by setting their Criticality to "Do not monitor this host".
- For classical operating systems (Linux/Windows/etc.), this indicates that you might have plugins/local checks with quite a long runtime. Increasing the number of fetchers further here is not constructive. Instead, you must identify the long-running plugins/local checks and set them to asynchronous execution and/or define (generous) cache settings or even timeouts, especially for them.
- For SNMP devices, you might have poorly performing SNMP devices. To troubleshoot those, have a look at this blog post.
Related articles