Hi
We have this alert on a regular base (several times a day) and it's getting annoying.
Our relevant settings:
We have adapted the trigger to average on 40 minutes:
Our Housekeping settings were adapted like this:
A normal log extract would be something like this:
The server is running on an SSD RAID 10 with two SSD's. We are fully aware that these delete statements are very I/O expensive and would set up database partitioning for the next update. That will though not be right now.
My questions are:
We have this alert on a regular base (several times a day) and it's getting annoying.
Our relevant settings:
Code:
# HousekeepingFrequency=1 # MaxHousekeeperDelete=5000
Code:
{Template App Zabbix Proxy:zabbix[process,housekeeper,avg,busy].avg(40m)}>75
- History - Data storage period: 14d
- Trends - Data storage period: 365d
- Events and alerts - Trigger data storage period 365d (all others are set to 1d)
A normal log extract would be something like this:
Code:
housekeeper [deleted 831062 hist/trends, 0 items/triggers, 6 events, 1 problems, 0 sessions, 0 alarms, 0 audit items in 2294.853656 sec, idle for 1 hour(s)]
My questions are:
- what could we do better in order to keep the housekeeper below the threshold?
- why is it only our four zabbix proxys reporting this alarm, but when I check on the zabbix server itself with the graph "Zabbix internal process busy %" it looks like the server itself is sleeping? Is the cleanup not executed on the zabbix server itself?

Comment