• The Horde webmail has been deprecated. Its complete removal is scheduled for April 2025. For details and recommended actions, see the Feature and Deprecation Plan.
  • We’re working on enhancing the Monitoring feature in Plesk, and we could really use your expertise! If you’re open to sharing your experiences with server and website monitoring or providing feedback, we’d love to have a one-hour online meeting with you.

Resolved Alert "average(cpu usage average) is higher 75% for 10 minutes" triggered by backup job

carini

Basic Pleskian
Server operating system version
Ubuntu 20.04 x86_64
Plesk version and microupdate number
Plesk Obsidian 18.0.44.2
Wen Plesk build a full backup on S3-Compatible the monitoring send an alert (and a SMS is consumed).

Is possible to take account of backup processes (or any other processes that runs on high niceness) and spew the alarm messages only if the backup runs more than <x> hours?

Screenshot 2022-06-19 at 00.29.53.png
 
Hi,

What are your backup settings? You can lower the priority of the backup processes so they won't reach 100% CPU time.
 
Backup process uses two threads and default priority for CPU (10) and I/O (7)

As shown on the graph the CPU usage is almost entirely due to nice processes.
 
We had the same issue and ended up using these settings:

Run scheduled backup processes with low priority:
Priority: 19
IOPriority: 7

Run all backup processes with low priority:
Priority: 19
IOPriority: 7

Compression level: No compression

The compression level was the biggest cause of the high CPU usage. As soon as we lowered it to less or no compression, the CPU graphs never reached 50% again.
 
Thanks @maartenv .

Two question:

  1. How much space do you waste by using no compression?
  2. Old backups remains recoverable if you change compression level?

Thanks in advance.
 
The size of "no compression" backups is not as much as I expected, but you should just try it. If you are using the "fast" level, try "fastest" and see if that solves the CPU problem. If the size of the backups is acceptable, try "no compression".

It also helps to use an incremental backup scheme like weekly, which gives a full backup once a week and incremental backups on the other days. It works perfectly fine if you want to recover files from the backup.

The old backups are still recoverable if you change the level of compression.
 
The compression level was the biggest cause of the high CPU usage. As soon as we lowered it to less or no compression, the CPU graphs never reached 50% again.
That's because the compressor is the only part that can actually use multithreading.
Also I don't see how this is a problem. Why have a powerful CPU if you only use a few % of it?
 
That's because the compressor is the only part that can actually use multithreading.
Also I don't see how this is a problem. Why have a powerful CPU if you only use a few % of it?
The only problem in using 100% CPU for a long time is the triggering of unnecessary notifications from platform360.io
 
The size of "no compression" backups is not as much as I expected, but you should just try it. If you are using the "fast" level, try "fastest" and see if that solves the CPU problem. If the size of the backups is acceptable, try "no compression".

It also helps to use an incremental backup scheme like weekly, which gives a full backup once a week and incremental backups on the other days. It works perfectly fine if you want to recover files from the backup.

The old backups are still recoverable if you change the level of compression.

I tried to set up with pri = 19 and compression normal, it seems to work without loosing out too much space plese mark as solved
 
Back
Top