If you’ve seen OOM killer events, that’s definitely what’s happening, and it’s disastrous. You can’t run a production server that randomly kills processes.
This should be plenty. You must have something leaking memory, or it’s a very heavily loaded server with tons of services.
If it is not (e.g. it’s physical hardware or VM running under KVM or Xen), it isn’t possible for it to be overly aggressive. It only kills processes when it’s literally out of memory. At that point it has no choice. Something literally must be killed.
OOMScoreAdjust should never come into play. You need to fix it so processes aren’t being randomly killed. But, if you can’t do that, I guess you can make your least important processes first on the chopping block.
But, figure out what’s taking up your memory and fix it. Running out of memory is a disaster to be avoided at any cost. You must reduce your usage or increase available memory (or choose a host that doesn’t use OpenVZ or Virtuozzo and oversell memory, which causes random memory errors no matter how much memory it claims you have).
OK, they use KVM or Xen, so it’s not oversold. So, yes, you need to reduce memory. You cannot run production services on a system that’s having OOM killer events. Nothing me or Webmin can do about that.
The only right number of OOM killer events is zero.
Mar 9 00:00:02 vps1 systemd[1]: lfd.service: Main process exited, code=killed, status=9/KILL
Mar 9 01:36:30 vps1 systemd[1]: packagekit.service: Main process exited, code=killed, status=15/TERM
Mar 9 03:46:08 vps1 systemd[1]: packagekit.service: Main process exited, code=killed, status=15/TERM
Mar 9 05:00:45 vps1 systemd[1]: lfd.service: Main process exited, code=killed, status=9/KILL
Mar 9 06:51:50 vps1 spamd[15781]: spamd: child [16087] killed successfully: interrupted, signal 2 (0002)
Mar 9 06:51:50 vps1 spamd[15781]: spamd: child [16088] killed successfully: interrupted, signal 2 (0002)
Mar 9 07:00:10 vps1 systemd[1]: lfd.service: Main process exited, code=killed, status=9/KILL
Mar 9 07:37:42 vps1 systemd[1]: packagekit.service: Main process exited, code=killed, status=15/TERM
Mar 9 11:32:22 vps1 systemd[1]: lfd.service: Main process exited, code=killed, status=9/KILL
Mar 9 13:40:50 vps1 systemd[1]: packagekit.service: Main process exited, code=killed, status=15/TERM
Mar 9 14:52:16 vps1 systemd[1]: lfd.service: Main process exited, code=killed, status=9/KILL
Mar 9 15:27:19 vps1 systemd[1]: lfd.service: Main process exited, code=killed, status=9/KILL
Mar 9 19:35:54 vps1 usermin[32541]: /etc/usermin/stop: 4: kill: No such process
Mar 9 19:41:20 vps1 systemd[1]: packagekit.service: Main process exited, code=killed, status=15/TERM
Yesterday is even worse…
*Edit
It should state invoked oom-killer though, which it isn’t.
OK, then it’s some other problem. I don’t have good guesses about what. It seems like it has to have been Webmin being killed while it was writing the file (or it ran out of space), and if it’s not the OOM killer, I don’t know what else it’d be.