Scrubber processes getting killed by OOM killer
On the production server, some scrubber processes are killed by the OOM killer
Jul 10 18:37:59 scrubber1 kernel: Out of memory: Kill process 3192 (swh) score 116 or sacrifice child Jul 10 18:37:59 scrubber1 kernel: Killed process 3192 (swh) total-vm:297888kB, anon-rss:229668kB, file-rss:24kB, shmem-rss:0kB Jul 10 18:37:59 scrubber1 kernel: oom_reaper: reaped process 3192 (swh), now anon-rss:0kB, file-rss:0kB, shmem-rss:0kB Jul 10 18:37:59 scrubber1 swh: INFO:swh.scrubber.storage_checker:Processing revision range bd30e8 to bd30e9 Jul 10 18:37:59 scrubber1 swh: INFO:swh.scrubber.storage_checker:Processing revision range 48b4e6 to 48b4e7 Jul 10 18:37:59 scrubber1 systemd: email@example.com: Main process exited, code=killed, status=9/KILL
free -h total used free shared buff/cache available Mem: 978Mi 756Mi 109Mi 0.0Ki 112Mi 47Mi Swap: 975Mi 962Mi 13Mi
Despite the ballooning allowed to 4g, it seems the server remains to 1Go.
Migrated from T4387 (view on Phabricator)