Re: spurious out of swap kills

From: Konstantin Belousov <kostikbel_at_gmail.com>
Date: Fri, 13 Sep 2019 08:53:32 +0300
On Thu, Sep 12, 2019 at 05:42:00PM -0700, Don Lewis wrote:
> On 12 Sep, Mark Johnston wrote:
> > On Thu, Sep 12, 2019 at 04:00:17PM -0700, Don Lewis wrote:
> >> My poudriere machine is running 13.0-CURRENT and gets updated to the
> >> latest version of -CURRENT periodically.  At least in the last week or
> >> so, I've been seeing occasional port build failures when building my
> >> default set of ports, and I finally had some time to do some
> >> investigation.
> >> 
> >> It's a 16-thread Ryzen machine, with 64 GB of RAM and 40 GB of swap.
> >> Poudriere is configured with
> >>   USE_TMPFS="wrkdir data localbase"
> >> and I have
> >>   .if ${.CURDIR:M*/www/chromium}
> >>   MAKE_JOBS_NUMBER=16
> >>   .else
> >>   MAKE_JOBS_NUMBER=7
> >>   .endif
> >> in /usr/local/etc/poudriere.d/make.conf, since this gives me the best
> >> overall build time for my set of ports.  This hits memory pretty hard,
> >> especially when chromium, firefox, libreoffice, and both versions of
> >> openoffice are all building at the same time.  During this time, the
> >> amount of space consumed by tmpfs for /wrkdir gets large when building
> >> these large ports.  There is not enough RAM to hold it all, so some of
> >> the older data spills over to swap.  Swap usage peaks at about 10 GB,
> >> leaving about 30 GB of free swap.  Nevertheless, I see these errors,
> >> with rustc being the usual victim:
> >> 
> >> Sep 11 23:21:43 zipper kernel: pid 16581 (rustc), jid 43, uid 65534, was killed: out of swap space
> >> Sep 12 02:48:23 zipper kernel: pid 1209 (rustc), jid 62, uid 65534, was killed: out of swap space
> >> 
> >> Top shows the size of rustc being about 2 GB, so I doubt that it
> >> suddenly needs an additional 30 GB of swap.
> >> 
> >> I'm wondering if there might be a transient kmem shortage that is
> >> causing a malloc(..., M_NOWAIT) failure in the swap allocation path
> >> that is the cause of the problem.
> > 
> > Perhaps this is a consequence of r351114?  To confirm this, you might
> > try increasing the value of vm.pfault_oom_wait to a larger value, like
> > 20 or 30, and see if the OOM kills still occur.
> 
> I wonder if increasing vm.pfault_oom_attempts might also be a good idea.
If you are sure that you cannot exhaust your swap space, set
attempts to -1 to disable this mechanism.

Basically, page fault handler waits for vm.pfault_oom_wait *
vm.pfault_oom_attempts for a page allocation before killing the process.
Default is 30 secs, and if you cannot get a page for 30 secs, there is
something very wrong with the machine.
Received on Fri Sep 13 2019 - 03:53:40 UTC

This archive was generated by hypermail 2.4.0 : Wed May 19 2021 - 11:41:21 UTC