Re: Strange ARC/Swap/CPU on yesterday's -CURRENT

From: Michael Gmelin <freebsd_at_grem.de>
Date: Sun, 11 Mar 2018 01:40:58 +0100
> On 11. Mar 2018, at 00:47, O. Hartmann <ohartmann_at_walstatt.org> wrote:
> 
> Am Wed, 7 Mar 2018 14:39:13 +0400
> Roman Bogorodskiy <novel_at_FreeBSD.org> schrieb:
> 
>>  Danilo G. Baio wrote:
>> 
>>>> On Tue, Mar 06, 2018 at 01:36:45PM -0600, Larry Rosenman wrote:  
>>>> On Tue, Mar 06, 2018 at 10:16:36AM -0800, Rodney W. Grimes wrote:  
>>>>>> On Tue, Mar 06, 2018 at 08:40:10AM -0800, Rodney W. Grimes wrote:  
>>>>>>>> On Mon, 5 Mar 2018 14:39-0600, Larry Rosenman wrote:
>>>>>>>> 
>>>>>>>>> Upgraded to:
>>>>>>>>> 
>>>>>>>>> FreeBSD borg.lerctr.org 12.0-CURRENT FreeBSD 12.0-CURRENT #11 r330385:
>>>>>>>>> Sun Mar  4 12:48:52 CST 2018
>>>>>>>>> root_at_borg.lerctr.org:/usr/obj/usr/src/amd64.amd64/sys/VT-LER  amd64
>>>>>>>>> +1200060 1200060
>>>>>>>>> 
>>>>>>>>> Yesterday, and I'm seeing really strange slowness, ARC use, and SWAP use
>>>>>>>>> and swapping.
>>>>>>>>> 
>>>>>>>>> See http://www.lerctr.org/~ler/FreeBSD/Swapuse.png  
>>>>>>>> 
>>>>>>>> I see these symptoms on stable/11. One of my servers has 32 GiB of 
>>>>>>>> RAM. After a reboot all is well. ARC starts to fill up, and I still 
>>>>>>>> have more than half of the memory available for user processes.
>>>>>>>> 
>>>>>>>> After running the periodic jobs at night, the amount of wired memory 
>>>>>>>> goes sky high. /etc/periodic/weekly/310.locate is a particular nasty 
>>>>>>>> one.  
>>>>>>> 
>>>>>>> I would like to find out if this is the same person I have
>>>>>>> reporting this problem from another source, or if this is
>>>>>>> a confirmation of a bug I was helping someone else with.
>>>>>>> 
>>>>>>> Have you been in contact with Michael Dexter about this
>>>>>>> issue, or any other forum/mailing list/etc?    
>>>>>> Just IRC/Slack, with no response.  
>>>>>>> 
>>>>>>> If not then we have at least 2 reports of this unbound
>>>>>>> wired memory growth, if so hopefully someone here can
>>>>>>> take you further in the debug than we have been able
>>>>>>> to get.  
>>>>>> What can I provide?  The system is still in this state as the full backup is
>>>>>> slow.  
>>>>> 
>>>>> One place to look is to see if this is the recently fixed:
>>>>> https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=222288
>>>>> g_bio leak.
>>>>> 
>>>>> vmstat -z | egrep 'ITEM|g_bio|UMA'
>>>>> 
>>>>> would be a good first look
>>>>> 
>>>> borg.lerctr.org /home/ler $ vmstat -z | egrep 'ITEM|g_bio|UMA'
>>>> ITEM                   SIZE  LIMIT     USED     FREE      REQ FAIL SLEEP
>>>> UMA Kegs:               280,      0,     346,       5,     560,   0,   0
>>>> UMA Zones:             1928,      0,     363,       1,     577,   0,   0
>>>> UMA Slabs:              112,      0,25384098,  977762,102033225,   0,   0
>>>> UMA Hash:               256,      0,      59,      16,     105,   0,   0
>>>> g_bio:                  384,      0,      33,    1627,542482056,   0,   0
>>>> borg.lerctr.org /home/ler $  
>>>>>>>> Limiting the ARC to, say, 16 GiB, has no effect of the high amount of 
>>>>>>>> wired memory. After a few more days, the kernel consumes virtually all 
>>>>>>>> memory, forcing processes in and out of the swap device.  
>>>>>>> 
>>>>>>> Our experience as well.
>>>>>>> 
>>>>>>> ...
>>>>>>> 
>>>>>>> Thanks,
>>>>>>> Rod Grimes
>>>>>>> rgrimes_at_freebsd.org  
>>>>>> Larry Rosenman                     http://www.lerctr.org/~ler  
>>>>> 
>>>>> -- 
>>>>> Rod Grimes                                                 rgrimes_at_freebsd.org  
>>>> 
>>>> -- 
>>>> Larry Rosenman                     http://www.lerctr.org/~ler
>>>> Phone: +1 214-642-9640                 E-Mail: ler_at_lerctr.org
>>>> US Mail: 5708 Sabbia Drive, Round Rock, TX 78665-2106  
>>> 
>>> 
>>> Hi.
>>> 
>>> I noticed this behavior as well and changed vfs.zfs.arc_max for a smaller size.
>>> 
>>> For me it started when I upgraded to 1200058, in this box I'm only using
>>> poudriere for building tests.  
>> 
>> I've noticed that as well.
>> 
>> I have 16G of RAM and two disks, the first one is UFS with the system
>> installation and the second one is ZFS which I use to store media and
>> data files and for poudreire.
>> 
>> I don't recall the exact date, but it started fairly recently. System would
>> swap like crazy to a point when I cannot even ssh to it, and can hardly
>> login through tty: it might take 10-15 minutes to see a command typed in
>> the shell.
>> 
>> I've updated loader.conf to have the following:
>> 
>> vfs.zfs.arc_max="4G"
>> vfs.zfs.prefetch_disable="1"
>> 
>> It fixed the problem, but introduced a new one. When I'm building stuff
>> with poudriere with ccache enabled, it takes hours to build even small
>> projects like curl or gnutls.
>> 
>> For example, current build:
>> 
>> [10i386-default] [2018-03-07_07h44m45s] [parallel_build:] Queued: 3  Built: 1  Failed:
>> 0  Skipped: 0  Ignored: 0  Tobuild: 2   Time: 06:48:35 [02]: security/gnutls
>> | gnutls-3.5.18             build           (06:47:51)
>> 
>> Almost 7 hours already and still going!
>> 
>> gstat output looks like this:
>> 
>> dT: 1.002s  w: 1.000s
>> L(q)  ops/s    r/s   kBps   ms/r    w/s   kBps   ms/w   %busy Name
>>    0      0      0      0    0.0      0      0    0.0    0.0  da0
>>    0      1      0      0    0.0      1    128    0.7    0.1  ada0
>>    1    106    106    439   64.6      0      0    0.0   98.8  ada1
>>    0      1      0      0    0.0      1    128    0.7    0.1  ada0s1
>>    0      0      0      0    0.0      0      0    0.0    0.0  ada0s1a
>>    0      0      0      0    0.0      0      0    0.0    0.0  ada0s1b
>>    0      1      0      0    0.0      1    128    0.7    0.1  ada0s1d
>> 
>> ada0 here is UFS driver, and ada1 is ZFS.
>> 
>>> Regards.
>>> -- 
>>> Danilo G. Baio (dbaio)  
>> 
>> 
>> 
>> Roman Bogorodskiy
> 
> 
> This is from a APU, no ZFS, UFS on a small mSATA device, the APU (PCenigine) works as a
> firewall, router, PBX):
> 
> last pid:  9665;  load averages:  0.13,  0.13,  0.11
> up 3+06:53:55  00:26:26 19 processes:  1 running, 18 sleeping CPU:  0.3% user,  0.0%
> nice,  0.2% system,  0.0% interrupt, 99.5% idle Mem: 27M Active, 6200K Inact, 83M
> Laundry, 185M Wired, 128K Buf, 675M Free Swap: 7808M Total, 2856K Used, 7805M Free
> [...]
> 
> The APU is running CURRENT ( FreeBSD 12.0-CURRENT #42 r330608: Wed Mar  7 16:55:59 CET
> 2018 amd64). Usually, the APU never(!) uses swap, now it is starting to swap like hell
> for a couple of days and I have to reboot it failty often.
> 
> Another box, 16 GB RAM, ZFS, poudriere, the packaging box, is right now unresponsible:
> after hours of building packages, I tried to copy the repository from one location on
> the same ZFS volume to another - usually this task takes a couple of minutes for ~ 2200
> ports. Now, I has taken 2 1/2 hours and the box got stuck, Ctrl-T  on the console
> delivers:
> load: 0.00  cmd: make 91199 [pfault] 7239.56r 0.03u 0.04s 0% 740k
> 
> No response from the box anymore.
> 
> 
> The problem of swapping like hell and performing slow isn't an issue of the past days, it
> is present at least since 1 1/2 weeks for now, even more. Since I build ports fairly
> often, time taken on that specific box has increased from 2 to 3 days for all ~2200
> ports. The system has 16 GB of RAM, IvyBridge 4-core XEON at 3,4 GHz, if this information
> matters. The box is consuming swap really fast.
> 
> Today is the first time the machine got inresponsible (no ssh, no console login so far).
> Need to coldstart. OS is CURRENT as well.
> 

Any chance this is related to meltdown/spectre mitigation patches?

Best,
Michael



> Regards,
> 
> O. Hartmann
> 
> 
> -- 
> O. Hartmann
> 
> Ich widerspreche der Nutzung oder Übermittlung meiner Daten für
> Werbezwecke oder für die Markt- oder Meinungsforschung (§ 28 Abs. 4 BDSG).
Received on Sat Mar 10 2018 - 23:47:43 UTC

This archive was generated by hypermail 2.4.0 : Wed May 19 2021 - 11:41:15 UTC