I updated my main amd64 workstation to r332158 from something much earlier (mid Jan). Upon reboot, all seemed well. However, I later realized that the vmm.ko module was not loaded at boot, because bhyve PCI passthru did not work. My loader.conf looks like (I'm passing a USB interface through): ####### vmm_load="YES" opensolaris_load="YES" zfs_load="YES" nvidia_load="YES" nvidia-modeset_load="YES" # Tune ZFS Arc Size - Change to adjust memory used for disk cache vfs.zfs.arc_max="4096M" hint.xhci.2.disabled="1" pptdevs="8/0/0" hw.dmar.enable="0" cuse_load="YES" ####### The problem seems "random". I rebooted into single-user to see if somehow, vmm.ko was loaded at boot and something was unloading vmm.ko. However, on this boot it was loaded. I then ^D'ed and continued to multi-user, where X failed to start because this time, the nvidia modules were not loaded. (but nvidia had been loaded on the 1st boot). So it *seems* like different modules are randomly not loaded by the loader, at boot. The ZFS config is: config: NAME STATE READ WRITE CKSUM tank ONLINE 0 0 0 mirror-0 ONLINE 0 0 0 ada0p2 ONLINE 0 0 0 da3p2 ONLINE 0 0 0 mirror-1 ONLINE 0 0 0 ada1p2 ONLINE 0 0 0 da0p2 ONLINE 0 0 0 cache da2s1d ONLINE 0 0 0 The data drives in the pool are all exactly like this: => 34 9767541101 ada0 GPT (4.5T) 34 6 - free - (3.0K) 40 204800 1 efi (100M) 204840 9763209216 2 freebsd-zfs (4.5T) 9763414056 4096000 3 freebsd-swap (2.0G) 9767510056 31079 - free - (15M) There is about 1.44T used in the pool. I have no idea how ZFS mirrors work, but I'm wondering if somehow this is a 2T problem, and there are issues with blocks on difference sides of the mirror being across the 2T boundary. Sorry to be so vague.. but this is the one machine I *don't* have a serial console on, so I don't have good logs. DrewReceived on Mon Apr 09 2018 - 21:17:48 UTC
This archive was generated by hypermail 2.4.0 : Wed May 19 2021 - 11:41:15 UTC