On Tue, Apr 17, 2018 at 10:03:55AM -0700, John Baldwin wrote: > On Tuesday, April 17, 2018 10:01:41 AM John Baldwin wrote: > > My laptop running recent head panicked this morning, apparently from hitting > > a key to stop the screensaver (at which point xscreensaver prompts for a > > password to unlock). > > (Sorry, buggy mail client sent this early) > > panic: Lock vm object not locked _at_ /usr/src/sys/vm/vm_page.c:4135 > > #4 0xffffffff805e4893 in panic (fmt=<unavailable>) > at /usr/src/sys/kern/kern_shutdown.c:764 > #5 0xffffffff805dff22 in __rw_assert (c=<optimized out>, > what=<optimized out>, file=<unavailable>, line=<unavailable>) > at /usr/src/sys/kern/kern_rwlock.c:1397 > #6 0xffffffff80882723 in vm_page_ps_test (m=0xfffff80431c2e980, flags=7, > skip_m=0xfffff80431c34890) at /usr/src/sys/vm/vm_page.c:4135 > #7 0xffffffff80867d84 in vm_fault_soft_fast (vaddr=<optimized out>, > prot=<optimized out>, fault_type=<optimized out>, > fault_flags=<optimized out>, wired=0, fs=<optimized out>, > m_hold=<optimized out>) at /usr/src/sys/vm/vm_fault.c:307 > #8 vm_fault_hold (map=0xfffff8000832a000, vaddr=<optimized out>, > fault_type=<optimized out>, fault_flags=<optimized out>, m_hold=0x0) > at /usr/src/sys/vm/vm_fault.c:610 > #9 0xffffffff80866cf5 in vm_fault (map=0xfffff8000832a000, > vaddr=<optimized out>, fault_type=2 '\002', fault_flags=0) > at /usr/src/sys/vm/vm_fault.c:514 > #10 0xffffffff808bc64c in trap_pfault (frame=0xfffffe008b1dbac0, usermode=1) > at /usr/src/sys/amd64/amd64/trap.c:728 > #11 0xffffffff808bbe1e in trap (frame=0xfffffe008b1dbac0) > #12 <signal handler called> > #13 0x0000000805b51556 in ?? () > > (kgdb) frame 6 > #6 0xffffffff80882723 in vm_page_ps_test (m=0xfffff80431c2e980, flags=7, > skip_m=0xfffff80431c34890) at /usr/src/sys/vm/vm_page.c:4135 > (kgdb) l > 4130 { > 4131 vm_object_t object; > 4132 int i, npages; > 4133 > 4134 object = m->object; > 4135 VM_OBJECT_ASSERT_LOCKED(object); > 4136 npages = atop(pagesizes[m->psind]); > 4137 > 4138 /* > 4139 * The physically contiguous pages that make up a superpage, i.e., a > (kgdb) p m->object > $1 = (vm_object_t) 0xfffff80190785900 > (kgdb) p pagesizes[m->psind] > $3 = 2097152 > (kgdb) up > #7 0xffffffff80867d84 in vm_fault_soft_fast (vaddr=<optimized out>, > prot=<optimized out>, fault_type=<optimized out>, > fault_flags=<optimized out>, wired=0, fs=<optimized out>, > m_hold=<optimized out>) at /usr/src/sys/vm/vm_fault.c:307 > 307 if (vm_page_ps_test(m_super, flags, m)) { > (kgdb) p m->object > $4 = (vm_object_t) 0xfffff80190116a00 > (kgdb) p/x m->flags > $5 = 0x0 > > So 'm' (original page fault page) and 'm_super' are from different VM > objects. Why are they part of the same reservation? > > (kgdb) p m->phys_addr >> (9 + 12) > $7 = 4514 > (kgdb) p vm_reserv_array[$7] > $8 = {lock = {lock_object = {lo_name = 0xffffffff8099112c "vm reserv", > lo_flags = 16973824, lo_data = 0, lo_witness = 0x0}, mtx_lock = 0}, > partpopq = {tqe_next = 0x0, tqe_prev = 0xfffff80423656680}, objq = { > le_next = 0xfffff8042365b0c0, le_prev = 0xfffff80190116ab8}, > object = 0xfffff80190116a00, pindex = 1760, pages = 0xfffff80431c2e980, > domain = 0, popcnt = 512, inpartpopq = 0 '\000', popmap = { > 18446744073709551615, 18446744073709551615, 18446744073709551615, > 18446744073709551615, 18446744073709551615, 18446744073709551615, > 18446744073709551615, 18446744073709551615}} > (kgdb) set $rv = vm_reserv_array[$7] > (kgdb) p $rv.object > $9 = (vm_object_t) 0xfffff80190116a00 > > So rv->object matches m->object ($4) but not m_super->object ($1). > Double-checking: > > (kgdb) p m_super->object > $10 = (vm_object_t) 0xfffff80190785900 > > Other conditions in vm_reserv_to_superpage() are true: > > (kgdb) p $rv.pages > $11 = (vm_page_t) 0xfffff80431c2e980 > (kgdb) p m_super > $12 = (vm_page_t) 0xfffff80431c2e980 > (kgdb) p $rv.popcnt > $13 = 512 > > Both objects are OBJT_DEFAULT objects: > > (kgdb) p *m->object > $14 = {lock = {lock_object = {lo_name = 0xffffffff8095e7ce "vm object", > lo_flags = 627245056, lo_data = 0, lo_witness = 0x0}, rw_lock = 41}, > object_list = {tqe_next = 0xfffff80190116b00, > tqe_prev = 0xfffff80190116920}, shadow_head = {lh_first = 0x0}, > shadow_list = {le_next = 0x0, le_prev = 0xfffff80190785930}, memq = { > tqh_first = 0xfffff80431ddf878, tqh_last = 0xfffff80431e2a900}, rtree = { > rt_root = 18446735284333515328}, size = 2829, domain = {dr_policy = 0x0, > dr_iterator = 0}, generation = 1, ref_count = 3, shadow_count = 0, > memattr = 6 '\006', type = 0 '\000', flags = 12352, pg_color = 1824, > paging_in_progress = 1, resident_page_count = 1024, > backing_object = 0xfffff80190785900, backing_object_offset = 0, > pager_object_list = {tqe_next = 0x0, tqe_prev = 0x0}, rvq = { > lh_first = 0xfffff80423659540}, handle = 0x0, un_pager = {vnp = { > vnp_size = 0, writemappings = 0}, devp = {devp_pglist = { > tqh_first = 0x0, tqh_last = 0x0}, ops = 0x0, dev = 0x0}, sgp = { > sgp_pglist = {tqh_first = 0x0, tqh_last = 0x0}}, swp = {swp_tmpfs = 0x0, > swp_blks = {pt_root = 0}}}, cred = 0xfffff80008d99500, > charge = 11587584, umtx_data = 0x0} > (kgdb) p *m_super->object > $15 = {lock = {lock_object = {lo_name = 0xffffffff8095e7ce "vm object", > lo_flags = 627245056, lo_data = 0, lo_witness = 0x0}, rw_lock = 1}, > object_list = {tqe_next = 0xfffff80190785a00, > tqe_prev = 0xfffff80190785820}, shadow_head = { > lh_first = 0xfffff80190116a00}, shadow_list = { > le_next = 0xfffff801902b2100, le_prev = 0xfffff80008d61d30}, memq = { > tqh_first = 0xfffff80431bd0720, tqh_last = 0xfffff80431c49b70}, rtree = { > rt_root = 18446735284336303488}, size = 2829, domain = {dr_policy = 0x0, > dr_iterator = 0}, generation = 1, ref_count = 1, shadow_count = 1, > memattr = 6 '\006', type = 0 '\000', flags = 4096, pg_color = 42784, > paging_in_progress = 0, resident_page_count = 2820, backing_object = 0x0, > backing_object_offset = 0, pager_object_list = {tqe_next = 0x0, > tqe_prev = 0x0}, rvq = {lh_first = 0xfffff80423659680}, handle = 0x0, > un_pager = {vnp = {vnp_size = 0, writemappings = 0}, devp = {devp_pglist = { > tqh_first = 0x0, tqh_last = 0x0}, ops = 0x0, dev = 0x0}, sgp = { > sgp_pglist = {tqh_first = 0x0, tqh_last = 0x0}}, swp = {swp_tmpfs = 0x0, > swp_blks = {pt_root = 0}}}, cred = 0xfffff80008d99500, > charge = 11587584, umtx_data = 0x0} > > Huh, m_super's object is the backing object of the object for 'm' and > 'rv': My suspicion is that there's a thread collapsing m->object with its backing object. As it moves pages into the backing object, vm_reserv_rename() can cause a situation where pages allocated from a reservation do not belong to the same object as that reservation. I think that the fast path should simply verify that m_super->object == m->object. If that's not true then we have to fall back to the slow path anyway because the "object consistency test" in vm_page_ps_test() will fail. diff --git a/sys/vm/vm_fault.c b/sys/vm/vm_fault.c index 2a951759325b..b546be7ea4d8 100644 --- a/sys/vm/vm_fault.c +++ b/sys/vm/vm_fault.c _at__at_ -287,6 +287,7 _at__at_ vm_fault_soft_fast(struct faultstate *fs, vm_offset_t vaddr, vm_prot_t prot, #if defined(__amd64__) && VM_NRESERVLEVEL > 0 if ((m->flags & PG_FICTITIOUS) == 0 && (m_super = vm_reserv_to_superpage(m)) != NULL && + m_super->object == m->object && rounddown2(vaddr, pagesizes[m_super->psind]) >= fs->entry->start && roundup2(vaddr + 1, pagesizes[m_super->psind]) <= fs->entry->end && (vaddr & (pagesizes[m_super->psind] - 1)) == (VM_PAGE_TO_PHYS(m) &Received on Tue Apr 17 2018 - 15:38:16 UTC
This archive was generated by hypermail 2.4.0 : Wed May 19 2021 - 11:41:15 UTC