On Tue, 2004/05/18 at 12:57:34 -0700, Julian Elischer wrote: > On Mon, 10 May 2004, Thomas Moestl wrote: > > On Sun, 2004/05/09 at 15:44:40 -0700, Julian Elischer wrote: > > > in vm_machdep.c the sparc64 code has > > > void > > > cpu_sched_exit(struct thread *td) > > > { > > > struct vmspace *vm; > > > struct pcpu *pc; > > > struct proc *p; > > > > > > mtx_assert(&sched_lock, MA_OWNED); > > > > > > p = td->td_proc; > > > vm = p->p_vmspace; > > > if (vm->vm_refcnt > 1) > > > return; > > > SLIST_FOREACH(pc, &cpuhead, pc_allcpu) { > > > if (pc->pc_vmspace == vm) { > > > vm->vm_pmap.pm_active &= ~pc->pc_cpumask; > > > vm->vm_pmap.pm_context[pc->pc_cpuid] = -1; > > > pc->pc_vmspace = NULL; > > > } > > > } > > > } > > > > > > > > > > > > This is thw only architecture that has this.. > > > What does it do? And what does it have to do with the scheduler? > > to answer question 2,, > nothing.. in my sources I renamed it to cpu_exit2() The name probably derives from the fact that it needs to be called after the sched lock is obtained, as was mentioned in the commit message. > > To quote from the commit log: > > date: 2002/06/24 15:48:01; author: jake; state: Exp; lines: +1 -0 > > Add an MD callout like cpu_exit, but which is called after sched_lock is > > obtained, when all other scheduling activity is suspended. This is needed > > on sparc64 to deactivate the vmspace of the exiting process on all cpus. > > Otherwise if another unrelated process gets the exact same vmspace structure > > allocated to it (same address), its address space will not be activated > > properly. This seems to fix some spontaneous signal 11 problems with smp > > on sparc64. > > > > To elaborate on that a bit: > > The sparc64 cpu_switch() has an optimization to avoid needlessly > > invalidating TLB entries: when we switch to a kernel thread, we need > > not switch VM contexts at all, and do with whatever vmspace was active > > before. When we switch to a thread that has the vmspace that is > > already in use currently, we need not load a new context register > > value (which is analog to flushing the TLB). > > > > We identify vmspaces by their pointers for this purpose, so there can > > be a race between freeing the struct vmspace by wait()ing (on another > > processor) and switching to another thread (on the first > > processor). Specifically, the first processor could be switching to a > > newly created thread that has the same struct vmspace that was just > > freed, so we would mistakenly assume that we need not bother loading > > the context register, and continue using outdated TLB entries. > > > > To prevent this, cpu_sched_exit() zeros the respective PCPU variables > > holding the active vmspace if it is going to be destroyed, so it will > > never match any other during the next cpu_switch(). > > I'm not convinced that this is valid.. > consider.. > When you cycle through the processors above and remove the pointers to > the vmspace, then proceed to destroy this vmspace, there is nothing done > to make sure that the other procerssors are actually > not USING the page tables etc. associated with the vmspace. > > If we reclame the page tables.. surely there is a danger that another > cpu by still be using them? No. The code will only invalidate the vmspace pointers on the exit of the last process that uses them; it will therefore primarily do so on the CPU that is processing the exit(). The other case, as mentioned above, can happen due to the switch optimization that keeps the vmspace of the switched-out process when switching to a kernel thread. Kernel threads only need kernel mappings, which are global, so they will never access the active vmspace or its pmap (sparc64 has different TSBs, which are somewhat analogue to page tables, for kernel and user space, and only the user space one is kept in the pmap, the kernel one being global). So, there are two cases for the per-CPU vmspace pointer: - the vmspace was used by the currently running thread. The thread is the last one using it, and exiting, so we can zero the vmspace pointer and free the vmspace. - the vmspace is not used by the currently running thread, which is a kernel thread. It is kept only as a hint that we can optimize a switch-back to that vmspace. This optimization is not possible any more, so we can zero the vmspace pointer, and free the vmspace which a kernel thread does not use. > I think that even "temporary" users of vmspaces, such as kernel threads, > should have reference counts and be capable of freeing the vmspace > should it go to 0 when they complete using it. That would require reference counting at switch time, which would be prohibitively expensive (requiring Giant, for example). Also, it would complicate the freeing of vmspaces in the rest of the kernel code. As mentioned above, kernel threads do not use the vmspace themselves; it is just kept as a hint for later switching. The per-CPU vmspace pointers are not a strong reference, which needs to stay valid; rather, they are a weak reference, and we just need notification in case they become invalid. - Thomas -- Thomas Moestl <t.moestl_at_tu-bs.de> http://www.tu-bs.de/~y0015675/ <tmm_at_FreeBSD.org> http://people.FreeBSD.org/~tmm/ "The most crucial career decision is picking a good 'ism' so everyone knows how to categorize you without understanding the work." -- Calvin and HobbesReceived on Thu May 20 2004 - 03:24:24 UTC
This archive was generated by hypermail 2.4.0 : Wed May 19 2021 - 11:37:54 UTC