Lines Matching full:pid

130 	unsigned int pid;  in local_flush_tlb_mm()  local
133 pid = mm->context.id; in local_flush_tlb_mm()
134 if (pid != MMU_NO_CONTEXT) in local_flush_tlb_mm()
135 _tlbil_pid(pid); in local_flush_tlb_mm()
143 unsigned int pid; in __local_flush_tlb_page() local
146 pid = mm ? mm->context.id : 0; in __local_flush_tlb_page()
147 if (pid != MMU_NO_CONTEXT) in __local_flush_tlb_page()
148 _tlbil_va(vmaddr, pid, tsize, ind); in __local_flush_tlb_page()
177 unsigned int pid; member
186 _tlbil_pid(p ? p->pid : 0); in do_flush_tlb_mm_ipi()
193 _tlbil_va(p->addr, p->pid, p->tsize, p->ind); in do_flush_tlb_page_ipi()
197 /* Note on invalidations and PID:
199 * We snapshot the PID with preempt disabled. At this point, it can still
201 * - our context is being stolen (PID -> NO_CONTEXT) on another CPU
203 * and is concurrently acquiring a new PID on another CPU
204 * - some other CPU is re-acquiring a lost PID for this mm
209 * don't care about the PID changing, and invalidating a stale PID
215 unsigned int pid; in flush_tlb_mm() local
218 pid = mm->context.id; in flush_tlb_mm()
219 if (unlikely(pid == MMU_NO_CONTEXT)) in flush_tlb_mm()
222 struct tlb_flush_param p = { .pid = pid }; in flush_tlb_mm()
227 _tlbil_pid(pid); in flush_tlb_mm()
237 unsigned int pid; in __flush_tlb_page() local
247 pid = mm->context.id; in __flush_tlb_page()
248 if (unlikely(pid == MMU_NO_CONTEXT)) in __flush_tlb_page()
257 _tlbivax_bcast(vmaddr, pid, tsize, ind); in __flush_tlb_page()
263 .pid = pid, in __flush_tlb_page()
273 _tlbil_va(vmaddr, pid, tsize, ind); in __flush_tlb_page()