Lines Matching full:stride

241  *	__flush_tlb_range(vma, start, end, stride, last_level, tlb_level)
245 * determined by 'stride' and only affect any walk-cache entries
369 * @stride: Flush granularity
377 * entries one by one at the granularity of 'stride'. If the TLB
396 #define __flush_tlb_range_op(op, start, pages, stride, \ argument
414 __flush_start += stride; \
415 __flush_pages -= stride >> PAGE_SHIFT; \
433 #define __flush_s2_tlb_range_op(op, start, pages, stride, tlb_level) \ argument
434 __flush_tlb_range_op(op, start, pages, stride, 0, tlb_level, false, kvm_lpa2_is_enabled());
437 unsigned long end, unsigned long pages, unsigned long stride) in __flush_tlb_range_limit_excess() argument
446 (end - start) >= (MAX_DVM_OPS * stride)) || in __flush_tlb_range_limit_excess()
455 unsigned long stride, bool last_level, in __flush_tlb_range_nosync() argument
460 start = round_down(start, stride); in __flush_tlb_range_nosync()
461 end = round_up(end, stride); in __flush_tlb_range_nosync()
464 if (__flush_tlb_range_limit_excess(start, end, pages, stride)) { in __flush_tlb_range_nosync()
473 __flush_tlb_range_op(vale1is, start, pages, stride, asid, in __flush_tlb_range_nosync()
476 __flush_tlb_range_op(vae1is, start, pages, stride, asid, in __flush_tlb_range_nosync()
484 unsigned long stride, bool last_level, in __flush_tlb_range() argument
487 __flush_tlb_range_nosync(vma, start, end, stride, in __flush_tlb_range()
506 const unsigned long stride = PAGE_SIZE; in flush_tlb_kernel_range() local
509 start = round_down(start, stride); in flush_tlb_kernel_range()
510 end = round_up(end, stride); in flush_tlb_kernel_range()
513 if (__flush_tlb_range_limit_excess(start, end, pages, stride)) { in flush_tlb_kernel_range()
519 __flush_tlb_range_op(vaale1is, start, pages, stride, 0, in flush_tlb_kernel_range()