Searched refs:need_flush (Results 1 – 9 of 9) sorted by relevance
52 get_new_mmu_context(struct mm_struct *mm, unsigned long cpu, bool *need_flush) in get_new_mmu_context() argument57 *need_flush = true; /* start new asid cycle */ in get_new_mmu_context()91 bool need_flush = false; in switch_mm_irqs_off() local96 get_new_mmu_context(next, cpu, &need_flush); in switch_mm_irqs_off()103 if (need_flush) in switch_mm_irqs_off()151 bool need_flush = false; in drop_mmu_context() local154 get_new_mmu_context(mm, cpu, &need_flush); in drop_mmu_context()157 if (need_flush) in drop_mmu_context()
368 unsigned int subregions, bool need_flush) in mpu_setup_region() argument387 if (need_flush) in mpu_setup_region()441 bool need_flush = region == PMSAv7_RAM_REGION; in pmsav7_setup() local448 xip[i].subreg, need_flush); in pmsav7_setup()
19 static inline bool need_flush(struct iommu_map_table *iommu) in need_flush() function206 (n < pool->hint || need_flush(iommu))) { in iommu_tbl_range_alloc()
199 int need_flush = 0; in flush_all_zero_pkmaps() local232 need_flush = 1; in flush_all_zero_pkmaps()234 if (need_flush) in flush_all_zero_pkmaps()
529 int need_flush = 0; in mlx4_ib_cm_paravirt_clean() local537 need_flush |= !cancel_delayed_work(&map->timeout); in mlx4_ib_cm_paravirt_clean()543 if (need_flush) in mlx4_ib_cm_paravirt_clean()
583 capsnap->need_flush = true; in ceph_queue_cap_snap()595 ceph_cap_string(dirty), capsnap->need_flush ? "" : "no_flush"); in ceph_queue_cap_snap()
1475 if (capsnap->need_flush) { in __prep_cap()1640 BUG_ON(!capsnap->need_flush); in __ceph_flush_snaps()3203 if (!capsnap->need_flush && in ceph_try_drop_cap_snap()
1000 bool need_flush = false; in writecache_resume() local1071 need_flush = true; in writecache_resume()1092 need_flush = true; in writecache_resume()1099 if (need_flush) { in writecache_resume()
4684 int need_flush = 0; in reshape_request() local4725 need_flush = 1; in reshape_request()4747 need_flush = 1; in reshape_request()4757 if (need_flush || in reshape_request()