Loading mm/vmalloc.c +16 −12 Original line number Diff line number Diff line Loading @@ -667,7 +667,7 @@ static atomic_t vmap_lazy_nr = ATOMIC_INIT(0); * by this look, but we want to avoid concurrent calls for performance * reasons and to make the pcpu_get_vm_areas more deterministic. */ static DEFINE_SPINLOCK(vmap_purge_lock); static DEFINE_MUTEX(vmap_purge_lock); /* for per-CPU blocks */ static void purge_fragmented_blocks_allcpus(void); Loading @@ -689,7 +689,7 @@ static bool __purge_vmap_area_lazy(unsigned long start, unsigned long end) struct llist_node *valist; struct vmap_area *va; struct vmap_area *n_va; int nr = 0; bool do_free = false; lockdep_assert_held(&vmap_purge_lock); Loading @@ -699,18 +699,22 @@ static bool __purge_vmap_area_lazy(unsigned long start, unsigned long end) start = va->va_start; if (va->va_end > end) end = va->va_end; nr += (va->va_end - va->va_start) >> PAGE_SHIFT; do_free = true; } if (!nr) if (!do_free) return false; atomic_sub(nr, &vmap_lazy_nr); flush_tlb_kernel_range(start, end); spin_lock(&vmap_area_lock); llist_for_each_entry_safe(va, n_va, valist, purge_list) llist_for_each_entry_safe(va, n_va, valist, purge_list) { int nr = (va->va_end - va->va_start) >> PAGE_SHIFT; __free_vmap_area(va); atomic_sub(nr, &vmap_lazy_nr); cond_resched_lock(&vmap_area_lock); } spin_unlock(&vmap_area_lock); return true; } Loading @@ -721,9 +725,9 @@ static bool __purge_vmap_area_lazy(unsigned long start, unsigned long end) */ static void try_purge_vmap_area_lazy(void) { if (spin_trylock(&vmap_purge_lock)) { if (mutex_trylock(&vmap_purge_lock)) { __purge_vmap_area_lazy(ULONG_MAX, 0); spin_unlock(&vmap_purge_lock); mutex_unlock(&vmap_purge_lock); } } Loading @@ -732,10 +736,10 @@ static void try_purge_vmap_area_lazy(void) */ static void purge_vmap_area_lazy(void) { spin_lock(&vmap_purge_lock); mutex_lock(&vmap_purge_lock); purge_fragmented_blocks_allcpus(); __purge_vmap_area_lazy(ULONG_MAX, 0); spin_unlock(&vmap_purge_lock); mutex_unlock(&vmap_purge_lock); } /* Loading Loading @@ -1124,11 +1128,11 @@ void vm_unmap_aliases(void) rcu_read_unlock(); } spin_lock(&vmap_purge_lock); mutex_lock(&vmap_purge_lock); purge_fragmented_blocks_allcpus(); if (!__purge_vmap_area_lazy(start, end) && flush) flush_tlb_kernel_range(start, end); spin_unlock(&vmap_purge_lock); mutex_unlock(&vmap_purge_lock); } EXPORT_SYMBOL_GPL(vm_unmap_aliases); Loading Loading
mm/vmalloc.c +16 −12 Original line number Diff line number Diff line Loading @@ -667,7 +667,7 @@ static atomic_t vmap_lazy_nr = ATOMIC_INIT(0); * by this look, but we want to avoid concurrent calls for performance * reasons and to make the pcpu_get_vm_areas more deterministic. */ static DEFINE_SPINLOCK(vmap_purge_lock); static DEFINE_MUTEX(vmap_purge_lock); /* for per-CPU blocks */ static void purge_fragmented_blocks_allcpus(void); Loading @@ -689,7 +689,7 @@ static bool __purge_vmap_area_lazy(unsigned long start, unsigned long end) struct llist_node *valist; struct vmap_area *va; struct vmap_area *n_va; int nr = 0; bool do_free = false; lockdep_assert_held(&vmap_purge_lock); Loading @@ -699,18 +699,22 @@ static bool __purge_vmap_area_lazy(unsigned long start, unsigned long end) start = va->va_start; if (va->va_end > end) end = va->va_end; nr += (va->va_end - va->va_start) >> PAGE_SHIFT; do_free = true; } if (!nr) if (!do_free) return false; atomic_sub(nr, &vmap_lazy_nr); flush_tlb_kernel_range(start, end); spin_lock(&vmap_area_lock); llist_for_each_entry_safe(va, n_va, valist, purge_list) llist_for_each_entry_safe(va, n_va, valist, purge_list) { int nr = (va->va_end - va->va_start) >> PAGE_SHIFT; __free_vmap_area(va); atomic_sub(nr, &vmap_lazy_nr); cond_resched_lock(&vmap_area_lock); } spin_unlock(&vmap_area_lock); return true; } Loading @@ -721,9 +725,9 @@ static bool __purge_vmap_area_lazy(unsigned long start, unsigned long end) */ static void try_purge_vmap_area_lazy(void) { if (spin_trylock(&vmap_purge_lock)) { if (mutex_trylock(&vmap_purge_lock)) { __purge_vmap_area_lazy(ULONG_MAX, 0); spin_unlock(&vmap_purge_lock); mutex_unlock(&vmap_purge_lock); } } Loading @@ -732,10 +736,10 @@ static void try_purge_vmap_area_lazy(void) */ static void purge_vmap_area_lazy(void) { spin_lock(&vmap_purge_lock); mutex_lock(&vmap_purge_lock); purge_fragmented_blocks_allcpus(); __purge_vmap_area_lazy(ULONG_MAX, 0); spin_unlock(&vmap_purge_lock); mutex_unlock(&vmap_purge_lock); } /* Loading Loading @@ -1124,11 +1128,11 @@ void vm_unmap_aliases(void) rcu_read_unlock(); } spin_lock(&vmap_purge_lock); mutex_lock(&vmap_purge_lock); purge_fragmented_blocks_allcpus(); if (!__purge_vmap_area_lazy(start, end) && flush) flush_tlb_kernel_range(start, end); spin_unlock(&vmap_purge_lock); mutex_unlock(&vmap_purge_lock); } EXPORT_SYMBOL_GPL(vm_unmap_aliases); Loading