Loading mm/memory.c +12 −44 Original line number Diff line number Diff line Loading @@ -84,9 +84,6 @@ #include "internal.h" #define CREATE_TRACE_POINTS #include <trace/events/pagefault.h> #if defined(LAST_CPUPID_NOT_IN_PAGE_FLAGS) && !defined(CONFIG_COMPILE_TEST) #warning Unfortunate NUMA and NUMA Balancing config, growing page-frame for last_cpupid. #endif Loading Loading @@ -2171,10 +2168,8 @@ static bool pte_spinlock(struct vm_fault *vmf) } local_irq_disable(); if (vma_has_changed(vmf)) { trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address); if (vma_has_changed(vmf)) goto out; } #ifdef CONFIG_TRANSPARENT_HUGEPAGE /* Loading @@ -2182,21 +2177,16 @@ static bool pte_spinlock(struct vm_fault *vmf) * is not a huge collapse operation in progress in our back. */ pmdval = READ_ONCE(*vmf->pmd); if (!pmd_same(pmdval, vmf->orig_pmd)) { trace_spf_pmd_changed(_RET_IP_, vmf->vma, vmf->address); if (!pmd_same(pmdval, vmf->orig_pmd)) goto out; } #endif vmf->ptl = pte_lockptr(vmf->vma->vm_mm, vmf->pmd); if (unlikely(!spin_trylock(vmf->ptl))) { trace_spf_pte_lock(_RET_IP_, vmf->vma, vmf->address); if (unlikely(!spin_trylock(vmf->ptl))) goto out; } if (vma_has_changed(vmf)) { spin_unlock(vmf->ptl); trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address); goto out; } Loading Loading @@ -2229,10 +2219,8 @@ static bool pte_map_lock(struct vm_fault *vmf) * block on the PTL and thus we're safe. */ local_irq_disable(); if (vma_has_changed(vmf)) { trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address); if (vma_has_changed(vmf)) goto out; } #ifdef CONFIG_TRANSPARENT_HUGEPAGE /* Loading @@ -2240,10 +2228,8 @@ static bool pte_map_lock(struct vm_fault *vmf) * is not a huge collapse operation in progress in our back. */ pmdval = READ_ONCE(*vmf->pmd); if (!pmd_same(pmdval, vmf->orig_pmd)) { trace_spf_pmd_changed(_RET_IP_, vmf->vma, vmf->address); if (!pmd_same(pmdval, vmf->orig_pmd)) goto out; } #endif /* Loading @@ -2257,13 +2243,11 @@ static bool pte_map_lock(struct vm_fault *vmf) pte = pte_offset_map(vmf->pmd, vmf->address); if (unlikely(!spin_trylock(ptl))) { pte_unmap(pte); trace_spf_pte_lock(_RET_IP_, vmf->vma, vmf->address); goto out; } if (vma_has_changed(vmf)) { pte_unmap_unlock(pte, ptl); trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address); goto out; } Loading Loading @@ -4294,45 +4278,35 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address, /* rmb <-> seqlock,vma_rb_erase() */ seq = raw_read_seqcount(&vmf.vma->vm_sequence); if (seq & 1) { trace_spf_vma_changed(_RET_IP_, vmf.vma, address); if (seq & 1) return VM_FAULT_RETRY; } /* * __anon_vma_prepare() requires the mmap_sem to be held * because vm_next and vm_prev must be safe. This can't be guaranteed * in the speculative path. */ if (unlikely(vma_is_anonymous(vmf.vma) && !vmf.vma->anon_vma)) { trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); if (unlikely(vma_is_anonymous(vmf.vma) && !vmf.vma->anon_vma)) return VM_FAULT_RETRY; } vmf.vma_flags = READ_ONCE(vmf.vma->vm_flags); vmf.vma_page_prot = READ_ONCE(vmf.vma->vm_page_prot); /* Can't call userland page fault handler in the speculative path */ if (unlikely(vmf.vma_flags & VM_UFFD_MISSING)) { trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); if (unlikely(vmf.vma_flags & VM_UFFD_MISSING)) return VM_FAULT_RETRY; } if (vmf.vma_flags & VM_GROWSDOWN || vmf.vma_flags & VM_GROWSUP) { if (vmf.vma_flags & VM_GROWSDOWN || vmf.vma_flags & VM_GROWSUP) /* * This could be detected by the check address against VMA's * boundaries but we want to trace it as not supported instead * of changed. */ trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); return VM_FAULT_RETRY; } if (address < READ_ONCE(vmf.vma->vm_start) || READ_ONCE(vmf.vma->vm_end) <= address) { trace_spf_vma_changed(_RET_IP_, vmf.vma, address); || READ_ONCE(vmf.vma->vm_end) <= address) return VM_FAULT_RETRY; } if (!arch_vma_access_permitted(vmf.vma, flags & FAULT_FLAG_WRITE, flags & FAULT_FLAG_INSTRUCTION, Loading @@ -4358,10 +4332,8 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address, if (!pol) pol = get_task_policy(current); if (!pol) if (pol && pol->mode == MPOL_INTERLEAVE) { trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); if (pol && pol->mode == MPOL_INTERLEAVE) return VM_FAULT_RETRY; } #endif /* Loading Loading @@ -4433,10 +4405,8 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address, * We need to re-validate the VMA after checking the bounds, otherwise * we might have a false positive on the bounds. */ if (read_seqcount_retry(&vmf.vma->vm_sequence, seq)) { trace_spf_vma_changed(_RET_IP_, vmf.vma, address); if (read_seqcount_retry(&vmf.vma->vm_sequence, seq)) return VM_FAULT_RETRY; } mem_cgroup_enter_user_fault(); ret = handle_pte_fault(&vmf); Loading Loading @@ -4465,12 +4435,10 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address, return ret; out_walk: trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); local_irq_enable(); return VM_FAULT_RETRY; out_segv: trace_spf_vma_access(_RET_IP_, vmf.vma, address); /* * We don't return VM_FAULT_RETRY so the caller is not expected to * retrieve the fetched VMA. Loading Loading
mm/memory.c +12 −44 Original line number Diff line number Diff line Loading @@ -84,9 +84,6 @@ #include "internal.h" #define CREATE_TRACE_POINTS #include <trace/events/pagefault.h> #if defined(LAST_CPUPID_NOT_IN_PAGE_FLAGS) && !defined(CONFIG_COMPILE_TEST) #warning Unfortunate NUMA and NUMA Balancing config, growing page-frame for last_cpupid. #endif Loading Loading @@ -2171,10 +2168,8 @@ static bool pte_spinlock(struct vm_fault *vmf) } local_irq_disable(); if (vma_has_changed(vmf)) { trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address); if (vma_has_changed(vmf)) goto out; } #ifdef CONFIG_TRANSPARENT_HUGEPAGE /* Loading @@ -2182,21 +2177,16 @@ static bool pte_spinlock(struct vm_fault *vmf) * is not a huge collapse operation in progress in our back. */ pmdval = READ_ONCE(*vmf->pmd); if (!pmd_same(pmdval, vmf->orig_pmd)) { trace_spf_pmd_changed(_RET_IP_, vmf->vma, vmf->address); if (!pmd_same(pmdval, vmf->orig_pmd)) goto out; } #endif vmf->ptl = pte_lockptr(vmf->vma->vm_mm, vmf->pmd); if (unlikely(!spin_trylock(vmf->ptl))) { trace_spf_pte_lock(_RET_IP_, vmf->vma, vmf->address); if (unlikely(!spin_trylock(vmf->ptl))) goto out; } if (vma_has_changed(vmf)) { spin_unlock(vmf->ptl); trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address); goto out; } Loading Loading @@ -2229,10 +2219,8 @@ static bool pte_map_lock(struct vm_fault *vmf) * block on the PTL and thus we're safe. */ local_irq_disable(); if (vma_has_changed(vmf)) { trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address); if (vma_has_changed(vmf)) goto out; } #ifdef CONFIG_TRANSPARENT_HUGEPAGE /* Loading @@ -2240,10 +2228,8 @@ static bool pte_map_lock(struct vm_fault *vmf) * is not a huge collapse operation in progress in our back. */ pmdval = READ_ONCE(*vmf->pmd); if (!pmd_same(pmdval, vmf->orig_pmd)) { trace_spf_pmd_changed(_RET_IP_, vmf->vma, vmf->address); if (!pmd_same(pmdval, vmf->orig_pmd)) goto out; } #endif /* Loading @@ -2257,13 +2243,11 @@ static bool pte_map_lock(struct vm_fault *vmf) pte = pte_offset_map(vmf->pmd, vmf->address); if (unlikely(!spin_trylock(ptl))) { pte_unmap(pte); trace_spf_pte_lock(_RET_IP_, vmf->vma, vmf->address); goto out; } if (vma_has_changed(vmf)) { pte_unmap_unlock(pte, ptl); trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address); goto out; } Loading Loading @@ -4294,45 +4278,35 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address, /* rmb <-> seqlock,vma_rb_erase() */ seq = raw_read_seqcount(&vmf.vma->vm_sequence); if (seq & 1) { trace_spf_vma_changed(_RET_IP_, vmf.vma, address); if (seq & 1) return VM_FAULT_RETRY; } /* * __anon_vma_prepare() requires the mmap_sem to be held * because vm_next and vm_prev must be safe. This can't be guaranteed * in the speculative path. */ if (unlikely(vma_is_anonymous(vmf.vma) && !vmf.vma->anon_vma)) { trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); if (unlikely(vma_is_anonymous(vmf.vma) && !vmf.vma->anon_vma)) return VM_FAULT_RETRY; } vmf.vma_flags = READ_ONCE(vmf.vma->vm_flags); vmf.vma_page_prot = READ_ONCE(vmf.vma->vm_page_prot); /* Can't call userland page fault handler in the speculative path */ if (unlikely(vmf.vma_flags & VM_UFFD_MISSING)) { trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); if (unlikely(vmf.vma_flags & VM_UFFD_MISSING)) return VM_FAULT_RETRY; } if (vmf.vma_flags & VM_GROWSDOWN || vmf.vma_flags & VM_GROWSUP) { if (vmf.vma_flags & VM_GROWSDOWN || vmf.vma_flags & VM_GROWSUP) /* * This could be detected by the check address against VMA's * boundaries but we want to trace it as not supported instead * of changed. */ trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); return VM_FAULT_RETRY; } if (address < READ_ONCE(vmf.vma->vm_start) || READ_ONCE(vmf.vma->vm_end) <= address) { trace_spf_vma_changed(_RET_IP_, vmf.vma, address); || READ_ONCE(vmf.vma->vm_end) <= address) return VM_FAULT_RETRY; } if (!arch_vma_access_permitted(vmf.vma, flags & FAULT_FLAG_WRITE, flags & FAULT_FLAG_INSTRUCTION, Loading @@ -4358,10 +4332,8 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address, if (!pol) pol = get_task_policy(current); if (!pol) if (pol && pol->mode == MPOL_INTERLEAVE) { trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); if (pol && pol->mode == MPOL_INTERLEAVE) return VM_FAULT_RETRY; } #endif /* Loading Loading @@ -4433,10 +4405,8 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address, * We need to re-validate the VMA after checking the bounds, otherwise * we might have a false positive on the bounds. */ if (read_seqcount_retry(&vmf.vma->vm_sequence, seq)) { trace_spf_vma_changed(_RET_IP_, vmf.vma, address); if (read_seqcount_retry(&vmf.vma->vm_sequence, seq)) return VM_FAULT_RETRY; } mem_cgroup_enter_user_fault(); ret = handle_pte_fault(&vmf); Loading Loading @@ -4465,12 +4435,10 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address, return ret; out_walk: trace_spf_vma_notsup(_RET_IP_, vmf.vma, address); local_irq_enable(); return VM_FAULT_RETRY; out_segv: trace_spf_vma_access(_RET_IP_, vmf.vma, address); /* * We don't return VM_FAULT_RETRY so the caller is not expected to * retrieve the fetched VMA. Loading