Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 9a54b720 authored by Laurent Dufour's avatar Laurent Dufour Committed by Vinayak Menon
Browse files

mm: adding speculative page fault failure trace events



This patch a set of new trace events to collect the speculative page fault
event failures.

Signed-off-by: default avatarLaurent Dufour <ldufour@linux.vnet.ibm.com>
Change-Id: I9dafba293edf40bdad4ae241d105ecdfb42579c1
Patch-mainline: linux-mm @ Tue, 17 Apr 2018 16:33:25
[vinmenon@codeaurora.org: trivial merge conflicts]
Signed-off-by: default avatarVinayak Menon <vinmenon@codeaurora.org>
parent ce37f07c
Loading
Loading
Loading
Loading
+88 −0
Original line number Diff line number Diff line
/* SPDX-License-Identifier: GPL-2.0 */
#undef TRACE_SYSTEM
#define TRACE_SYSTEM pagefault

#if !defined(_TRACE_PAGEFAULT_H) || defined(TRACE_HEADER_MULTI_READ)
#define _TRACE_PAGEFAULT_H

#include <linux/tracepoint.h>
#include <linux/mm.h>

DECLARE_EVENT_CLASS(spf,

	TP_PROTO(unsigned long caller,
		 struct vm_area_struct *vma, unsigned long address),

	TP_ARGS(caller, vma, address),

	TP_STRUCT__entry(
		__field(unsigned long, caller)
		__field(unsigned long, vm_start)
		__field(unsigned long, vm_end)
		__field(unsigned long, address)
	),

	TP_fast_assign(
		__entry->caller		= caller;
		__entry->vm_start	= vma->vm_start;
		__entry->vm_end		= vma->vm_end;
		__entry->address	= address;
	),

	TP_printk("ip:%lx vma:%lx-%lx address:%lx",
		  __entry->caller, __entry->vm_start, __entry->vm_end,
		  __entry->address)
);

DEFINE_EVENT(spf, spf_pte_lock,

	TP_PROTO(unsigned long caller,
		 struct vm_area_struct *vma, unsigned long address),

	TP_ARGS(caller, vma, address)
);

DEFINE_EVENT(spf, spf_vma_changed,

	TP_PROTO(unsigned long caller,
		 struct vm_area_struct *vma, unsigned long address),

	TP_ARGS(caller, vma, address)
);

DEFINE_EVENT(spf, spf_vma_noanon,

	TP_PROTO(unsigned long caller,
		 struct vm_area_struct *vma, unsigned long address),

	TP_ARGS(caller, vma, address)
);

DEFINE_EVENT(spf, spf_vma_notsup,

	TP_PROTO(unsigned long caller,
		 struct vm_area_struct *vma, unsigned long address),

	TP_ARGS(caller, vma, address)
);

DEFINE_EVENT(spf, spf_vma_access,

	TP_PROTO(unsigned long caller,
		 struct vm_area_struct *vma, unsigned long address),

	TP_ARGS(caller, vma, address)
);

DEFINE_EVENT(spf, spf_pmd_changed,

	TP_PROTO(unsigned long caller,
		 struct vm_area_struct *vma, unsigned long address),

	TP_ARGS(caller, vma, address)
);

#endif /* _TRACE_PAGEFAULT_H */

/* This part must be outside protection */
#include <trace/define_trace.h>
+51 −14
Original line number Diff line number Diff line
@@ -81,6 +81,9 @@

#include "internal.h"

#define CREATE_TRACE_POINTS
#include <trace/events/pagefault.h>

#if defined(LAST_CPUPID_NOT_IN_PAGE_FLAGS) && !defined(CONFIG_COMPILE_TEST)
#warning Unfortunate NUMA and NUMA Balancing config, growing page-frame for last_cpupid.
#endif
@@ -2299,8 +2302,10 @@ static bool pte_spinlock(struct vm_fault *vmf)
	}

	local_irq_disable();
	if (vma_has_changed(vmf))
	if (vma_has_changed(vmf)) {
		trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address);
		goto out;
	}

#ifdef CONFIG_TRANSPARENT_HUGEPAGE
	/*
@@ -2308,16 +2313,21 @@ static bool pte_spinlock(struct vm_fault *vmf)
	 * is not a huge collapse operation in progress in our back.
	 */
	pmdval = READ_ONCE(*vmf->pmd);
	if (!pmd_same(pmdval, vmf->orig_pmd))
	if (!pmd_same(pmdval, vmf->orig_pmd)) {
		trace_spf_pmd_changed(_RET_IP_, vmf->vma, vmf->address);
		goto out;
	}
#endif

	vmf->ptl = pte_lockptr(vmf->vma->vm_mm, vmf->pmd);
	if (unlikely(!spin_trylock(vmf->ptl)))
	if (unlikely(!spin_trylock(vmf->ptl))) {
		trace_spf_pte_lock(_RET_IP_, vmf->vma, vmf->address);
		goto out;
	}

	if (vma_has_changed(vmf)) {
		spin_unlock(vmf->ptl);
		trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address);
		goto out;
	}

@@ -2350,8 +2360,10 @@ static bool pte_map_lock(struct vm_fault *vmf)
	 * block on the PTL and thus we're safe.
	 */
	local_irq_disable();
	if (vma_has_changed(vmf))
	if (vma_has_changed(vmf)) {
		trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address);
		goto out;
	}

#ifdef CONFIG_TRANSPARENT_HUGEPAGE
	/*
@@ -2359,8 +2371,10 @@ static bool pte_map_lock(struct vm_fault *vmf)
	 * is not a huge collapse operation in progress in our back.
	 */
	pmdval = READ_ONCE(*vmf->pmd);
	if (!pmd_same(pmdval, vmf->orig_pmd))
	if (!pmd_same(pmdval, vmf->orig_pmd)) {
		trace_spf_pmd_changed(_RET_IP_, vmf->vma, vmf->address);
		goto out;
	}
#endif

	/*
@@ -2374,11 +2388,13 @@ static bool pte_map_lock(struct vm_fault *vmf)
	pte = pte_offset_map(vmf->pmd, vmf->address);
	if (unlikely(!spin_trylock(ptl))) {
		pte_unmap(pte);
		trace_spf_pte_lock(_RET_IP_, vmf->vma, vmf->address);
		goto out;
	}

	if (vma_has_changed(vmf)) {
		pte_unmap_unlock(pte, ptl);
		trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address);
		goto out;
	}

@@ -4281,47 +4297,60 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,

	/* rmb <-> seqlock,vma_rb_erase() */
	seq = raw_read_seqcount(&vma->vm_sequence);
	if (seq & 1)
	if (seq & 1) {
		trace_spf_vma_changed(_RET_IP_, vma, address);
		goto out_put;
	}

	/*
	 * Can't call vm_ops service has we don't know what they would do
	 * with the VMA.
	 * This include huge page from hugetlbfs.
	 */
	if (vma->vm_ops)
	if (vma->vm_ops) {
		trace_spf_vma_notsup(_RET_IP_, vma, address);
		goto out_put;
	}

	/*
	 * __anon_vma_prepare() requires the mmap_sem to be held
	 * because vm_next and vm_prev must be safe. This can't be guaranteed
	 * in the speculative path.
	 */
	if (unlikely(!vma->anon_vma))
	if (unlikely(!vma->anon_vma)) {
		trace_spf_vma_notsup(_RET_IP_, vma, address);
		goto out_put;
	}

	vmf.vma_flags = READ_ONCE(vma->vm_flags);
	vmf.vma_page_prot = READ_ONCE(vma->vm_page_prot);

	/* Can't call userland page fault handler in the speculative path */
	if (unlikely(vmf.vma_flags & VM_UFFD_MISSING))
	if (unlikely(vmf.vma_flags & VM_UFFD_MISSING)) {
		trace_spf_vma_notsup(_RET_IP_, vma, address);
		goto out_put;
	}

	if (vmf.vma_flags & VM_GROWSDOWN || vmf.vma_flags & VM_GROWSUP)
	if (vmf.vma_flags & VM_GROWSDOWN || vmf.vma_flags & VM_GROWSUP) {
		/*
		 * This could be detected by the check address against VMA's
		 * boundaries but we want to trace it as not supported instead
		 * of changed.
		 */
		trace_spf_vma_notsup(_RET_IP_, vma, address);
		goto out_put;
	}

	if (address < READ_ONCE(vma->vm_start)
	    || READ_ONCE(vma->vm_end) <= address)
	    || READ_ONCE(vma->vm_end) <= address) {
		trace_spf_vma_changed(_RET_IP_, vma, address);
		goto out_put;
	}

	if (!arch_vma_access_permitted(vma, flags & FAULT_FLAG_WRITE,
				       flags & FAULT_FLAG_INSTRUCTION,
				       flags & FAULT_FLAG_REMOTE)) {
		trace_spf_vma_access(_RET_IP_, vma, address);
		ret = VM_FAULT_SIGSEGV;
		goto out_put;
	}
@@ -4329,10 +4358,12 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,
	/* This is one is required to check that the VMA has write access set */
	if (flags & FAULT_FLAG_WRITE) {
		if (unlikely(!(vmf.vma_flags & VM_WRITE))) {
			trace_spf_vma_access(_RET_IP_, vma, address);
			ret = VM_FAULT_SIGSEGV;
			goto out_put;
		}
	} else if (unlikely(!(vmf.vma_flags & (VM_READ|VM_EXEC|VM_WRITE)))) {
		trace_spf_vma_access(_RET_IP_, vma, address);
		ret = VM_FAULT_SIGSEGV;
		goto out_put;
	}
@@ -4348,8 +4379,11 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,
	pol = __get_vma_policy(vma, address);
	if (!pol)
		pol = get_task_policy(current);
	if (pol && pol->mode == MPOL_INTERLEAVE)
	if (!pol)
		if (pol && pol->mode == MPOL_INTERLEAVE) {
			trace_spf_vma_notsup(_RET_IP_, vma, address);
			goto out_put;
		}
#endif

	/*
@@ -4422,8 +4456,10 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,
	 * We need to re-validate the VMA after checking the bounds, otherwise
	 * we might have a false positive on the bounds.
	 */
	if (read_seqcount_retry(&vma->vm_sequence, seq))
	if (read_seqcount_retry(&vma->vm_sequence, seq)) {
		trace_spf_vma_changed(_RET_IP_, vma, address);
		goto out_put;
	}

	mem_cgroup_oom_enable();
	ret = handle_pte_fault(&vmf);
@@ -4442,6 +4478,7 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,
	return ret;

out_walk:
	trace_spf_vma_notsup(_RET_IP_, vma, address);
	local_irq_enable();
out_put:
	put_vma(vma);