Loading arch/sh/mm/fault_32.c +0 −1 Original line number Diff line number Diff line Loading @@ -200,7 +200,6 @@ asmlinkage void __kprobes do_page_fault(struct pt_regs *regs, * make sure we exit gracefully rather than endlessly redo * the fault. */ survive: fault = handle_mm_fault(mm, vma, address, writeaccess ? FAULT_FLAG_WRITE : 0); if (unlikely(fault & VM_FAULT_ERROR)) { if (fault & VM_FAULT_OOM) Loading arch/sh/mm/tlbflush_64.c +0 −1 Original line number Diff line number Diff line Loading @@ -189,7 +189,6 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long writeaccess, * make sure we exit gracefully rather than endlessly redo * the fault. */ survive: fault = handle_mm_fault(mm, vma, address, writeaccess ? FAULT_FLAG_WRITE : 0); if (unlikely(fault & VM_FAULT_ERROR)) { if (fault & VM_FAULT_OOM) Loading Loading
arch/sh/mm/fault_32.c +0 −1 Original line number Diff line number Diff line Loading @@ -200,7 +200,6 @@ asmlinkage void __kprobes do_page_fault(struct pt_regs *regs, * make sure we exit gracefully rather than endlessly redo * the fault. */ survive: fault = handle_mm_fault(mm, vma, address, writeaccess ? FAULT_FLAG_WRITE : 0); if (unlikely(fault & VM_FAULT_ERROR)) { if (fault & VM_FAULT_OOM) Loading
arch/sh/mm/tlbflush_64.c +0 −1 Original line number Diff line number Diff line Loading @@ -189,7 +189,6 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long writeaccess, * make sure we exit gracefully rather than endlessly redo * the fault. */ survive: fault = handle_mm_fault(mm, vma, address, writeaccess ? FAULT_FLAG_WRITE : 0); if (unlikely(fault & VM_FAULT_ERROR)) { if (fault & VM_FAULT_OOM) Loading