Loading Documentation/memory-barriers.txt +1 −1 Original line number Diff line number Diff line Loading @@ -602,7 +602,7 @@ Consider the following sequence of events: This sequence of events is committed to the memory coherence system in an order that the rest of the system might perceive as the unordered set of { STORE A, STORE B, STORE C } all occuring before the unordered set of { STORE D, STORE E STORE B, STORE C } all occurring before the unordered set of { STORE D, STORE E }: +-------+ : : Loading arch/i386/mm/fault.c +1 −1 Original line number Diff line number Diff line Loading @@ -389,7 +389,7 @@ fastcall void __kprobes do_page_fault(struct pt_regs *regs, /* When running in the kernel we expect faults to occur only to * addresses in user space. All other faults represent errors in the * kernel and should generate an OOPS. Unfortunatly, in the case of an * erroneous fault occuring in a code path which already holds mmap_sem * erroneous fault occurring in a code path which already holds mmap_sem * we will deadlock attempting to validate the fault against the * address space. Luckily the kernel only validly references user * space from well defined areas of code, which are listed in the Loading arch/m32r/mm/fault.c +1 −1 Original line number Diff line number Diff line Loading @@ -148,7 +148,7 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long error_code, /* When running in the kernel we expect faults to occur only to * addresses in user space. All other faults represent errors in the * kernel and should generate an OOPS. Unfortunatly, in the case of an * erroneous fault occuring in a code path which already holds mmap_sem * erroneous fault occurring in a code path which already holds mmap_sem * we will deadlock attempting to validate the fault against the * address space. Luckily the kernel only validly references user * space from well defined areas of code, which are listed in the Loading arch/powerpc/platforms/powermac/cpufreq_32.c +1 −1 Original line number Diff line number Diff line Loading @@ -268,7 +268,7 @@ static int pmu_set_cpu_speed(int low_speed) /* Make sure the decrementer won't interrupt us */ asm volatile("mtdec %0" : : "r" (0x7fffffff)); /* Make sure any pending DEC interrupt occuring while we did /* Make sure any pending DEC interrupt occurring while we did * the above didn't re-enable the DEC */ mb(); asm volatile("mtdec %0" : : "r" (0x7fffffff)); Loading arch/powerpc/platforms/pseries/eeh_driver.c +1 −1 Original line number Diff line number Diff line Loading @@ -175,7 +175,7 @@ static void eeh_report_failure(struct pci_dev *dev, void *userdata) * * pSeries systems will isolate a PCI slot if the PCI-Host * bridge detects address or data parity errors, DMA's * occuring to wild addresses (which usually happen due to * occurring to wild addresses (which usually happen due to * bugs in device drivers or in PCI adapter firmware). * Slot isolations also occur if #SERR, #PERR or other misc * PCI-related errors are detected. Loading Loading
Documentation/memory-barriers.txt +1 −1 Original line number Diff line number Diff line Loading @@ -602,7 +602,7 @@ Consider the following sequence of events: This sequence of events is committed to the memory coherence system in an order that the rest of the system might perceive as the unordered set of { STORE A, STORE B, STORE C } all occuring before the unordered set of { STORE D, STORE E STORE B, STORE C } all occurring before the unordered set of { STORE D, STORE E }: +-------+ : : Loading
arch/i386/mm/fault.c +1 −1 Original line number Diff line number Diff line Loading @@ -389,7 +389,7 @@ fastcall void __kprobes do_page_fault(struct pt_regs *regs, /* When running in the kernel we expect faults to occur only to * addresses in user space. All other faults represent errors in the * kernel and should generate an OOPS. Unfortunatly, in the case of an * erroneous fault occuring in a code path which already holds mmap_sem * erroneous fault occurring in a code path which already holds mmap_sem * we will deadlock attempting to validate the fault against the * address space. Luckily the kernel only validly references user * space from well defined areas of code, which are listed in the Loading
arch/m32r/mm/fault.c +1 −1 Original line number Diff line number Diff line Loading @@ -148,7 +148,7 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long error_code, /* When running in the kernel we expect faults to occur only to * addresses in user space. All other faults represent errors in the * kernel and should generate an OOPS. Unfortunatly, in the case of an * erroneous fault occuring in a code path which already holds mmap_sem * erroneous fault occurring in a code path which already holds mmap_sem * we will deadlock attempting to validate the fault against the * address space. Luckily the kernel only validly references user * space from well defined areas of code, which are listed in the Loading
arch/powerpc/platforms/powermac/cpufreq_32.c +1 −1 Original line number Diff line number Diff line Loading @@ -268,7 +268,7 @@ static int pmu_set_cpu_speed(int low_speed) /* Make sure the decrementer won't interrupt us */ asm volatile("mtdec %0" : : "r" (0x7fffffff)); /* Make sure any pending DEC interrupt occuring while we did /* Make sure any pending DEC interrupt occurring while we did * the above didn't re-enable the DEC */ mb(); asm volatile("mtdec %0" : : "r" (0x7fffffff)); Loading
arch/powerpc/platforms/pseries/eeh_driver.c +1 −1 Original line number Diff line number Diff line Loading @@ -175,7 +175,7 @@ static void eeh_report_failure(struct pci_dev *dev, void *userdata) * * pSeries systems will isolate a PCI slot if the PCI-Host * bridge detects address or data parity errors, DMA's * occuring to wild addresses (which usually happen due to * occurring to wild addresses (which usually happen due to * bugs in device drivers or in PCI adapter firmware). * Slot isolations also occur if #SERR, #PERR or other misc * PCI-related errors are detected. Loading