Loading arch/alpha/include/asm/pgtable.h +0 −3 Original line number Diff line number Diff line Loading @@ -354,9 +354,6 @@ extern inline pte_t mk_swap_pte(unsigned long type, unsigned long offset) #define kern_addr_valid(addr) (1) #endif #define io_remap_pfn_range(vma, start, pfn, size, prot) \ remap_pfn_range(vma, start, pfn, size, prot) #define pte_ERROR(e) \ printk("%s:%d: bad pte %016lx.\n", __FILE__, __LINE__, pte_val(e)) #define pmd_ERROR(e) \ Loading arch/arc/include/asm/pgtable.h +0 −3 Original line number Diff line number Diff line Loading @@ -394,9 +394,6 @@ void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, * remap a physical page `pfn' of size `size' with page protection `prot' * into virtual address `from' */ #define io_remap_pfn_range(vma, from, pfn, size, prot) \ remap_pfn_range(vma, from, pfn, size, prot) #include <asm-generic/pgtable.h> /* to cope with aliasing VIPT cache */ Loading arch/arm/include/asm/pgtable-nommu.h +0 −2 Original line number Diff line number Diff line Loading @@ -79,8 +79,6 @@ extern unsigned int kobjsize(const void *objp); * No page table caches to initialise. */ #define pgtable_cache_init() do { } while (0) #define io_remap_pfn_range remap_pfn_range /* * All 32bit addresses are effectively valid for vmalloc... Loading arch/arm/include/asm/pgtable.h +0 −7 Original line number Diff line number Diff line Loading @@ -318,13 +318,6 @@ static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) #define HAVE_ARCH_UNMAPPED_AREA #define HAVE_ARCH_UNMAPPED_AREA_TOPDOWN /* * remap a physical page `pfn' of size `size' with page protection `prot' * into virtual address `from' */ #define io_remap_pfn_range(vma,from,pfn,size,prot) \ remap_pfn_range(vma, from, pfn, size, prot) #define pgtable_cache_init() do { } while (0) #endif /* !__ASSEMBLY__ */ Loading arch/arm64/include/asm/pgtable.h +0 −7 Original line number Diff line number Diff line Loading @@ -320,13 +320,6 @@ extern int kern_addr_valid(unsigned long addr); #include <asm-generic/pgtable.h> /* * remap a physical page `pfn' of size `size' with page protection `prot' * into virtual address `from' */ #define io_remap_pfn_range(vma,from,pfn,size,prot) \ remap_pfn_range(vma, from, pfn, size, prot) #define pgtable_cache_init() do { } while (0) #endif /* !__ASSEMBLY__ */ Loading Loading
arch/alpha/include/asm/pgtable.h +0 −3 Original line number Diff line number Diff line Loading @@ -354,9 +354,6 @@ extern inline pte_t mk_swap_pte(unsigned long type, unsigned long offset) #define kern_addr_valid(addr) (1) #endif #define io_remap_pfn_range(vma, start, pfn, size, prot) \ remap_pfn_range(vma, start, pfn, size, prot) #define pte_ERROR(e) \ printk("%s:%d: bad pte %016lx.\n", __FILE__, __LINE__, pte_val(e)) #define pmd_ERROR(e) \ Loading
arch/arc/include/asm/pgtable.h +0 −3 Original line number Diff line number Diff line Loading @@ -394,9 +394,6 @@ void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, * remap a physical page `pfn' of size `size' with page protection `prot' * into virtual address `from' */ #define io_remap_pfn_range(vma, from, pfn, size, prot) \ remap_pfn_range(vma, from, pfn, size, prot) #include <asm-generic/pgtable.h> /* to cope with aliasing VIPT cache */ Loading
arch/arm/include/asm/pgtable-nommu.h +0 −2 Original line number Diff line number Diff line Loading @@ -79,8 +79,6 @@ extern unsigned int kobjsize(const void *objp); * No page table caches to initialise. */ #define pgtable_cache_init() do { } while (0) #define io_remap_pfn_range remap_pfn_range /* * All 32bit addresses are effectively valid for vmalloc... Loading
arch/arm/include/asm/pgtable.h +0 −7 Original line number Diff line number Diff line Loading @@ -318,13 +318,6 @@ static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) #define HAVE_ARCH_UNMAPPED_AREA #define HAVE_ARCH_UNMAPPED_AREA_TOPDOWN /* * remap a physical page `pfn' of size `size' with page protection `prot' * into virtual address `from' */ #define io_remap_pfn_range(vma,from,pfn,size,prot) \ remap_pfn_range(vma, from, pfn, size, prot) #define pgtable_cache_init() do { } while (0) #endif /* !__ASSEMBLY__ */ Loading
arch/arm64/include/asm/pgtable.h +0 −7 Original line number Diff line number Diff line Loading @@ -320,13 +320,6 @@ extern int kern_addr_valid(unsigned long addr); #include <asm-generic/pgtable.h> /* * remap a physical page `pfn' of size `size' with page protection `prot' * into virtual address `from' */ #define io_remap_pfn_range(vma,from,pfn,size,prot) \ remap_pfn_range(vma, from, pfn, size, prot) #define pgtable_cache_init() do { } while (0) #endif /* !__ASSEMBLY__ */ Loading