Loading arch/arc/include/asm/page.h +7 −12 Original line number Diff line number Diff line Loading @@ -10,7 +10,6 @@ #include <uapi/asm/page.h> #ifndef __ASSEMBLY__ #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) Loading Loading @@ -76,7 +75,9 @@ typedef unsigned long pgprot_t; typedef pte_t * pgtable_t; #define ARCH_PFN_OFFSET (CONFIG_LINUX_LINK_BASE >> PAGE_SHIFT) #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) #define ARCH_PFN_OFFSET virt_to_pfn(CONFIG_LINUX_LINK_BASE) #define pfn_valid(pfn) (((pfn) - ARCH_PFN_OFFSET) < max_mapnr) Loading @@ -85,21 +86,15 @@ typedef pte_t * pgtable_t; * * These macros have historically been misnamed * virt here means link-address/program-address as embedded in object code. * So if kernel img is linked at 0x8000_0000 onwards, 0x8010_0000 will be * 128th page, and virt_to_page( ) will return the struct page corresp to it. * mem_map[ ] is an array of struct page for each page frame in the system * * Independent of where linux is linked at, link-addr = physical address * So the old macro __pa = vaddr + PAGE_OFFSET - CONFIG_LINUX_LINK_BASE * would have been wrong in case kernel is not at 0x8zs * And for ARC, link-addr = physical address */ #define __pa(vaddr) ((unsigned long)vaddr) #define __va(paddr) ((void *)((unsigned long)(paddr))) #define virt_to_page(kaddr) \ (mem_map + ((__pa(kaddr) - CONFIG_LINUX_LINK_BASE) >> PAGE_SHIFT)) (mem_map + virt_to_pfn((kaddr) - CONFIG_LINUX_LINK_BASE)) #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) /* Default Permissions for stack/heaps pages (Non Executable) */ #define VM_DATA_DEFAULT_FLAGS (VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE) Loading arch/arc/include/asm/pgtable.h +4 −5 Original line number Diff line number Diff line Loading @@ -278,15 +278,14 @@ static inline void pmd_set(pmd_t *pmdp, pte_t *ptep) #define pmd_present(x) (pmd_val(x)) #define pmd_clear(xp) do { pmd_val(*(xp)) = 0; } while (0) #define pte_page(x) (mem_map + \ (unsigned long)(((pte_val(x) - CONFIG_LINUX_LINK_BASE) >> \ PAGE_SHIFT))) #define pte_page(pte) \ (mem_map + virt_to_pfn(pte_val(pte) - CONFIG_LINUX_LINK_BASE)) #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot) #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT) #define pte_pfn(pte) virt_to_pfn(pte_val(pte)) #define pfn_pte(pfn, prot) (__pte(((pte_t)(pfn) << PAGE_SHIFT) | \ pgprot_val(prot))) #define __pte_index(addr) (((addr) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) #define __pte_index(addr) (virt_to_pfn(addr) & (PTRS_PER_PTE - 1)) /* * pte_offset gets a @ptr to PMD entry (PGD in our 2-tier paging system) Loading Loading
arch/arc/include/asm/page.h +7 −12 Original line number Diff line number Diff line Loading @@ -10,7 +10,6 @@ #include <uapi/asm/page.h> #ifndef __ASSEMBLY__ #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) Loading Loading @@ -76,7 +75,9 @@ typedef unsigned long pgprot_t; typedef pte_t * pgtable_t; #define ARCH_PFN_OFFSET (CONFIG_LINUX_LINK_BASE >> PAGE_SHIFT) #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) #define ARCH_PFN_OFFSET virt_to_pfn(CONFIG_LINUX_LINK_BASE) #define pfn_valid(pfn) (((pfn) - ARCH_PFN_OFFSET) < max_mapnr) Loading @@ -85,21 +86,15 @@ typedef pte_t * pgtable_t; * * These macros have historically been misnamed * virt here means link-address/program-address as embedded in object code. * So if kernel img is linked at 0x8000_0000 onwards, 0x8010_0000 will be * 128th page, and virt_to_page( ) will return the struct page corresp to it. * mem_map[ ] is an array of struct page for each page frame in the system * * Independent of where linux is linked at, link-addr = physical address * So the old macro __pa = vaddr + PAGE_OFFSET - CONFIG_LINUX_LINK_BASE * would have been wrong in case kernel is not at 0x8zs * And for ARC, link-addr = physical address */ #define __pa(vaddr) ((unsigned long)vaddr) #define __va(paddr) ((void *)((unsigned long)(paddr))) #define virt_to_page(kaddr) \ (mem_map + ((__pa(kaddr) - CONFIG_LINUX_LINK_BASE) >> PAGE_SHIFT)) (mem_map + virt_to_pfn((kaddr) - CONFIG_LINUX_LINK_BASE)) #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) /* Default Permissions for stack/heaps pages (Non Executable) */ #define VM_DATA_DEFAULT_FLAGS (VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE) Loading
arch/arc/include/asm/pgtable.h +4 −5 Original line number Diff line number Diff line Loading @@ -278,15 +278,14 @@ static inline void pmd_set(pmd_t *pmdp, pte_t *ptep) #define pmd_present(x) (pmd_val(x)) #define pmd_clear(xp) do { pmd_val(*(xp)) = 0; } while (0) #define pte_page(x) (mem_map + \ (unsigned long)(((pte_val(x) - CONFIG_LINUX_LINK_BASE) >> \ PAGE_SHIFT))) #define pte_page(pte) \ (mem_map + virt_to_pfn(pte_val(pte) - CONFIG_LINUX_LINK_BASE)) #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot) #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT) #define pte_pfn(pte) virt_to_pfn(pte_val(pte)) #define pfn_pte(pfn, prot) (__pte(((pte_t)(pfn) << PAGE_SHIFT) | \ pgprot_val(prot))) #define __pte_index(addr) (((addr) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) #define __pte_index(addr) (virt_to_pfn(addr) & (PTRS_PER_PTE - 1)) /* * pte_offset gets a @ptr to PMD entry (PGD in our 2-tier paging system) Loading