Loading Documentation/cachetlb.txt +0 −9 Original line number Original line Diff line number Diff line Loading @@ -49,9 +49,6 @@ changes occur: page table operations such as what happens during page table operations such as what happens during fork, and exec. fork, and exec. Platform developers note that generic code will always invoke this interface without mm->page_table_lock held. 3) void flush_tlb_range(struct vm_area_struct *vma, 3) void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) unsigned long start, unsigned long end) Loading @@ -72,9 +69,6 @@ changes occur: call flush_tlb_page (see below) for each entry which may be call flush_tlb_page (see below) for each entry which may be modified. modified. Platform developers note that generic code will always invoke this interface with mm->page_table_lock held. 4) void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) 4) void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) This time we need to remove the PAGE_SIZE sized translation This time we need to remove the PAGE_SIZE sized translation Loading @@ -93,9 +87,6 @@ changes occur: This is used primarily during fault processing. This is used primarily during fault processing. Platform developers note that generic code will always invoke this interface with mm->page_table_lock held. 5) void flush_tlb_pgtables(struct mm_struct *mm, 5) void flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) unsigned long start, unsigned long end) Loading Documentation/kernel-parameters.txt +0 −2 Original line number Original line Diff line number Diff line Loading @@ -1460,8 +1460,6 @@ running once the system is up. stifb= [HW] stifb= [HW] Format: bpp:<bpp1>[:<bpp2>[:<bpp3>...]] Format: bpp:<bpp1>[:<bpp2>[:<bpp3>...]] stram_swap= [HW,M68k] swiotlb= [IA-64] Number of I/O TLB slabs swiotlb= [IA-64] Number of I/O TLB slabs switches= [HW,M68k] switches= [HW,M68k] Loading Documentation/m68k/kernel-options.txt +1 −23 Original line number Original line Diff line number Diff line Loading @@ -626,7 +626,7 @@ ignored (others aren't affected). can be performed in optimal order. Not all SCSI devices support can be performed in optimal order. Not all SCSI devices support tagged queuing (:-(). tagged queuing (:-(). 4.6 switches= 4.5 switches= ------------- ------------- Syntax: switches=<list of switches> Syntax: switches=<list of switches> Loading Loading @@ -661,28 +661,6 @@ correctly. earlier initialization ("ov_"-less) takes precedence. But the earlier initialization ("ov_"-less) takes precedence. But the switching-off on reset still happens in this case. switching-off on reset still happens in this case. 4.5) stram_swap= ---------------- Syntax: stram_swap=<do_swap>[,<max_swap>] This option is available only if the kernel has been compiled with CONFIG_STRAM_SWAP enabled. Normally, the kernel then determines dynamically whether to actually use ST-RAM as swap space. (Currently, the fraction of ST-RAM must be less or equal 1/3 of total memory to enable this swapping.) You can override the kernel's decision by specifying this option. 1 for <do_swap> means always enable the swap, even if you have less alternate RAM. 0 stands for never swap to ST-RAM, even if it's small enough compared to the rest of memory. If ST-RAM swapping is enabled, the kernel usually uses all free ST-RAM as swap "device". If the kernel resides in ST-RAM, the region allocated by it is obviously never used for swapping :-) You can also limit this amount by specifying the second parameter, <max_swap>, if you want to use parts of ST-RAM as normal system memory. <max_swap> is in kBytes and the number should be a multiple of 4 (otherwise: rounded down). 5) Options for Amiga Only: 5) Options for Amiga Only: ========================== ========================== Loading arch/alpha/mm/numa.c +3 −0 Original line number Original line Diff line number Diff line Loading @@ -371,6 +371,8 @@ show_mem(void) show_free_areas(); show_free_areas(); printk("Free swap: %6ldkB\n", nr_swap_pages<<(PAGE_SHIFT-10)); printk("Free swap: %6ldkB\n", nr_swap_pages<<(PAGE_SHIFT-10)); for_each_online_node(nid) { for_each_online_node(nid) { unsigned long flags; pgdat_resize_lock(NODE_DATA(nid), &flags); i = node_spanned_pages(nid); i = node_spanned_pages(nid); while (i-- > 0) { while (i-- > 0) { struct page *page = nid_page_nr(nid, i); struct page *page = nid_page_nr(nid, i); Loading @@ -384,6 +386,7 @@ show_mem(void) else else shared += page_count(page) - 1; shared += page_count(page) - 1; } } pgdat_resize_unlock(NODE_DATA(nid), &flags); } } printk("%ld pages of RAM\n",total); printk("%ld pages of RAM\n",total); printk("%ld free pages\n",free); printk("%ld free pages\n",free); Loading arch/alpha/mm/remap.c +1 −5 Original line number Original line Diff line number Diff line Loading @@ -2,7 +2,6 @@ #include <asm/pgalloc.h> #include <asm/pgalloc.h> #include <asm/cacheflush.h> #include <asm/cacheflush.h> /* called with the page_table_lock held */ static inline void static inline void remap_area_pte(pte_t * pte, unsigned long address, unsigned long size, remap_area_pte(pte_t * pte, unsigned long address, unsigned long size, unsigned long phys_addr, unsigned long flags) unsigned long phys_addr, unsigned long flags) Loading Loading @@ -31,7 +30,6 @@ remap_area_pte(pte_t * pte, unsigned long address, unsigned long size, } while (address && (address < end)); } while (address && (address < end)); } } /* called with the page_table_lock held */ static inline int static inline int remap_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size, remap_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size, unsigned long phys_addr, unsigned long flags) unsigned long phys_addr, unsigned long flags) Loading @@ -46,7 +44,7 @@ remap_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size, if (address >= end) if (address >= end) BUG(); BUG(); do { do { pte_t * pte = pte_alloc_kernel(&init_mm, pmd, address); pte_t * pte = pte_alloc_kernel(pmd, address); if (!pte) if (!pte) return -ENOMEM; return -ENOMEM; remap_area_pte(pte, address, end - address, remap_area_pte(pte, address, end - address, Loading @@ -70,7 +68,6 @@ __alpha_remap_area_pages(unsigned long address, unsigned long phys_addr, flush_cache_all(); flush_cache_all(); if (address >= end) if (address >= end) BUG(); BUG(); spin_lock(&init_mm.page_table_lock); do { do { pmd_t *pmd; pmd_t *pmd; pmd = pmd_alloc(&init_mm, dir, address); pmd = pmd_alloc(&init_mm, dir, address); Loading @@ -84,7 +81,6 @@ __alpha_remap_area_pages(unsigned long address, unsigned long phys_addr, address = (address + PGDIR_SIZE) & PGDIR_MASK; address = (address + PGDIR_SIZE) & PGDIR_MASK; dir++; dir++; } while (address && (address < end)); } while (address && (address < end)); spin_unlock(&init_mm.page_table_lock); return error; return error; } } Loading
Documentation/cachetlb.txt +0 −9 Original line number Original line Diff line number Diff line Loading @@ -49,9 +49,6 @@ changes occur: page table operations such as what happens during page table operations such as what happens during fork, and exec. fork, and exec. Platform developers note that generic code will always invoke this interface without mm->page_table_lock held. 3) void flush_tlb_range(struct vm_area_struct *vma, 3) void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) unsigned long start, unsigned long end) Loading @@ -72,9 +69,6 @@ changes occur: call flush_tlb_page (see below) for each entry which may be call flush_tlb_page (see below) for each entry which may be modified. modified. Platform developers note that generic code will always invoke this interface with mm->page_table_lock held. 4) void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) 4) void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) This time we need to remove the PAGE_SIZE sized translation This time we need to remove the PAGE_SIZE sized translation Loading @@ -93,9 +87,6 @@ changes occur: This is used primarily during fault processing. This is used primarily during fault processing. Platform developers note that generic code will always invoke this interface with mm->page_table_lock held. 5) void flush_tlb_pgtables(struct mm_struct *mm, 5) void flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) unsigned long start, unsigned long end) Loading
Documentation/kernel-parameters.txt +0 −2 Original line number Original line Diff line number Diff line Loading @@ -1460,8 +1460,6 @@ running once the system is up. stifb= [HW] stifb= [HW] Format: bpp:<bpp1>[:<bpp2>[:<bpp3>...]] Format: bpp:<bpp1>[:<bpp2>[:<bpp3>...]] stram_swap= [HW,M68k] swiotlb= [IA-64] Number of I/O TLB slabs swiotlb= [IA-64] Number of I/O TLB slabs switches= [HW,M68k] switches= [HW,M68k] Loading
Documentation/m68k/kernel-options.txt +1 −23 Original line number Original line Diff line number Diff line Loading @@ -626,7 +626,7 @@ ignored (others aren't affected). can be performed in optimal order. Not all SCSI devices support can be performed in optimal order. Not all SCSI devices support tagged queuing (:-(). tagged queuing (:-(). 4.6 switches= 4.5 switches= ------------- ------------- Syntax: switches=<list of switches> Syntax: switches=<list of switches> Loading Loading @@ -661,28 +661,6 @@ correctly. earlier initialization ("ov_"-less) takes precedence. But the earlier initialization ("ov_"-less) takes precedence. But the switching-off on reset still happens in this case. switching-off on reset still happens in this case. 4.5) stram_swap= ---------------- Syntax: stram_swap=<do_swap>[,<max_swap>] This option is available only if the kernel has been compiled with CONFIG_STRAM_SWAP enabled. Normally, the kernel then determines dynamically whether to actually use ST-RAM as swap space. (Currently, the fraction of ST-RAM must be less or equal 1/3 of total memory to enable this swapping.) You can override the kernel's decision by specifying this option. 1 for <do_swap> means always enable the swap, even if you have less alternate RAM. 0 stands for never swap to ST-RAM, even if it's small enough compared to the rest of memory. If ST-RAM swapping is enabled, the kernel usually uses all free ST-RAM as swap "device". If the kernel resides in ST-RAM, the region allocated by it is obviously never used for swapping :-) You can also limit this amount by specifying the second parameter, <max_swap>, if you want to use parts of ST-RAM as normal system memory. <max_swap> is in kBytes and the number should be a multiple of 4 (otherwise: rounded down). 5) Options for Amiga Only: 5) Options for Amiga Only: ========================== ========================== Loading
arch/alpha/mm/numa.c +3 −0 Original line number Original line Diff line number Diff line Loading @@ -371,6 +371,8 @@ show_mem(void) show_free_areas(); show_free_areas(); printk("Free swap: %6ldkB\n", nr_swap_pages<<(PAGE_SHIFT-10)); printk("Free swap: %6ldkB\n", nr_swap_pages<<(PAGE_SHIFT-10)); for_each_online_node(nid) { for_each_online_node(nid) { unsigned long flags; pgdat_resize_lock(NODE_DATA(nid), &flags); i = node_spanned_pages(nid); i = node_spanned_pages(nid); while (i-- > 0) { while (i-- > 0) { struct page *page = nid_page_nr(nid, i); struct page *page = nid_page_nr(nid, i); Loading @@ -384,6 +386,7 @@ show_mem(void) else else shared += page_count(page) - 1; shared += page_count(page) - 1; } } pgdat_resize_unlock(NODE_DATA(nid), &flags); } } printk("%ld pages of RAM\n",total); printk("%ld pages of RAM\n",total); printk("%ld free pages\n",free); printk("%ld free pages\n",free); Loading
arch/alpha/mm/remap.c +1 −5 Original line number Original line Diff line number Diff line Loading @@ -2,7 +2,6 @@ #include <asm/pgalloc.h> #include <asm/pgalloc.h> #include <asm/cacheflush.h> #include <asm/cacheflush.h> /* called with the page_table_lock held */ static inline void static inline void remap_area_pte(pte_t * pte, unsigned long address, unsigned long size, remap_area_pte(pte_t * pte, unsigned long address, unsigned long size, unsigned long phys_addr, unsigned long flags) unsigned long phys_addr, unsigned long flags) Loading Loading @@ -31,7 +30,6 @@ remap_area_pte(pte_t * pte, unsigned long address, unsigned long size, } while (address && (address < end)); } while (address && (address < end)); } } /* called with the page_table_lock held */ static inline int static inline int remap_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size, remap_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size, unsigned long phys_addr, unsigned long flags) unsigned long phys_addr, unsigned long flags) Loading @@ -46,7 +44,7 @@ remap_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size, if (address >= end) if (address >= end) BUG(); BUG(); do { do { pte_t * pte = pte_alloc_kernel(&init_mm, pmd, address); pte_t * pte = pte_alloc_kernel(pmd, address); if (!pte) if (!pte) return -ENOMEM; return -ENOMEM; remap_area_pte(pte, address, end - address, remap_area_pte(pte, address, end - address, Loading @@ -70,7 +68,6 @@ __alpha_remap_area_pages(unsigned long address, unsigned long phys_addr, flush_cache_all(); flush_cache_all(); if (address >= end) if (address >= end) BUG(); BUG(); spin_lock(&init_mm.page_table_lock); do { do { pmd_t *pmd; pmd_t *pmd; pmd = pmd_alloc(&init_mm, dir, address); pmd = pmd_alloc(&init_mm, dir, address); Loading @@ -84,7 +81,6 @@ __alpha_remap_area_pages(unsigned long address, unsigned long phys_addr, address = (address + PGDIR_SIZE) & PGDIR_MASK; address = (address + PGDIR_SIZE) & PGDIR_MASK; dir++; dir++; } while (address && (address < end)); } while (address && (address < end)); spin_unlock(&init_mm.page_table_lock); return error; return error; } }