mm/rmap.c | 95 +++++++++++++++++++++++++++----------------------------------- 1 files changed, 42 insertions(+), 53 deletions(-) diff -puN mm/rmap.c~rmap-cleanup mm/rmap.c --- bk-linux/mm/rmap.c~rmap-cleanup 2004-11-08 15:08:23.692440784 +0300 +++ bk-linux-nikita/mm/rmap.c 2004-11-08 15:08:23.697440024 +0300 @@ -250,6 +250,34 @@ unsigned long page_address_in_vma(struct } /* + * Check that @page is mapped at @address into @mm. + * + * On success returns with mapped pte and locked mm->page_table_lock. + */ +static inline pte_t *page_check_address(struct page *page, struct mm_struct *mm, + unsigned long address) +{ + pgd_t *pgd; + pmd_t *pmd; + pte_t *pte; + + spin_lock(&mm->page_table_lock); + pgd = pgd_offset(mm, address); + if (likely(pgd_present(*pgd))) { + pmd = pmd_offset(pgd, address); + if (likely(pmd_present(*pmd))) { + pte = pte_offset_map(pmd, address); + if (likely(pte_present(*pte) && + page_to_pfn(page) == pte_pfn(*pte))) + return pte; + pte_unmap(pte); + } + } + spin_unlock(&mm->page_table_lock); + return ERR_PTR(-ENOENT); +} + +/* * Subfunctions of page_referenced: page_referenced_one called * repeatedly from either page_referenced_anon or page_referenced_file. */ @@ -258,8 +286,6 @@ static int page_referenced_one(struct pa { struct mm_struct *mm = vma->vm_mm; unsigned long address; - pgd_t *pgd; - pmd_t *pmd; pte_t *pte; int referenced = 0; @@ -269,35 +295,18 @@ static int page_referenced_one(struct pa if (address == -EFAULT) goto out; - spin_lock(&mm->page_table_lock); - - pgd = pgd_offset(mm, address); - if (!pgd_present(*pgd)) - goto out_unlock; - - pmd = pmd_offset(pgd, address); - if (!pmd_present(*pmd)) - goto out_unlock; - - pte = pte_offset_map(pmd, address); - if (!pte_present(*pte)) - goto out_unmap; - - if (page_to_pfn(page) != pte_pfn(*pte)) - goto out_unmap; - - if (ptep_clear_flush_young(vma, address, pte)) - referenced++; - - if (mm != current->mm && has_swap_token(mm)) - referenced++; + pte = page_check_address(page, mm, address); + if (!IS_ERR(pte)) { + if (ptep_clear_flush_young(vma, address, pte)) + referenced++; - (*mapcount)--; + if (mm != current->mm && has_swap_token(mm)) + referenced++; -out_unmap: - pte_unmap(pte); -out_unlock: - spin_unlock(&mm->page_table_lock); + (*mapcount)--; + pte_unmap(pte); + spin_unlock(&mm->page_table_lock); + } out: return referenced; } @@ -496,8 +505,6 @@ static int try_to_unmap_one(struct page { struct mm_struct *mm = vma->vm_mm; unsigned long address; - pgd_t *pgd; - pmd_t *pmd; pte_t *pte; pte_t pteval; int ret = SWAP_AGAIN; @@ -508,26 +515,9 @@ static int try_to_unmap_one(struct page if (address == -EFAULT) goto out; - /* - * We need the page_table_lock to protect us from page faults, - * munmap, fork, etc... - */ - spin_lock(&mm->page_table_lock); - - pgd = pgd_offset(mm, address); - if (!pgd_present(*pgd)) - goto out_unlock; - - pmd = pmd_offset(pgd, address); - if (!pmd_present(*pmd)) - goto out_unlock; - - pte = pte_offset_map(pmd, address); - if (!pte_present(*pte)) - goto out_unmap; - - if (page_to_pfn(page) != pte_pfn(*pte)) - goto out_unmap; + pte = page_check_address(page, mm, address); + if (IS_ERR(pte)) + goto out; /* * If the page is mlock()d, we cannot swap it out. @@ -592,7 +582,6 @@ static int try_to_unmap_one(struct page out_unmap: pte_unmap(pte); -out_unlock: spin_unlock(&mm->page_table_lock); out: return ret; @@ -691,7 +680,6 @@ static void try_to_unmap_cluster(unsigne } pte_unmap(pte); - out_unlock: spin_unlock(&mm->page_table_lock); } @@ -843,3 +831,4 @@ int try_to_unmap(struct page *page) ret = SWAP_SUCCESS; return ret; } + _