mm: further cleanup page_add_new_anon_rmap
Moving lru_cache_add_active_or_unevictable() into page_add_new_anon_rmap() was good but stupid: we can and should SetPageSwapBacked() there too; and we know for sure that this anonymous, swap-backed page is not file cache. Signed-off-by: Hugh Dickins <hugh@veritas.com> Cc: Lee Schermerhorn <lee.schermerhorn@hp.com> Cc: Nick Piggin <nickpiggin@yahoo.com.au> Acked-by: Rik van Riel <riel@redhat.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
committed by
Linus Torvalds
parent
2afd1c928f
commit
cbf84b7add
@ -47,7 +47,6 @@
|
||||
#include <linux/rmap.h>
|
||||
#include <linux/rcupdate.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/mm_inline.h>
|
||||
#include <linux/kallsyms.h>
|
||||
#include <linux/memcontrol.h>
|
||||
#include <linux/mmu_notifier.h>
|
||||
@ -673,10 +672,11 @@ void page_add_new_anon_rmap(struct page *page,
|
||||
struct vm_area_struct *vma, unsigned long address)
|
||||
{
|
||||
VM_BUG_ON(address < vma->vm_start || address >= vma->vm_end);
|
||||
atomic_set(&page->_mapcount, 0); /* elevate count by 1 (starts at -1) */
|
||||
SetPageSwapBacked(page);
|
||||
atomic_set(&page->_mapcount, 0); /* increment count (starts at -1) */
|
||||
__page_set_anon_rmap(page, vma, address);
|
||||
if (page_evictable(page, vma))
|
||||
lru_cache_add_lru(page, LRU_ACTIVE + page_is_file_cache(page));
|
||||
lru_cache_add_lru(page, LRU_ACTIVE_ANON);
|
||||
else
|
||||
add_page_to_unevictable_list(page);
|
||||
}
|
||||
|
Reference in New Issue
Block a user