mm, thp: count thp_fault_fallback anytime thp fault fails
Currently, thp_fault_fallback in vmstat only gets incremented if a hugepage allocation fails. If current's memcg hits its limit or the page fault handler returns an error, it is incorrectly accounted as a successful thp_fault_alloc. Count thp_fault_fallback anytime the page fault handler falls back to using regular pages and only count thp_fault_alloc when a hugepage has actually been faulted. Signed-off-by: David Rientjes <rientjes@google.com> Cc: Mel Gorman <mgorman@suse.de> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: "Kirill A. Shutemov" <kirill.shutemov@linux.intel.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
committed by
Linus Torvalds
parent
c02925540c
commit
17766dde36
@@ -820,17 +820,19 @@ int do_huge_pmd_anonymous_page(struct mm_struct *mm, struct vm_area_struct *vma,
|
|||||||
count_vm_event(THP_FAULT_FALLBACK);
|
count_vm_event(THP_FAULT_FALLBACK);
|
||||||
return VM_FAULT_FALLBACK;
|
return VM_FAULT_FALLBACK;
|
||||||
}
|
}
|
||||||
count_vm_event(THP_FAULT_ALLOC);
|
|
||||||
if (unlikely(mem_cgroup_newpage_charge(page, mm, GFP_KERNEL))) {
|
if (unlikely(mem_cgroup_newpage_charge(page, mm, GFP_KERNEL))) {
|
||||||
put_page(page);
|
put_page(page);
|
||||||
|
count_vm_event(THP_FAULT_FALLBACK);
|
||||||
return VM_FAULT_FALLBACK;
|
return VM_FAULT_FALLBACK;
|
||||||
}
|
}
|
||||||
if (unlikely(__do_huge_pmd_anonymous_page(mm, vma, haddr, pmd, page))) {
|
if (unlikely(__do_huge_pmd_anonymous_page(mm, vma, haddr, pmd, page))) {
|
||||||
mem_cgroup_uncharge_page(page);
|
mem_cgroup_uncharge_page(page);
|
||||||
put_page(page);
|
put_page(page);
|
||||||
|
count_vm_event(THP_FAULT_FALLBACK);
|
||||||
return VM_FAULT_FALLBACK;
|
return VM_FAULT_FALLBACK;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
count_vm_event(THP_FAULT_ALLOC);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1143,7 +1145,6 @@ alloc:
|
|||||||
new_page = NULL;
|
new_page = NULL;
|
||||||
|
|
||||||
if (unlikely(!new_page)) {
|
if (unlikely(!new_page)) {
|
||||||
count_vm_event(THP_FAULT_FALLBACK);
|
|
||||||
if (is_huge_zero_pmd(orig_pmd)) {
|
if (is_huge_zero_pmd(orig_pmd)) {
|
||||||
ret = do_huge_pmd_wp_zero_page_fallback(mm, vma,
|
ret = do_huge_pmd_wp_zero_page_fallback(mm, vma,
|
||||||
address, pmd, orig_pmd, haddr);
|
address, pmd, orig_pmd, haddr);
|
||||||
@@ -1154,9 +1155,9 @@ alloc:
|
|||||||
split_huge_page(page);
|
split_huge_page(page);
|
||||||
put_page(page);
|
put_page(page);
|
||||||
}
|
}
|
||||||
|
count_vm_event(THP_FAULT_FALLBACK);
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
count_vm_event(THP_FAULT_ALLOC);
|
|
||||||
|
|
||||||
if (unlikely(mem_cgroup_newpage_charge(new_page, mm, GFP_KERNEL))) {
|
if (unlikely(mem_cgroup_newpage_charge(new_page, mm, GFP_KERNEL))) {
|
||||||
put_page(new_page);
|
put_page(new_page);
|
||||||
@@ -1164,10 +1165,13 @@ alloc:
|
|||||||
split_huge_page(page);
|
split_huge_page(page);
|
||||||
put_page(page);
|
put_page(page);
|
||||||
}
|
}
|
||||||
|
count_vm_event(THP_FAULT_FALLBACK);
|
||||||
ret |= VM_FAULT_OOM;
|
ret |= VM_FAULT_OOM;
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
count_vm_event(THP_FAULT_ALLOC);
|
||||||
|
|
||||||
if (is_huge_zero_pmd(orig_pmd))
|
if (is_huge_zero_pmd(orig_pmd))
|
||||||
clear_huge_page(new_page, haddr, HPAGE_PMD_NR);
|
clear_huge_page(new_page, haddr, HPAGE_PMD_NR);
|
||||||
else
|
else
|
||||||
|
Reference in New Issue
Block a user