forked from mirrors/linux
		
	mm: account pmd page tables to the process
Dave noticed that unprivileged process can allocate significant amount of
memory -- >500 MiB on x86_64 -- and stay unnoticed by oom-killer and
memory cgroup.  The trick is to allocate a lot of PMD page tables.  Linux
kernel doesn't account PMD tables to the process, only PTE.
The use-cases below use few tricks to allocate a lot of PMD page tables
while keeping VmRSS and VmPTE low.  oom_score for the process will be 0.
	#include <errno.h>
	#include <stdio.h>
	#include <stdlib.h>
	#include <unistd.h>
	#include <sys/mman.h>
	#include <sys/prctl.h>
	#define PUD_SIZE (1UL << 30)
	#define PMD_SIZE (1UL << 21)
	#define NR_PUD 130000
	int main(void)
	{
		char *addr = NULL;
		unsigned long i;
		prctl(PR_SET_THP_DISABLE);
		for (i = 0; i < NR_PUD ; i++) {
			addr = mmap(addr + PUD_SIZE, PUD_SIZE, PROT_WRITE|PROT_READ,
					MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
			if (addr == MAP_FAILED) {
				perror("mmap");
				break;
			}
			*addr = 'x';
			munmap(addr, PMD_SIZE);
			mmap(addr, PMD_SIZE, PROT_WRITE|PROT_READ,
					MAP_ANONYMOUS|MAP_PRIVATE|MAP_FIXED, -1, 0);
			if (addr == MAP_FAILED)
				perror("re-mmap"), exit(1);
		}
		printf("PID %d consumed %lu KiB in PMD page tables\n",
				getpid(), i * 4096 >> 10);
		return pause();
	}
The patch addresses the issue by account PMD tables to the process the
same way we account PTE.
The main place where PMD tables is accounted is __pmd_alloc() and
free_pmd_range(). But there're few corner cases:
 - HugeTLB can share PMD page tables. The patch handles by accounting
   the table to all processes who share it.
 - x86 PAE pre-allocates few PMD tables on fork.
 - Architectures with FIRST_USER_ADDRESS > 0. We need to adjust sanity
   check on exit(2).
Accounting only happens on configuration where PMD page table's level is
present (PMD is not folded).  As with nr_ptes we use per-mm counter.  The
counter value is used to calculate baseline for badness score by
oom-killer.
Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Reported-by: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Hugh Dickins <hughd@google.com>
Reviewed-by: Cyrill Gorcunov <gorcunov@openvz.org>
Cc: Pavel Emelyanov <xemul@openvz.org>
Cc: David Rientjes <rientjes@google.com>
Tested-by: Sedat Dilek <sedat.dilek@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
			
			
This commit is contained in:
		
							parent
							
								
									8aa76875dc
								
							
						
					
					
						commit
						dc6c9a35b6
					
				
					 11 changed files with 75 additions and 29 deletions
				
			
		|  | @ -555,12 +555,12 @@ this is causing problems for your system/application. | ||||||
| 
 | 
 | ||||||
| oom_dump_tasks | oom_dump_tasks | ||||||
| 
 | 
 | ||||||
| Enables a system-wide task dump (excluding kernel threads) to be | Enables a system-wide task dump (excluding kernel threads) to be produced | ||||||
| produced when the kernel performs an OOM-killing and includes such | when the kernel performs an OOM-killing and includes such information as | ||||||
| information as pid, uid, tgid, vm size, rss, nr_ptes, swapents, | pid, uid, tgid, vm size, rss, nr_ptes, nr_pmds, swapents, oom_score_adj | ||||||
| oom_score_adj score, and name.  This is helpful to determine why the | score, and name.  This is helpful to determine why the OOM killer was | ||||||
| OOM killer was invoked, to identify the rogue task that caused it, | invoked, to identify the rogue task that caused it, and to determine why | ||||||
| and to determine why the OOM killer chose the task it did to kill. | the OOM killer chose the task it did to kill. | ||||||
| 
 | 
 | ||||||
| If this is set to zero, this information is suppressed.  On very | If this is set to zero, this information is suppressed.  On very | ||||||
| large systems with thousands of tasks it may not be feasible to dump | large systems with thousands of tasks it may not be feasible to dump | ||||||
|  |  | ||||||
|  | @ -190,7 +190,7 @@ void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd) | ||||||
| 
 | 
 | ||||||
| #endif	/* CONFIG_X86_PAE */ | #endif	/* CONFIG_X86_PAE */ | ||||||
| 
 | 
 | ||||||
| static void free_pmds(pmd_t *pmds[]) | static void free_pmds(struct mm_struct *mm, pmd_t *pmds[]) | ||||||
| { | { | ||||||
| 	int i; | 	int i; | ||||||
| 
 | 
 | ||||||
|  | @ -198,10 +198,11 @@ static void free_pmds(pmd_t *pmds[]) | ||||||
| 		if (pmds[i]) { | 		if (pmds[i]) { | ||||||
| 			pgtable_pmd_page_dtor(virt_to_page(pmds[i])); | 			pgtable_pmd_page_dtor(virt_to_page(pmds[i])); | ||||||
| 			free_page((unsigned long)pmds[i]); | 			free_page((unsigned long)pmds[i]); | ||||||
|  | 			mm_dec_nr_pmds(mm); | ||||||
| 		} | 		} | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| static int preallocate_pmds(pmd_t *pmds[]) | static int preallocate_pmds(struct mm_struct *mm, pmd_t *pmds[]) | ||||||
| { | { | ||||||
| 	int i; | 	int i; | ||||||
| 	bool failed = false; | 	bool failed = false; | ||||||
|  | @ -215,11 +216,13 @@ static int preallocate_pmds(pmd_t *pmds[]) | ||||||
| 			pmd = NULL; | 			pmd = NULL; | ||||||
| 			failed = true; | 			failed = true; | ||||||
| 		} | 		} | ||||||
|  | 		if (pmd) | ||||||
|  | 			mm_inc_nr_pmds(mm); | ||||||
| 		pmds[i] = pmd; | 		pmds[i] = pmd; | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	if (failed) { | 	if (failed) { | ||||||
| 		free_pmds(pmds); | 		free_pmds(mm, pmds); | ||||||
| 		return -ENOMEM; | 		return -ENOMEM; | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
|  | @ -246,6 +249,7 @@ static void pgd_mop_up_pmds(struct mm_struct *mm, pgd_t *pgdp) | ||||||
| 
 | 
 | ||||||
| 			paravirt_release_pmd(pgd_val(pgd) >> PAGE_SHIFT); | 			paravirt_release_pmd(pgd_val(pgd) >> PAGE_SHIFT); | ||||||
| 			pmd_free(mm, pmd); | 			pmd_free(mm, pmd); | ||||||
|  | 			mm_dec_nr_pmds(mm); | ||||||
| 		} | 		} | ||||||
| 	} | 	} | ||||||
| } | } | ||||||
|  | @ -283,7 +287,7 @@ pgd_t *pgd_alloc(struct mm_struct *mm) | ||||||
| 
 | 
 | ||||||
| 	mm->pgd = pgd; | 	mm->pgd = pgd; | ||||||
| 
 | 
 | ||||||
| 	if (preallocate_pmds(pmds) != 0) | 	if (preallocate_pmds(mm, pmds) != 0) | ||||||
| 		goto out_free_pgd; | 		goto out_free_pgd; | ||||||
| 
 | 
 | ||||||
| 	if (paravirt_pgd_alloc(mm) != 0) | 	if (paravirt_pgd_alloc(mm) != 0) | ||||||
|  | @ -304,7 +308,7 @@ pgd_t *pgd_alloc(struct mm_struct *mm) | ||||||
| 	return pgd; | 	return pgd; | ||||||
| 
 | 
 | ||||||
| out_free_pmds: | out_free_pmds: | ||||||
| 	free_pmds(pmds); | 	free_pmds(mm, pmds); | ||||||
| out_free_pgd: | out_free_pgd: | ||||||
| 	free_page((unsigned long)pgd); | 	free_page((unsigned long)pgd); | ||||||
| out: | out: | ||||||
|  |  | ||||||
|  | @ -21,7 +21,7 @@ | ||||||
| 
 | 
 | ||||||
| void task_mem(struct seq_file *m, struct mm_struct *mm) | void task_mem(struct seq_file *m, struct mm_struct *mm) | ||||||
| { | { | ||||||
| 	unsigned long data, text, lib, swap; | 	unsigned long data, text, lib, swap, ptes, pmds; | ||||||
| 	unsigned long hiwater_vm, total_vm, hiwater_rss, total_rss; | 	unsigned long hiwater_vm, total_vm, hiwater_rss, total_rss; | ||||||
| 
 | 
 | ||||||
| 	/*
 | 	/*
 | ||||||
|  | @ -42,6 +42,8 @@ void task_mem(struct seq_file *m, struct mm_struct *mm) | ||||||
| 	text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) >> 10; | 	text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) >> 10; | ||||||
| 	lib = (mm->exec_vm << (PAGE_SHIFT-10)) - text; | 	lib = (mm->exec_vm << (PAGE_SHIFT-10)) - text; | ||||||
| 	swap = get_mm_counter(mm, MM_SWAPENTS); | 	swap = get_mm_counter(mm, MM_SWAPENTS); | ||||||
|  | 	ptes = PTRS_PER_PTE * sizeof(pte_t) * atomic_long_read(&mm->nr_ptes); | ||||||
|  | 	pmds = PTRS_PER_PMD * sizeof(pmd_t) * mm_nr_pmds(mm); | ||||||
| 	seq_printf(m, | 	seq_printf(m, | ||||||
| 		"VmPeak:\t%8lu kB\n" | 		"VmPeak:\t%8lu kB\n" | ||||||
| 		"VmSize:\t%8lu kB\n" | 		"VmSize:\t%8lu kB\n" | ||||||
|  | @ -54,6 +56,7 @@ void task_mem(struct seq_file *m, struct mm_struct *mm) | ||||||
| 		"VmExe:\t%8lu kB\n" | 		"VmExe:\t%8lu kB\n" | ||||||
| 		"VmLib:\t%8lu kB\n" | 		"VmLib:\t%8lu kB\n" | ||||||
| 		"VmPTE:\t%8lu kB\n" | 		"VmPTE:\t%8lu kB\n" | ||||||
|  | 		"VmPMD:\t%8lu kB\n" | ||||||
| 		"VmSwap:\t%8lu kB\n", | 		"VmSwap:\t%8lu kB\n", | ||||||
| 		hiwater_vm << (PAGE_SHIFT-10), | 		hiwater_vm << (PAGE_SHIFT-10), | ||||||
| 		total_vm << (PAGE_SHIFT-10), | 		total_vm << (PAGE_SHIFT-10), | ||||||
|  | @ -63,8 +66,8 @@ void task_mem(struct seq_file *m, struct mm_struct *mm) | ||||||
| 		total_rss << (PAGE_SHIFT-10), | 		total_rss << (PAGE_SHIFT-10), | ||||||
| 		data << (PAGE_SHIFT-10), | 		data << (PAGE_SHIFT-10), | ||||||
| 		mm->stack_vm << (PAGE_SHIFT-10), text, lib, | 		mm->stack_vm << (PAGE_SHIFT-10), text, lib, | ||||||
| 		(PTRS_PER_PTE * sizeof(pte_t) * | 		ptes >> 10, | ||||||
| 		 atomic_long_read(&mm->nr_ptes)) >> 10, | 		pmds >> 10, | ||||||
| 		swap << (PAGE_SHIFT-10)); | 		swap << (PAGE_SHIFT-10)); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
|  | @ -1438,8 +1438,32 @@ static inline int __pmd_alloc(struct mm_struct *mm, pud_t *pud, | ||||||
| { | { | ||||||
| 	return 0; | 	return 0; | ||||||
| } | } | ||||||
|  | 
 | ||||||
|  | static inline unsigned long mm_nr_pmds(struct mm_struct *mm) | ||||||
|  | { | ||||||
|  | 	return 0; | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | static inline void mm_inc_nr_pmds(struct mm_struct *mm) {} | ||||||
|  | static inline void mm_dec_nr_pmds(struct mm_struct *mm) {} | ||||||
|  | 
 | ||||||
| #else | #else | ||||||
| int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address); | int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address); | ||||||
|  | 
 | ||||||
|  | static inline unsigned long mm_nr_pmds(struct mm_struct *mm) | ||||||
|  | { | ||||||
|  | 	return atomic_long_read(&mm->nr_pmds); | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | static inline void mm_inc_nr_pmds(struct mm_struct *mm) | ||||||
|  | { | ||||||
|  | 	atomic_long_inc(&mm->nr_pmds); | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | static inline void mm_dec_nr_pmds(struct mm_struct *mm) | ||||||
|  | { | ||||||
|  | 	atomic_long_dec(&mm->nr_pmds); | ||||||
|  | } | ||||||
| #endif | #endif | ||||||
| 
 | 
 | ||||||
| int __pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, | int __pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, | ||||||
|  |  | ||||||
|  | @ -363,7 +363,8 @@ struct mm_struct { | ||||||
| 	pgd_t * pgd; | 	pgd_t * pgd; | ||||||
| 	atomic_t mm_users;			/* How many users with user space? */ | 	atomic_t mm_users;			/* How many users with user space? */ | ||||||
| 	atomic_t mm_count;			/* How many references to "struct mm_struct" (users count as 1) */ | 	atomic_t mm_count;			/* How many references to "struct mm_struct" (users count as 1) */ | ||||||
| 	atomic_long_t nr_ptes;			/* Page table pages */ | 	atomic_long_t nr_ptes;			/* PTE page table pages */ | ||||||
|  | 	atomic_long_t nr_pmds;			/* PMD page table pages */ | ||||||
| 	int map_count;				/* number of VMAs */ | 	int map_count;				/* number of VMAs */ | ||||||
| 
 | 
 | ||||||
| 	spinlock_t page_table_lock;		/* Protects page tables and some counters */ | 	spinlock_t page_table_lock;		/* Protects page tables and some counters */ | ||||||
|  |  | ||||||
|  | @ -555,6 +555,9 @@ static struct mm_struct *mm_init(struct mm_struct *mm, struct task_struct *p) | ||||||
| 	INIT_LIST_HEAD(&mm->mmlist); | 	INIT_LIST_HEAD(&mm->mmlist); | ||||||
| 	mm->core_state = NULL; | 	mm->core_state = NULL; | ||||||
| 	atomic_long_set(&mm->nr_ptes, 0); | 	atomic_long_set(&mm->nr_ptes, 0); | ||||||
|  | #ifndef __PAGETABLE_PMD_FOLDED | ||||||
|  | 	atomic_long_set(&mm->nr_pmds, 0); | ||||||
|  | #endif | ||||||
| 	mm->map_count = 0; | 	mm->map_count = 0; | ||||||
| 	mm->locked_vm = 0; | 	mm->locked_vm = 0; | ||||||
| 	mm->pinned_vm = 0; | 	mm->pinned_vm = 0; | ||||||
|  |  | ||||||
|  | @ -173,7 +173,7 @@ void dump_mm(const struct mm_struct *mm) | ||||||
| 		"get_unmapped_area %p\n" | 		"get_unmapped_area %p\n" | ||||||
| #endif | #endif | ||||||
| 		"mmap_base %lu mmap_legacy_base %lu highest_vm_end %lu\n" | 		"mmap_base %lu mmap_legacy_base %lu highest_vm_end %lu\n" | ||||||
| 		"pgd %p mm_users %d mm_count %d nr_ptes %lu map_count %d\n" | 		"pgd %p mm_users %d mm_count %d nr_ptes %lu nr_pmds %lu map_count %d\n" | ||||||
| 		"hiwater_rss %lx hiwater_vm %lx total_vm %lx locked_vm %lx\n" | 		"hiwater_rss %lx hiwater_vm %lx total_vm %lx locked_vm %lx\n" | ||||||
| 		"pinned_vm %lx shared_vm %lx exec_vm %lx stack_vm %lx\n" | 		"pinned_vm %lx shared_vm %lx exec_vm %lx stack_vm %lx\n" | ||||||
| 		"start_code %lx end_code %lx start_data %lx end_data %lx\n" | 		"start_code %lx end_code %lx start_data %lx end_data %lx\n" | ||||||
|  | @ -206,6 +206,7 @@ void dump_mm(const struct mm_struct *mm) | ||||||
| 		mm->pgd, atomic_read(&mm->mm_users), | 		mm->pgd, atomic_read(&mm->mm_users), | ||||||
| 		atomic_read(&mm->mm_count), | 		atomic_read(&mm->mm_count), | ||||||
| 		atomic_long_read((atomic_long_t *)&mm->nr_ptes), | 		atomic_long_read((atomic_long_t *)&mm->nr_ptes), | ||||||
|  | 		mm_nr_pmds((struct mm_struct *)mm), | ||||||
| 		mm->map_count, | 		mm->map_count, | ||||||
| 		mm->hiwater_rss, mm->hiwater_vm, mm->total_vm, mm->locked_vm, | 		mm->hiwater_rss, mm->hiwater_vm, mm->total_vm, mm->locked_vm, | ||||||
| 		mm->pinned_vm, mm->shared_vm, mm->exec_vm, mm->stack_vm, | 		mm->pinned_vm, mm->shared_vm, mm->exec_vm, mm->stack_vm, | ||||||
|  |  | ||||||
|  | @ -3598,6 +3598,7 @@ pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud) | ||||||
| 		if (saddr) { | 		if (saddr) { | ||||||
| 			spte = huge_pte_offset(svma->vm_mm, saddr); | 			spte = huge_pte_offset(svma->vm_mm, saddr); | ||||||
| 			if (spte) { | 			if (spte) { | ||||||
|  | 				mm_inc_nr_pmds(mm); | ||||||
| 				get_page(virt_to_page(spte)); | 				get_page(virt_to_page(spte)); | ||||||
| 				break; | 				break; | ||||||
| 			} | 			} | ||||||
|  | @ -3609,11 +3610,13 @@ pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud) | ||||||
| 
 | 
 | ||||||
| 	ptl = huge_pte_lockptr(hstate_vma(vma), mm, spte); | 	ptl = huge_pte_lockptr(hstate_vma(vma), mm, spte); | ||||||
| 	spin_lock(ptl); | 	spin_lock(ptl); | ||||||
| 	if (pud_none(*pud)) | 	if (pud_none(*pud)) { | ||||||
| 		pud_populate(mm, pud, | 		pud_populate(mm, pud, | ||||||
| 				(pmd_t *)((unsigned long)spte & PAGE_MASK)); | 				(pmd_t *)((unsigned long)spte & PAGE_MASK)); | ||||||
| 	else | 	} else { | ||||||
| 		put_page(virt_to_page(spte)); | 		put_page(virt_to_page(spte)); | ||||||
|  | 		mm_inc_nr_pmds(mm); | ||||||
|  | 	} | ||||||
| 	spin_unlock(ptl); | 	spin_unlock(ptl); | ||||||
| out: | out: | ||||||
| 	pte = (pte_t *)pmd_alloc(mm, pud, addr); | 	pte = (pte_t *)pmd_alloc(mm, pud, addr); | ||||||
|  | @ -3644,6 +3647,7 @@ int huge_pmd_unshare(struct mm_struct *mm, unsigned long *addr, pte_t *ptep) | ||||||
| 
 | 
 | ||||||
| 	pud_clear(pud); | 	pud_clear(pud); | ||||||
| 	put_page(virt_to_page(ptep)); | 	put_page(virt_to_page(ptep)); | ||||||
|  | 	mm_dec_nr_pmds(mm); | ||||||
| 	*addr = ALIGN(*addr, HPAGE_SIZE * PTRS_PER_PTE) - HPAGE_SIZE; | 	*addr = ALIGN(*addr, HPAGE_SIZE * PTRS_PER_PTE) - HPAGE_SIZE; | ||||||
| 	return 1; | 	return 1; | ||||||
| } | } | ||||||
|  |  | ||||||
							
								
								
									
										15
									
								
								mm/memory.c
									
									
									
									
									
								
							
							
						
						
									
										15
									
								
								mm/memory.c
									
									
									
									
									
								
							|  | @ -428,6 +428,7 @@ static inline void free_pmd_range(struct mmu_gather *tlb, pud_t *pud, | ||||||
| 	pmd = pmd_offset(pud, start); | 	pmd = pmd_offset(pud, start); | ||||||
| 	pud_clear(pud); | 	pud_clear(pud); | ||||||
| 	pmd_free_tlb(tlb, pmd, start); | 	pmd_free_tlb(tlb, pmd, start); | ||||||
|  | 	mm_dec_nr_pmds(tlb->mm); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| static inline void free_pud_range(struct mmu_gather *tlb, pgd_t *pgd, | static inline void free_pud_range(struct mmu_gather *tlb, pgd_t *pgd, | ||||||
|  | @ -3322,15 +3323,17 @@ int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address) | ||||||
| 
 | 
 | ||||||
| 	spin_lock(&mm->page_table_lock); | 	spin_lock(&mm->page_table_lock); | ||||||
| #ifndef __ARCH_HAS_4LEVEL_HACK | #ifndef __ARCH_HAS_4LEVEL_HACK | ||||||
| 	if (pud_present(*pud))		/* Another has populated it */ | 	if (!pud_present(*pud)) { | ||||||
| 		pmd_free(mm, new); | 		mm_inc_nr_pmds(mm); | ||||||
| 	else |  | ||||||
| 		pud_populate(mm, pud, new); | 		pud_populate(mm, pud, new); | ||||||
| #else | 	} else	/* Another has populated it */ | ||||||
| 	if (pgd_present(*pud))		/* Another has populated it */ |  | ||||||
| 		pmd_free(mm, new); | 		pmd_free(mm, new); | ||||||
| 	else | #else | ||||||
|  | 	if (!pgd_present(*pud)) { | ||||||
|  | 		mm_inc_nr_pmds(mm); | ||||||
| 		pgd_populate(mm, pud, new); | 		pgd_populate(mm, pud, new); | ||||||
|  | 	} else /* Another has populated it */ | ||||||
|  | 		pmd_free(mm, new); | ||||||
| #endif /* __ARCH_HAS_4LEVEL_HACK */ | #endif /* __ARCH_HAS_4LEVEL_HACK */ | ||||||
| 	spin_unlock(&mm->page_table_lock); | 	spin_unlock(&mm->page_table_lock); | ||||||
| 	return 0; | 	return 0; | ||||||
|  |  | ||||||
|  | @ -2853,7 +2853,9 @@ void exit_mmap(struct mm_struct *mm) | ||||||
| 	vm_unacct_memory(nr_accounted); | 	vm_unacct_memory(nr_accounted); | ||||||
| 
 | 
 | ||||||
| 	WARN_ON(atomic_long_read(&mm->nr_ptes) > | 	WARN_ON(atomic_long_read(&mm->nr_ptes) > | ||||||
| 			(FIRST_USER_ADDRESS+PMD_SIZE-1)>>PMD_SHIFT); | 			round_up(FIRST_USER_ADDRESS, PMD_SIZE) >> PMD_SHIFT); | ||||||
|  | 	WARN_ON(mm_nr_pmds(mm) > | ||||||
|  | 			round_up(FIRST_USER_ADDRESS, PUD_SIZE) >> PUD_SHIFT); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| /* Insert vm structure into process list sorted by address
 | /* Insert vm structure into process list sorted by address
 | ||||||
|  |  | ||||||
|  | @ -169,8 +169,8 @@ unsigned long oom_badness(struct task_struct *p, struct mem_cgroup *memcg, | ||||||
| 	 * The baseline for the badness score is the proportion of RAM that each | 	 * The baseline for the badness score is the proportion of RAM that each | ||||||
| 	 * task's rss, pagetable and swap space use. | 	 * task's rss, pagetable and swap space use. | ||||||
| 	 */ | 	 */ | ||||||
| 	points = get_mm_rss(p->mm) + atomic_long_read(&p->mm->nr_ptes) + | 	points = get_mm_rss(p->mm) + get_mm_counter(p->mm, MM_SWAPENTS) + | ||||||
| 		 get_mm_counter(p->mm, MM_SWAPENTS); | 		atomic_long_read(&p->mm->nr_ptes) + mm_nr_pmds(p->mm); | ||||||
| 	task_unlock(p); | 	task_unlock(p); | ||||||
| 
 | 
 | ||||||
| 	/*
 | 	/*
 | ||||||
|  | @ -351,7 +351,7 @@ static void dump_tasks(struct mem_cgroup *memcg, const nodemask_t *nodemask) | ||||||
| 	struct task_struct *p; | 	struct task_struct *p; | ||||||
| 	struct task_struct *task; | 	struct task_struct *task; | ||||||
| 
 | 
 | ||||||
| 	pr_info("[ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name\n"); | 	pr_info("[ pid ]   uid  tgid total_vm      rss nr_ptes nr_pmds swapents oom_score_adj name\n"); | ||||||
| 	rcu_read_lock(); | 	rcu_read_lock(); | ||||||
| 	for_each_process(p) { | 	for_each_process(p) { | ||||||
| 		if (oom_unkillable_task(p, memcg, nodemask)) | 		if (oom_unkillable_task(p, memcg, nodemask)) | ||||||
|  | @ -367,10 +367,11 @@ static void dump_tasks(struct mem_cgroup *memcg, const nodemask_t *nodemask) | ||||||
| 			continue; | 			continue; | ||||||
| 		} | 		} | ||||||
| 
 | 
 | ||||||
| 		pr_info("[%5d] %5d %5d %8lu %8lu %7ld %8lu         %5hd %s\n", | 		pr_info("[%5d] %5d %5d %8lu %8lu %7ld %7ld %8lu         %5hd %s\n", | ||||||
| 			task->pid, from_kuid(&init_user_ns, task_uid(task)), | 			task->pid, from_kuid(&init_user_ns, task_uid(task)), | ||||||
| 			task->tgid, task->mm->total_vm, get_mm_rss(task->mm), | 			task->tgid, task->mm->total_vm, get_mm_rss(task->mm), | ||||||
| 			atomic_long_read(&task->mm->nr_ptes), | 			atomic_long_read(&task->mm->nr_ptes), | ||||||
|  | 			mm_nr_pmds(task->mm), | ||||||
| 			get_mm_counter(task->mm, MM_SWAPENTS), | 			get_mm_counter(task->mm, MM_SWAPENTS), | ||||||
| 			task->signal->oom_score_adj, task->comm); | 			task->signal->oom_score_adj, task->comm); | ||||||
| 		task_unlock(task); | 		task_unlock(task); | ||||||
|  |  | ||||||
		Loading…
	
		Reference in a new issue
	
	 Kirill A. Shutemov
						Kirill A. Shutemov