mirror of
https://github.com/torvalds/linux.git
synced 2025-11-12 22:49:37 +02:00
documented (hopefully adequately) in the respective changelogs. Notable
series include:
- Lucas Stach has provided some page-mapping
cleanup/consolidation/maintainability work in the series "mm/treewide:
Remove pXd_huge() API".
- In the series "Allow migrate on protnone reference with
MPOL_PREFERRED_MANY policy", Donet Tom has optimized mempolicy's
MPOL_PREFERRED_MANY mode, yielding almost doubled performance in one
test.
- In their series "Memory allocation profiling" Kent Overstreet and
Suren Baghdasaryan have contributed a means of determining (via
/proc/allocinfo) whereabouts in the kernel memory is being allocated:
number of calls and amount of memory.
- Matthew Wilcox has provided the series "Various significant MM
patches" which does a number of rather unrelated things, but in largely
similar code sites.
- In his series "mm: page_alloc: freelist migratetype hygiene" Johannes
Weiner has fixed the page allocator's handling of migratetype requests,
with resulting improvements in compaction efficiency.
- In the series "make the hugetlb migration strategy consistent" Baolin
Wang has fixed a hugetlb migration issue, which should improve hugetlb
allocation reliability.
- Liu Shixin has hit an I/O meltdown caused by readahead in a
memory-tight memcg. Addressed in the series "Fix I/O high when memory
almost met memcg limit".
- In the series "mm/filemap: optimize folio adding and splitting" Kairui
Song has optimized pagecache insertion, yielding ~10% performance
improvement in one test.
- Baoquan He has cleaned up and consolidated the early zone
initialization code in the series "mm/mm_init.c: refactor
free_area_init_core()".
- Baoquan has also redone some MM initializatio code in the series
"mm/init: minor clean up and improvement".
- MM helper cleanups from Christoph Hellwig in his series "remove
follow_pfn".
- More cleanups from Matthew Wilcox in the series "Various page->flags
cleanups".
- Vlastimil Babka has contributed maintainability improvements in the
series "memcg_kmem hooks refactoring".
- More folio conversions and cleanups in Matthew Wilcox's series
"Convert huge_zero_page to huge_zero_folio"
"khugepaged folio conversions"
"Remove page_idle and page_young wrappers"
"Use folio APIs in procfs"
"Clean up __folio_put()"
"Some cleanups for memory-failure"
"Remove page_mapping()"
"More folio compat code removal"
- David Hildenbrand chipped in with "fs/proc/task_mmu: convert hugetlb
functions to work on folis".
- Code consolidation and cleanup work related to GUP's handling of
hugetlbs in Peter Xu's series "mm/gup: Unify hugetlb, part 2".
- Rick Edgecombe has developed some fixes to stack guard gaps in the
series "Cover a guard gap corner case".
- Jinjiang Tu has fixed KSM's behaviour after a fork+exec in the series
"mm/ksm: fix ksm exec support for prctl".
- Baolin Wang has implemented NUMA balancing for multi-size THPs. This
is a simple first-cut implementation for now. The series is "support
multi-size THP numa balancing".
- Cleanups to vma handling helper functions from Matthew Wilcox in the
series "Unify vma_address and vma_pgoff_address".
- Some selftests maintenance work from Dev Jain in the series
"selftests/mm: mremap_test: Optimizations and style fixes".
- Improvements to the swapping of multi-size THPs from Ryan Roberts in
the series "Swap-out mTHP without splitting".
- Kefeng Wang has significantly optimized the handling of arm64's
permission page faults in the series
"arch/mm/fault: accelerate pagefault when badaccess"
"mm: remove arch's private VM_FAULT_BADMAP/BADACCESS"
- GUP cleanups from David Hildenbrand in "mm/gup: consistently call it
GUP-fast".
- hugetlb fault code cleanups from Vishal Moola in "Hugetlb fault path to
use struct vm_fault".
- selftests build fixes from John Hubbard in the series "Fix
selftests/mm build without requiring "make headers"".
- Memory tiering fixes/improvements from Ho-Ren (Jack) Chuang in the
series "Improved Memory Tier Creation for CPUless NUMA Nodes". Fixes
the initialization code so that migration between different memory types
works as intended.
- David Hildenbrand has improved follow_pte() and fixed an errant driver
in the series "mm: follow_pte() improvements and acrn follow_pte()
fixes".
- David also did some cleanup work on large folio mapcounts in his
series "mm: mapcount for large folios + page_mapcount() cleanups".
- Folio conversions in KSM in Alex Shi's series "transfer page to folio
in KSM".
- Barry Song has added some sysfs stats for monitoring multi-size THP's
in the series "mm: add per-order mTHP alloc and swpout counters".
- Some zswap cleanups from Yosry Ahmed in the series "zswap same-filled
and limit checking cleanups".
- Matthew Wilcox has been looking at buffer_head code and found the
documentation to be lacking. The series is "Improve buffer head
documentation".
- Multi-size THPs get more work, this time from Lance Yang. His series
"mm/madvise: enhance lazyfreeing with mTHP in madvise_free" optimizes
the freeing of these things.
- Kemeng Shi has added more userspace-visible writeback instrumentation
in the series "Improve visibility of writeback".
- Kemeng Shi then sent some maintenance work on top in the series "Fix
and cleanups to page-writeback".
- Matthew Wilcox reduces mmap_lock traffic in the anon vma code in the
series "Improve anon_vma scalability for anon VMAs". Intel's test bot
reported an improbable 3x improvement in one test.
- SeongJae Park adds some DAMON feature work in the series
"mm/damon: add a DAMOS filter type for page granularity access recheck"
"selftests/damon: add DAMOS quota goal test"
- Also some maintenance work in the series
"mm/damon/paddr: simplify page level access re-check for pageout"
"mm/damon: misc fixes and improvements"
- David Hildenbrand has disabled some known-to-fail selftests ni the
series "selftests: mm: cow: flag vmsplice() hugetlb tests as XFAIL".
- memcg metadata storage optimizations from Shakeel Butt in "memcg:
reduce memory consumption by memcg stats".
- DAX fixes and maintenance work from Vishal Verma in the series
"dax/bus.c: Fixups for dax-bus locking".
-----BEGIN PGP SIGNATURE-----
iHUEABYIAB0WIQTTMBEPP41GrTpTJgfdBJ7gKXxAjgUCZkgQYwAKCRDdBJ7gKXxA
jrdKAP9WVJdpEcXxpoub/vVE0UWGtffr8foifi9bCwrQrGh5mgEAx7Yf0+d/oBZB
nvA4E0DcPrUAFy144FNM0NTCb7u9vAw=
=V3R/
-----END PGP SIGNATURE-----
Merge tag 'mm-stable-2024-05-17-19-19' of git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm
Pull mm updates from Andrew Morton:
"The usual shower of singleton fixes and minor series all over MM,
documented (hopefully adequately) in the respective changelogs.
Notable series include:
- Lucas Stach has provided some page-mapping cleanup/consolidation/
maintainability work in the series "mm/treewide: Remove pXd_huge()
API".
- In the series "Allow migrate on protnone reference with
MPOL_PREFERRED_MANY policy", Donet Tom has optimized mempolicy's
MPOL_PREFERRED_MANY mode, yielding almost doubled performance in
one test.
- In their series "Memory allocation profiling" Kent Overstreet and
Suren Baghdasaryan have contributed a means of determining (via
/proc/allocinfo) whereabouts in the kernel memory is being
allocated: number of calls and amount of memory.
- Matthew Wilcox has provided the series "Various significant MM
patches" which does a number of rather unrelated things, but in
largely similar code sites.
- In his series "mm: page_alloc: freelist migratetype hygiene"
Johannes Weiner has fixed the page allocator's handling of
migratetype requests, with resulting improvements in compaction
efficiency.
- In the series "make the hugetlb migration strategy consistent"
Baolin Wang has fixed a hugetlb migration issue, which should
improve hugetlb allocation reliability.
- Liu Shixin has hit an I/O meltdown caused by readahead in a
memory-tight memcg. Addressed in the series "Fix I/O high when
memory almost met memcg limit".
- In the series "mm/filemap: optimize folio adding and splitting"
Kairui Song has optimized pagecache insertion, yielding ~10%
performance improvement in one test.
- Baoquan He has cleaned up and consolidated the early zone
initialization code in the series "mm/mm_init.c: refactor
free_area_init_core()".
- Baoquan has also redone some MM initializatio code in the series
"mm/init: minor clean up and improvement".
- MM helper cleanups from Christoph Hellwig in his series "remove
follow_pfn".
- More cleanups from Matthew Wilcox in the series "Various
page->flags cleanups".
- Vlastimil Babka has contributed maintainability improvements in the
series "memcg_kmem hooks refactoring".
- More folio conversions and cleanups in Matthew Wilcox's series:
"Convert huge_zero_page to huge_zero_folio"
"khugepaged folio conversions"
"Remove page_idle and page_young wrappers"
"Use folio APIs in procfs"
"Clean up __folio_put()"
"Some cleanups for memory-failure"
"Remove page_mapping()"
"More folio compat code removal"
- David Hildenbrand chipped in with "fs/proc/task_mmu: convert
hugetlb functions to work on folis".
- Code consolidation and cleanup work related to GUP's handling of
hugetlbs in Peter Xu's series "mm/gup: Unify hugetlb, part 2".
- Rick Edgecombe has developed some fixes to stack guard gaps in the
series "Cover a guard gap corner case".
- Jinjiang Tu has fixed KSM's behaviour after a fork+exec in the
series "mm/ksm: fix ksm exec support for prctl".
- Baolin Wang has implemented NUMA balancing for multi-size THPs.
This is a simple first-cut implementation for now. The series is
"support multi-size THP numa balancing".
- Cleanups to vma handling helper functions from Matthew Wilcox in
the series "Unify vma_address and vma_pgoff_address".
- Some selftests maintenance work from Dev Jain in the series
"selftests/mm: mremap_test: Optimizations and style fixes".
- Improvements to the swapping of multi-size THPs from Ryan Roberts
in the series "Swap-out mTHP without splitting".
- Kefeng Wang has significantly optimized the handling of arm64's
permission page faults in the series
"arch/mm/fault: accelerate pagefault when badaccess"
"mm: remove arch's private VM_FAULT_BADMAP/BADACCESS"
- GUP cleanups from David Hildenbrand in "mm/gup: consistently call
it GUP-fast".
- hugetlb fault code cleanups from Vishal Moola in "Hugetlb fault
path to use struct vm_fault".
- selftests build fixes from John Hubbard in the series "Fix
selftests/mm build without requiring "make headers"".
- Memory tiering fixes/improvements from Ho-Ren (Jack) Chuang in the
series "Improved Memory Tier Creation for CPUless NUMA Nodes".
Fixes the initialization code so that migration between different
memory types works as intended.
- David Hildenbrand has improved follow_pte() and fixed an errant
driver in the series "mm: follow_pte() improvements and acrn
follow_pte() fixes".
- David also did some cleanup work on large folio mapcounts in his
series "mm: mapcount for large folios + page_mapcount() cleanups".
- Folio conversions in KSM in Alex Shi's series "transfer page to
folio in KSM".
- Barry Song has added some sysfs stats for monitoring multi-size
THP's in the series "mm: add per-order mTHP alloc and swpout
counters".
- Some zswap cleanups from Yosry Ahmed in the series "zswap
same-filled and limit checking cleanups".
- Matthew Wilcox has been looking at buffer_head code and found the
documentation to be lacking. The series is "Improve buffer head
documentation".
- Multi-size THPs get more work, this time from Lance Yang. His
series "mm/madvise: enhance lazyfreeing with mTHP in madvise_free"
optimizes the freeing of these things.
- Kemeng Shi has added more userspace-visible writeback
instrumentation in the series "Improve visibility of writeback".
- Kemeng Shi then sent some maintenance work on top in the series
"Fix and cleanups to page-writeback".
- Matthew Wilcox reduces mmap_lock traffic in the anon vma code in
the series "Improve anon_vma scalability for anon VMAs". Intel's
test bot reported an improbable 3x improvement in one test.
- SeongJae Park adds some DAMON feature work in the series
"mm/damon: add a DAMOS filter type for page granularity access recheck"
"selftests/damon: add DAMOS quota goal test"
- Also some maintenance work in the series
"mm/damon/paddr: simplify page level access re-check for pageout"
"mm/damon: misc fixes and improvements"
- David Hildenbrand has disabled some known-to-fail selftests ni the
series "selftests: mm: cow: flag vmsplice() hugetlb tests as
XFAIL".
- memcg metadata storage optimizations from Shakeel Butt in "memcg:
reduce memory consumption by memcg stats".
- DAX fixes and maintenance work from Vishal Verma in the series
"dax/bus.c: Fixups for dax-bus locking""
* tag 'mm-stable-2024-05-17-19-19' of git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm: (426 commits)
memcg, oom: cleanup unused memcg_oom_gfp_mask and memcg_oom_order
selftests/mm: hugetlb_madv_vs_map: avoid test skipping by querying hugepage size at runtime
mm/hugetlb: add missing VM_FAULT_SET_HINDEX in hugetlb_wp
mm/hugetlb: add missing VM_FAULT_SET_HINDEX in hugetlb_fault
selftests: cgroup: add tests to verify the zswap writeback path
mm: memcg: make alloc_mem_cgroup_per_node_info() return bool
mm/damon/core: fix return value from damos_wmark_metric_value
mm: do not update memcg stats for NR_{FILE/SHMEM}_PMDMAPPED
selftests: cgroup: remove redundant enabling of memory controller
Docs/mm/damon/maintainer-profile: allow posting patches based on damon/next tree
Docs/mm/damon/maintainer-profile: change the maintainer's timezone from PST to PT
Docs/mm/damon/design: use a list for supported filters
Docs/admin-guide/mm/damon/usage: fix wrong schemes effective quota update command
Docs/admin-guide/mm/damon/usage: fix wrong example of DAMOS filter matching sysfs file
selftests/damon: classify tests for functionalities and regressions
selftests/damon/_damon_sysfs: use 'is' instead of '==' for 'None'
selftests/damon/_damon_sysfs: find sysfs mount point from /proc/mounts
selftests/damon/_damon_sysfs: check errors from nr_schemes file reads
mm/damon/core: initialize ->esz_bp from damos_quota_init_priv()
selftests/damon: add a test for DAMOS quota goal
...
1819 lines
44 KiB
C
1819 lines
44 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* COW (Copy On Write) tests.
|
|
*
|
|
* Copyright 2022, Red Hat, Inc.
|
|
*
|
|
* Author(s): David Hildenbrand <david@redhat.com>
|
|
*/
|
|
#define _GNU_SOURCE
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include <stdbool.h>
|
|
#include <stdint.h>
|
|
#include <unistd.h>
|
|
#include <errno.h>
|
|
#include <fcntl.h>
|
|
#include <assert.h>
|
|
#include <linux/mman.h>
|
|
#include <sys/mman.h>
|
|
#include <sys/ioctl.h>
|
|
#include <sys/wait.h>
|
|
#include <linux/memfd.h>
|
|
|
|
#include "local_config.h"
|
|
#ifdef LOCAL_CONFIG_HAVE_LIBURING
|
|
#include <liburing.h>
|
|
#endif /* LOCAL_CONFIG_HAVE_LIBURING */
|
|
|
|
#include "../../../../mm/gup_test.h"
|
|
#include "../kselftest.h"
|
|
#include "vm_util.h"
|
|
#include "thp_settings.h"
|
|
|
|
static size_t pagesize;
|
|
static int pagemap_fd;
|
|
static size_t pmdsize;
|
|
static int nr_thpsizes;
|
|
static size_t thpsizes[20];
|
|
static int nr_hugetlbsizes;
|
|
static size_t hugetlbsizes[10];
|
|
static int gup_fd;
|
|
static bool has_huge_zeropage;
|
|
|
|
static int sz2ord(size_t size)
|
|
{
|
|
return __builtin_ctzll(size / pagesize);
|
|
}
|
|
|
|
static int detect_thp_sizes(size_t sizes[], int max)
|
|
{
|
|
int count = 0;
|
|
unsigned long orders;
|
|
size_t kb;
|
|
int i;
|
|
|
|
/* thp not supported at all. */
|
|
if (!pmdsize)
|
|
return 0;
|
|
|
|
orders = 1UL << sz2ord(pmdsize);
|
|
orders |= thp_supported_orders();
|
|
|
|
for (i = 0; orders && count < max; i++) {
|
|
if (!(orders & (1UL << i)))
|
|
continue;
|
|
orders &= ~(1UL << i);
|
|
kb = (pagesize >> 10) << i;
|
|
sizes[count++] = kb * 1024;
|
|
ksft_print_msg("[INFO] detected THP size: %zu KiB\n", kb);
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
static void detect_huge_zeropage(void)
|
|
{
|
|
int fd = open("/sys/kernel/mm/transparent_hugepage/use_zero_page",
|
|
O_RDONLY);
|
|
size_t enabled = 0;
|
|
char buf[15];
|
|
int ret;
|
|
|
|
if (fd < 0)
|
|
return;
|
|
|
|
ret = pread(fd, buf, sizeof(buf), 0);
|
|
if (ret > 0 && ret < sizeof(buf)) {
|
|
buf[ret] = 0;
|
|
|
|
enabled = strtoul(buf, NULL, 10);
|
|
if (enabled == 1) {
|
|
has_huge_zeropage = true;
|
|
ksft_print_msg("[INFO] huge zeropage is enabled\n");
|
|
}
|
|
}
|
|
|
|
close(fd);
|
|
}
|
|
|
|
static bool range_is_swapped(void *addr, size_t size)
|
|
{
|
|
for (; size; addr += pagesize, size -= pagesize)
|
|
if (!pagemap_is_swapped(pagemap_fd, addr))
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
struct comm_pipes {
|
|
int child_ready[2];
|
|
int parent_ready[2];
|
|
};
|
|
|
|
static int setup_comm_pipes(struct comm_pipes *comm_pipes)
|
|
{
|
|
if (pipe(comm_pipes->child_ready) < 0)
|
|
return -errno;
|
|
if (pipe(comm_pipes->parent_ready) < 0) {
|
|
close(comm_pipes->child_ready[0]);
|
|
close(comm_pipes->child_ready[1]);
|
|
return -errno;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void close_comm_pipes(struct comm_pipes *comm_pipes)
|
|
{
|
|
close(comm_pipes->child_ready[0]);
|
|
close(comm_pipes->child_ready[1]);
|
|
close(comm_pipes->parent_ready[0]);
|
|
close(comm_pipes->parent_ready[1]);
|
|
}
|
|
|
|
static int child_memcmp_fn(char *mem, size_t size,
|
|
struct comm_pipes *comm_pipes)
|
|
{
|
|
char *old = malloc(size);
|
|
char buf;
|
|
|
|
/* Backup the original content. */
|
|
memcpy(old, mem, size);
|
|
|
|
/* Wait until the parent modified the page. */
|
|
write(comm_pipes->child_ready[1], "0", 1);
|
|
while (read(comm_pipes->parent_ready[0], &buf, 1) != 1)
|
|
;
|
|
|
|
/* See if we still read the old values. */
|
|
return memcmp(old, mem, size);
|
|
}
|
|
|
|
static int child_vmsplice_memcmp_fn(char *mem, size_t size,
|
|
struct comm_pipes *comm_pipes)
|
|
{
|
|
struct iovec iov = {
|
|
.iov_base = mem,
|
|
.iov_len = size,
|
|
};
|
|
ssize_t cur, total, transferred;
|
|
char *old, *new;
|
|
int fds[2];
|
|
char buf;
|
|
|
|
old = malloc(size);
|
|
new = malloc(size);
|
|
|
|
/* Backup the original content. */
|
|
memcpy(old, mem, size);
|
|
|
|
if (pipe(fds) < 0)
|
|
return -errno;
|
|
|
|
/* Trigger a read-only pin. */
|
|
transferred = vmsplice(fds[1], &iov, 1, 0);
|
|
if (transferred < 0)
|
|
return -errno;
|
|
if (transferred == 0)
|
|
return -EINVAL;
|
|
|
|
/* Unmap it from our page tables. */
|
|
if (munmap(mem, size) < 0)
|
|
return -errno;
|
|
|
|
/* Wait until the parent modified it. */
|
|
write(comm_pipes->child_ready[1], "0", 1);
|
|
while (read(comm_pipes->parent_ready[0], &buf, 1) != 1)
|
|
;
|
|
|
|
/* See if we still read the old values via the pipe. */
|
|
for (total = 0; total < transferred; total += cur) {
|
|
cur = read(fds[0], new + total, transferred - total);
|
|
if (cur < 0)
|
|
return -errno;
|
|
}
|
|
|
|
return memcmp(old, new, transferred);
|
|
}
|
|
|
|
typedef int (*child_fn)(char *mem, size_t size, struct comm_pipes *comm_pipes);
|
|
|
|
static void do_test_cow_in_parent(char *mem, size_t size, bool do_mprotect,
|
|
child_fn fn, bool xfail)
|
|
{
|
|
struct comm_pipes comm_pipes;
|
|
char buf;
|
|
int ret;
|
|
|
|
ret = setup_comm_pipes(&comm_pipes);
|
|
if (ret) {
|
|
ksft_test_result_fail("pipe() failed\n");
|
|
return;
|
|
}
|
|
|
|
ret = fork();
|
|
if (ret < 0) {
|
|
ksft_test_result_fail("fork() failed\n");
|
|
goto close_comm_pipes;
|
|
} else if (!ret) {
|
|
exit(fn(mem, size, &comm_pipes));
|
|
}
|
|
|
|
while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
|
|
;
|
|
|
|
if (do_mprotect) {
|
|
/*
|
|
* mprotect() optimizations might try avoiding
|
|
* write-faults by directly mapping pages writable.
|
|
*/
|
|
ret = mprotect(mem, size, PROT_READ);
|
|
ret |= mprotect(mem, size, PROT_READ|PROT_WRITE);
|
|
if (ret) {
|
|
ksft_test_result_fail("mprotect() failed\n");
|
|
write(comm_pipes.parent_ready[1], "0", 1);
|
|
wait(&ret);
|
|
goto close_comm_pipes;
|
|
}
|
|
}
|
|
|
|
/* Modify the page. */
|
|
memset(mem, 0xff, size);
|
|
write(comm_pipes.parent_ready[1], "0", 1);
|
|
|
|
wait(&ret);
|
|
if (WIFEXITED(ret))
|
|
ret = WEXITSTATUS(ret);
|
|
else
|
|
ret = -EINVAL;
|
|
|
|
if (!ret) {
|
|
ksft_test_result_pass("No leak from parent into child\n");
|
|
} else if (xfail) {
|
|
/*
|
|
* With hugetlb, some vmsplice() tests are currently expected to
|
|
* fail because (a) harder to fix and (b) nobody really cares.
|
|
* Flag them as expected failure for now.
|
|
*/
|
|
ksft_test_result_xfail("Leak from parent into child\n");
|
|
} else {
|
|
ksft_test_result_fail("Leak from parent into child\n");
|
|
}
|
|
close_comm_pipes:
|
|
close_comm_pipes(&comm_pipes);
|
|
}
|
|
|
|
static void test_cow_in_parent(char *mem, size_t size, bool is_hugetlb)
|
|
{
|
|
do_test_cow_in_parent(mem, size, false, child_memcmp_fn, false);
|
|
}
|
|
|
|
static void test_cow_in_parent_mprotect(char *mem, size_t size, bool is_hugetlb)
|
|
{
|
|
do_test_cow_in_parent(mem, size, true, child_memcmp_fn, false);
|
|
}
|
|
|
|
static void test_vmsplice_in_child(char *mem, size_t size, bool is_hugetlb)
|
|
{
|
|
do_test_cow_in_parent(mem, size, false, child_vmsplice_memcmp_fn,
|
|
is_hugetlb);
|
|
}
|
|
|
|
static void test_vmsplice_in_child_mprotect(char *mem, size_t size,
|
|
bool is_hugetlb)
|
|
{
|
|
do_test_cow_in_parent(mem, size, true, child_vmsplice_memcmp_fn,
|
|
is_hugetlb);
|
|
}
|
|
|
|
static void do_test_vmsplice_in_parent(char *mem, size_t size,
|
|
bool before_fork, bool xfail)
|
|
{
|
|
struct iovec iov = {
|
|
.iov_base = mem,
|
|
.iov_len = size,
|
|
};
|
|
ssize_t cur, total, transferred;
|
|
struct comm_pipes comm_pipes;
|
|
char *old, *new;
|
|
int ret, fds[2];
|
|
char buf;
|
|
|
|
old = malloc(size);
|
|
new = malloc(size);
|
|
|
|
memcpy(old, mem, size);
|
|
|
|
ret = setup_comm_pipes(&comm_pipes);
|
|
if (ret) {
|
|
ksft_test_result_fail("pipe() failed\n");
|
|
goto free;
|
|
}
|
|
|
|
if (pipe(fds) < 0) {
|
|
ksft_test_result_fail("pipe() failed\n");
|
|
goto close_comm_pipes;
|
|
}
|
|
|
|
if (before_fork) {
|
|
transferred = vmsplice(fds[1], &iov, 1, 0);
|
|
if (transferred <= 0) {
|
|
ksft_test_result_fail("vmsplice() failed\n");
|
|
goto close_pipe;
|
|
}
|
|
}
|
|
|
|
ret = fork();
|
|
if (ret < 0) {
|
|
ksft_test_result_fail("fork() failed\n");
|
|
goto close_pipe;
|
|
} else if (!ret) {
|
|
write(comm_pipes.child_ready[1], "0", 1);
|
|
while (read(comm_pipes.parent_ready[0], &buf, 1) != 1)
|
|
;
|
|
/* Modify page content in the child. */
|
|
memset(mem, 0xff, size);
|
|
exit(0);
|
|
}
|
|
|
|
if (!before_fork) {
|
|
transferred = vmsplice(fds[1], &iov, 1, 0);
|
|
if (transferred <= 0) {
|
|
ksft_test_result_fail("vmsplice() failed\n");
|
|
wait(&ret);
|
|
goto close_pipe;
|
|
}
|
|
}
|
|
|
|
while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
|
|
;
|
|
if (munmap(mem, size) < 0) {
|
|
ksft_test_result_fail("munmap() failed\n");
|
|
goto close_pipe;
|
|
}
|
|
write(comm_pipes.parent_ready[1], "0", 1);
|
|
|
|
/* Wait until the child is done writing. */
|
|
wait(&ret);
|
|
if (!WIFEXITED(ret)) {
|
|
ksft_test_result_fail("wait() failed\n");
|
|
goto close_pipe;
|
|
}
|
|
|
|
/* See if we still read the old values. */
|
|
for (total = 0; total < transferred; total += cur) {
|
|
cur = read(fds[0], new + total, transferred - total);
|
|
if (cur < 0) {
|
|
ksft_test_result_fail("read() failed\n");
|
|
goto close_pipe;
|
|
}
|
|
}
|
|
|
|
if (!memcmp(old, new, transferred)) {
|
|
ksft_test_result_pass("No leak from child into parent\n");
|
|
} else if (xfail) {
|
|
/*
|
|
* With hugetlb, some vmsplice() tests are currently expected to
|
|
* fail because (a) harder to fix and (b) nobody really cares.
|
|
* Flag them as expected failure for now.
|
|
*/
|
|
ksft_test_result_xfail("Leak from child into parent\n");
|
|
} else {
|
|
ksft_test_result_fail("Leak from child into parent\n");
|
|
}
|
|
close_pipe:
|
|
close(fds[0]);
|
|
close(fds[1]);
|
|
close_comm_pipes:
|
|
close_comm_pipes(&comm_pipes);
|
|
free:
|
|
free(old);
|
|
free(new);
|
|
}
|
|
|
|
static void test_vmsplice_before_fork(char *mem, size_t size, bool is_hugetlb)
|
|
{
|
|
do_test_vmsplice_in_parent(mem, size, true, is_hugetlb);
|
|
}
|
|
|
|
static void test_vmsplice_after_fork(char *mem, size_t size, bool is_hugetlb)
|
|
{
|
|
do_test_vmsplice_in_parent(mem, size, false, is_hugetlb);
|
|
}
|
|
|
|
#ifdef LOCAL_CONFIG_HAVE_LIBURING
|
|
static void do_test_iouring(char *mem, size_t size, bool use_fork)
|
|
{
|
|
struct comm_pipes comm_pipes;
|
|
struct io_uring_cqe *cqe;
|
|
struct io_uring_sqe *sqe;
|
|
struct io_uring ring;
|
|
ssize_t cur, total;
|
|
struct iovec iov;
|
|
char *buf, *tmp;
|
|
int ret, fd;
|
|
FILE *file;
|
|
|
|
ret = setup_comm_pipes(&comm_pipes);
|
|
if (ret) {
|
|
ksft_test_result_fail("pipe() failed\n");
|
|
return;
|
|
}
|
|
|
|
file = tmpfile();
|
|
if (!file) {
|
|
ksft_test_result_fail("tmpfile() failed\n");
|
|
goto close_comm_pipes;
|
|
}
|
|
fd = fileno(file);
|
|
assert(fd);
|
|
|
|
tmp = malloc(size);
|
|
if (!tmp) {
|
|
ksft_test_result_fail("malloc() failed\n");
|
|
goto close_file;
|
|
}
|
|
|
|
/* Skip on errors, as we might just lack kernel support. */
|
|
ret = io_uring_queue_init(1, &ring, 0);
|
|
if (ret < 0) {
|
|
ksft_test_result_skip("io_uring_queue_init() failed\n");
|
|
goto free_tmp;
|
|
}
|
|
|
|
/*
|
|
* Register the range as a fixed buffer. This will FOLL_WRITE | FOLL_PIN
|
|
* | FOLL_LONGTERM the range.
|
|
*
|
|
* Skip on errors, as we might just lack kernel support or might not
|
|
* have sufficient MEMLOCK permissions.
|
|
*/
|
|
iov.iov_base = mem;
|
|
iov.iov_len = size;
|
|
ret = io_uring_register_buffers(&ring, &iov, 1);
|
|
if (ret) {
|
|
ksft_test_result_skip("io_uring_register_buffers() failed\n");
|
|
goto queue_exit;
|
|
}
|
|
|
|
if (use_fork) {
|
|
/*
|
|
* fork() and keep the child alive until we're done. Note that
|
|
* we expect the pinned page to not get shared with the child.
|
|
*/
|
|
ret = fork();
|
|
if (ret < 0) {
|
|
ksft_test_result_fail("fork() failed\n");
|
|
goto unregister_buffers;
|
|
} else if (!ret) {
|
|
write(comm_pipes.child_ready[1], "0", 1);
|
|
while (read(comm_pipes.parent_ready[0], &buf, 1) != 1)
|
|
;
|
|
exit(0);
|
|
}
|
|
|
|
while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
|
|
;
|
|
} else {
|
|
/*
|
|
* Map the page R/O into the page table. Enable softdirty
|
|
* tracking to stop the page from getting mapped R/W immediately
|
|
* again by mprotect() optimizations. Note that we don't have an
|
|
* easy way to test if that worked (the pagemap does not export
|
|
* if the page is mapped R/O vs. R/W).
|
|
*/
|
|
ret = mprotect(mem, size, PROT_READ);
|
|
clear_softdirty();
|
|
ret |= mprotect(mem, size, PROT_READ | PROT_WRITE);
|
|
if (ret) {
|
|
ksft_test_result_fail("mprotect() failed\n");
|
|
goto unregister_buffers;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Modify the page and write page content as observed by the fixed
|
|
* buffer pin to the file so we can verify it.
|
|
*/
|
|
memset(mem, 0xff, size);
|
|
sqe = io_uring_get_sqe(&ring);
|
|
if (!sqe) {
|
|
ksft_test_result_fail("io_uring_get_sqe() failed\n");
|
|
goto quit_child;
|
|
}
|
|
io_uring_prep_write_fixed(sqe, fd, mem, size, 0, 0);
|
|
|
|
ret = io_uring_submit(&ring);
|
|
if (ret < 0) {
|
|
ksft_test_result_fail("io_uring_submit() failed\n");
|
|
goto quit_child;
|
|
}
|
|
|
|
ret = io_uring_wait_cqe(&ring, &cqe);
|
|
if (ret < 0) {
|
|
ksft_test_result_fail("io_uring_wait_cqe() failed\n");
|
|
goto quit_child;
|
|
}
|
|
|
|
if (cqe->res != size) {
|
|
ksft_test_result_fail("write_fixed failed\n");
|
|
goto quit_child;
|
|
}
|
|
io_uring_cqe_seen(&ring, cqe);
|
|
|
|
/* Read back the file content to the temporary buffer. */
|
|
total = 0;
|
|
while (total < size) {
|
|
cur = pread(fd, tmp + total, size - total, total);
|
|
if (cur < 0) {
|
|
ksft_test_result_fail("pread() failed\n");
|
|
goto quit_child;
|
|
}
|
|
total += cur;
|
|
}
|
|
|
|
/* Finally, check if we read what we expected. */
|
|
ksft_test_result(!memcmp(mem, tmp, size),
|
|
"Longterm R/W pin is reliable\n");
|
|
|
|
quit_child:
|
|
if (use_fork) {
|
|
write(comm_pipes.parent_ready[1], "0", 1);
|
|
wait(&ret);
|
|
}
|
|
unregister_buffers:
|
|
io_uring_unregister_buffers(&ring);
|
|
queue_exit:
|
|
io_uring_queue_exit(&ring);
|
|
free_tmp:
|
|
free(tmp);
|
|
close_file:
|
|
fclose(file);
|
|
close_comm_pipes:
|
|
close_comm_pipes(&comm_pipes);
|
|
}
|
|
|
|
static void test_iouring_ro(char *mem, size_t size, bool is_hugetlb)
|
|
{
|
|
do_test_iouring(mem, size, false);
|
|
}
|
|
|
|
static void test_iouring_fork(char *mem, size_t size, bool is_hugetlb)
|
|
{
|
|
do_test_iouring(mem, size, true);
|
|
}
|
|
|
|
#endif /* LOCAL_CONFIG_HAVE_LIBURING */
|
|
|
|
enum ro_pin_test {
|
|
RO_PIN_TEST,
|
|
RO_PIN_TEST_SHARED,
|
|
RO_PIN_TEST_PREVIOUSLY_SHARED,
|
|
RO_PIN_TEST_RO_EXCLUSIVE,
|
|
};
|
|
|
|
static void do_test_ro_pin(char *mem, size_t size, enum ro_pin_test test,
|
|
bool fast)
|
|
{
|
|
struct pin_longterm_test args;
|
|
struct comm_pipes comm_pipes;
|
|
char *tmp, buf;
|
|
__u64 tmp_val;
|
|
int ret;
|
|
|
|
if (gup_fd < 0) {
|
|
ksft_test_result_skip("gup_test not available\n");
|
|
return;
|
|
}
|
|
|
|
tmp = malloc(size);
|
|
if (!tmp) {
|
|
ksft_test_result_fail("malloc() failed\n");
|
|
return;
|
|
}
|
|
|
|
ret = setup_comm_pipes(&comm_pipes);
|
|
if (ret) {
|
|
ksft_test_result_fail("pipe() failed\n");
|
|
goto free_tmp;
|
|
}
|
|
|
|
switch (test) {
|
|
case RO_PIN_TEST:
|
|
break;
|
|
case RO_PIN_TEST_SHARED:
|
|
case RO_PIN_TEST_PREVIOUSLY_SHARED:
|
|
/*
|
|
* Share the pages with our child. As the pages are not pinned,
|
|
* this should just work.
|
|
*/
|
|
ret = fork();
|
|
if (ret < 0) {
|
|
ksft_test_result_fail("fork() failed\n");
|
|
goto close_comm_pipes;
|
|
} else if (!ret) {
|
|
write(comm_pipes.child_ready[1], "0", 1);
|
|
while (read(comm_pipes.parent_ready[0], &buf, 1) != 1)
|
|
;
|
|
exit(0);
|
|
}
|
|
|
|
/* Wait until our child is ready. */
|
|
while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
|
|
;
|
|
|
|
if (test == RO_PIN_TEST_PREVIOUSLY_SHARED) {
|
|
/*
|
|
* Tell the child to quit now and wait until it quit.
|
|
* The pages should now be mapped R/O into our page
|
|
* tables, but they are no longer shared.
|
|
*/
|
|
write(comm_pipes.parent_ready[1], "0", 1);
|
|
wait(&ret);
|
|
if (!WIFEXITED(ret))
|
|
ksft_print_msg("[INFO] wait() failed\n");
|
|
}
|
|
break;
|
|
case RO_PIN_TEST_RO_EXCLUSIVE:
|
|
/*
|
|
* Map the page R/O into the page table. Enable softdirty
|
|
* tracking to stop the page from getting mapped R/W immediately
|
|
* again by mprotect() optimizations. Note that we don't have an
|
|
* easy way to test if that worked (the pagemap does not export
|
|
* if the page is mapped R/O vs. R/W).
|
|
*/
|
|
ret = mprotect(mem, size, PROT_READ);
|
|
clear_softdirty();
|
|
ret |= mprotect(mem, size, PROT_READ | PROT_WRITE);
|
|
if (ret) {
|
|
ksft_test_result_fail("mprotect() failed\n");
|
|
goto close_comm_pipes;
|
|
}
|
|
break;
|
|
default:
|
|
assert(false);
|
|
}
|
|
|
|
/* Take a R/O pin. This should trigger unsharing. */
|
|
args.addr = (__u64)(uintptr_t)mem;
|
|
args.size = size;
|
|
args.flags = fast ? PIN_LONGTERM_TEST_FLAG_USE_FAST : 0;
|
|
ret = ioctl(gup_fd, PIN_LONGTERM_TEST_START, &args);
|
|
if (ret) {
|
|
if (errno == EINVAL)
|
|
ksft_test_result_skip("PIN_LONGTERM_TEST_START failed\n");
|
|
else
|
|
ksft_test_result_fail("PIN_LONGTERM_TEST_START failed\n");
|
|
goto wait;
|
|
}
|
|
|
|
/* Modify the page. */
|
|
memset(mem, 0xff, size);
|
|
|
|
/*
|
|
* Read back the content via the pin to the temporary buffer and
|
|
* test if we observed the modification.
|
|
*/
|
|
tmp_val = (__u64)(uintptr_t)tmp;
|
|
ret = ioctl(gup_fd, PIN_LONGTERM_TEST_READ, &tmp_val);
|
|
if (ret)
|
|
ksft_test_result_fail("PIN_LONGTERM_TEST_READ failed\n");
|
|
else
|
|
ksft_test_result(!memcmp(mem, tmp, size),
|
|
"Longterm R/O pin is reliable\n");
|
|
|
|
ret = ioctl(gup_fd, PIN_LONGTERM_TEST_STOP);
|
|
if (ret)
|
|
ksft_print_msg("[INFO] PIN_LONGTERM_TEST_STOP failed\n");
|
|
wait:
|
|
switch (test) {
|
|
case RO_PIN_TEST_SHARED:
|
|
write(comm_pipes.parent_ready[1], "0", 1);
|
|
wait(&ret);
|
|
if (!WIFEXITED(ret))
|
|
ksft_print_msg("[INFO] wait() failed\n");
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
close_comm_pipes:
|
|
close_comm_pipes(&comm_pipes);
|
|
free_tmp:
|
|
free(tmp);
|
|
}
|
|
|
|
static void test_ro_pin_on_shared(char *mem, size_t size, bool is_hugetlb)
|
|
{
|
|
do_test_ro_pin(mem, size, RO_PIN_TEST_SHARED, false);
|
|
}
|
|
|
|
static void test_ro_fast_pin_on_shared(char *mem, size_t size, bool is_hugetlb)
|
|
{
|
|
do_test_ro_pin(mem, size, RO_PIN_TEST_SHARED, true);
|
|
}
|
|
|
|
static void test_ro_pin_on_ro_previously_shared(char *mem, size_t size,
|
|
bool is_hugetlb)
|
|
{
|
|
do_test_ro_pin(mem, size, RO_PIN_TEST_PREVIOUSLY_SHARED, false);
|
|
}
|
|
|
|
static void test_ro_fast_pin_on_ro_previously_shared(char *mem, size_t size,
|
|
bool is_hugetlb)
|
|
{
|
|
do_test_ro_pin(mem, size, RO_PIN_TEST_PREVIOUSLY_SHARED, true);
|
|
}
|
|
|
|
static void test_ro_pin_on_ro_exclusive(char *mem, size_t size,
|
|
bool is_hugetlb)
|
|
{
|
|
do_test_ro_pin(mem, size, RO_PIN_TEST_RO_EXCLUSIVE, false);
|
|
}
|
|
|
|
static void test_ro_fast_pin_on_ro_exclusive(char *mem, size_t size,
|
|
bool is_hugetlb)
|
|
{
|
|
do_test_ro_pin(mem, size, RO_PIN_TEST_RO_EXCLUSIVE, true);
|
|
}
|
|
|
|
typedef void (*test_fn)(char *mem, size_t size, bool hugetlb);
|
|
|
|
static void do_run_with_base_page(test_fn fn, bool swapout)
|
|
{
|
|
char *mem;
|
|
int ret;
|
|
|
|
mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE,
|
|
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
return;
|
|
}
|
|
|
|
ret = madvise(mem, pagesize, MADV_NOHUGEPAGE);
|
|
/* Ignore if not around on a kernel. */
|
|
if (ret && errno != EINVAL) {
|
|
ksft_test_result_fail("MADV_NOHUGEPAGE failed\n");
|
|
goto munmap;
|
|
}
|
|
|
|
/* Populate a base page. */
|
|
memset(mem, 0, pagesize);
|
|
|
|
if (swapout) {
|
|
madvise(mem, pagesize, MADV_PAGEOUT);
|
|
if (!pagemap_is_swapped(pagemap_fd, mem)) {
|
|
ksft_test_result_skip("MADV_PAGEOUT did not work, is swap enabled?\n");
|
|
goto munmap;
|
|
}
|
|
}
|
|
|
|
fn(mem, pagesize, false);
|
|
munmap:
|
|
munmap(mem, pagesize);
|
|
}
|
|
|
|
static void run_with_base_page(test_fn fn, const char *desc)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with base page\n", desc);
|
|
do_run_with_base_page(fn, false);
|
|
}
|
|
|
|
static void run_with_base_page_swap(test_fn fn, const char *desc)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with swapped out base page\n", desc);
|
|
do_run_with_base_page(fn, true);
|
|
}
|
|
|
|
enum thp_run {
|
|
THP_RUN_PMD,
|
|
THP_RUN_PMD_SWAPOUT,
|
|
THP_RUN_PTE,
|
|
THP_RUN_PTE_SWAPOUT,
|
|
THP_RUN_SINGLE_PTE,
|
|
THP_RUN_SINGLE_PTE_SWAPOUT,
|
|
THP_RUN_PARTIAL_MREMAP,
|
|
THP_RUN_PARTIAL_SHARED,
|
|
};
|
|
|
|
static void do_run_with_thp(test_fn fn, enum thp_run thp_run, size_t thpsize)
|
|
{
|
|
char *mem, *mmap_mem, *tmp, *mremap_mem = MAP_FAILED;
|
|
size_t size, mmap_size, mremap_size;
|
|
int ret;
|
|
|
|
/* For alignment purposes, we need twice the thp size. */
|
|
mmap_size = 2 * thpsize;
|
|
mmap_mem = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
|
|
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
|
|
if (mmap_mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
return;
|
|
}
|
|
|
|
/* We need a THP-aligned memory area. */
|
|
mem = (char *)(((uintptr_t)mmap_mem + thpsize) & ~(thpsize - 1));
|
|
|
|
ret = madvise(mem, thpsize, MADV_HUGEPAGE);
|
|
if (ret) {
|
|
ksft_test_result_fail("MADV_HUGEPAGE failed\n");
|
|
goto munmap;
|
|
}
|
|
|
|
/*
|
|
* Try to populate a THP. Touch the first sub-page and test if
|
|
* we get the last sub-page populated automatically.
|
|
*/
|
|
mem[0] = 0;
|
|
if (!pagemap_is_populated(pagemap_fd, mem + thpsize - pagesize)) {
|
|
ksft_test_result_skip("Did not get a THP populated\n");
|
|
goto munmap;
|
|
}
|
|
memset(mem, 0, thpsize);
|
|
|
|
size = thpsize;
|
|
switch (thp_run) {
|
|
case THP_RUN_PMD:
|
|
case THP_RUN_PMD_SWAPOUT:
|
|
assert(thpsize == pmdsize);
|
|
break;
|
|
case THP_RUN_PTE:
|
|
case THP_RUN_PTE_SWAPOUT:
|
|
/*
|
|
* Trigger PTE-mapping the THP by temporarily mapping a single
|
|
* subpage R/O. This is a noop if the THP is not pmdsize (and
|
|
* therefore already PTE-mapped).
|
|
*/
|
|
ret = mprotect(mem + pagesize, pagesize, PROT_READ);
|
|
if (ret) {
|
|
ksft_test_result_fail("mprotect() failed\n");
|
|
goto munmap;
|
|
}
|
|
ret = mprotect(mem + pagesize, pagesize, PROT_READ | PROT_WRITE);
|
|
if (ret) {
|
|
ksft_test_result_fail("mprotect() failed\n");
|
|
goto munmap;
|
|
}
|
|
break;
|
|
case THP_RUN_SINGLE_PTE:
|
|
case THP_RUN_SINGLE_PTE_SWAPOUT:
|
|
/*
|
|
* Discard all but a single subpage of that PTE-mapped THP. What
|
|
* remains is a single PTE mapping a single subpage.
|
|
*/
|
|
ret = madvise(mem + pagesize, thpsize - pagesize, MADV_DONTNEED);
|
|
if (ret) {
|
|
ksft_test_result_fail("MADV_DONTNEED failed\n");
|
|
goto munmap;
|
|
}
|
|
size = pagesize;
|
|
break;
|
|
case THP_RUN_PARTIAL_MREMAP:
|
|
/*
|
|
* Remap half of the THP. We need some new memory location
|
|
* for that.
|
|
*/
|
|
mremap_size = thpsize / 2;
|
|
mremap_mem = mmap(NULL, mremap_size, PROT_NONE,
|
|
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
goto munmap;
|
|
}
|
|
tmp = mremap(mem + mremap_size, mremap_size, mremap_size,
|
|
MREMAP_MAYMOVE | MREMAP_FIXED, mremap_mem);
|
|
if (tmp != mremap_mem) {
|
|
ksft_test_result_fail("mremap() failed\n");
|
|
goto munmap;
|
|
}
|
|
size = mremap_size;
|
|
break;
|
|
case THP_RUN_PARTIAL_SHARED:
|
|
/*
|
|
* Share the first page of the THP with a child and quit the
|
|
* child. This will result in some parts of the THP never
|
|
* have been shared.
|
|
*/
|
|
ret = madvise(mem + pagesize, thpsize - pagesize, MADV_DONTFORK);
|
|
if (ret) {
|
|
ksft_test_result_fail("MADV_DONTFORK failed\n");
|
|
goto munmap;
|
|
}
|
|
ret = fork();
|
|
if (ret < 0) {
|
|
ksft_test_result_fail("fork() failed\n");
|
|
goto munmap;
|
|
} else if (!ret) {
|
|
exit(0);
|
|
}
|
|
wait(&ret);
|
|
/* Allow for sharing all pages again. */
|
|
ret = madvise(mem + pagesize, thpsize - pagesize, MADV_DOFORK);
|
|
if (ret) {
|
|
ksft_test_result_fail("MADV_DOFORK failed\n");
|
|
goto munmap;
|
|
}
|
|
break;
|
|
default:
|
|
assert(false);
|
|
}
|
|
|
|
switch (thp_run) {
|
|
case THP_RUN_PMD_SWAPOUT:
|
|
case THP_RUN_PTE_SWAPOUT:
|
|
case THP_RUN_SINGLE_PTE_SWAPOUT:
|
|
madvise(mem, size, MADV_PAGEOUT);
|
|
if (!range_is_swapped(mem, size)) {
|
|
ksft_test_result_skip("MADV_PAGEOUT did not work, is swap enabled?\n");
|
|
goto munmap;
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
fn(mem, size, false);
|
|
munmap:
|
|
munmap(mmap_mem, mmap_size);
|
|
if (mremap_mem != MAP_FAILED)
|
|
munmap(mremap_mem, mremap_size);
|
|
}
|
|
|
|
static void run_with_thp(test_fn fn, const char *desc, size_t size)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with THP (%zu kB)\n",
|
|
desc, size / 1024);
|
|
do_run_with_thp(fn, THP_RUN_PMD, size);
|
|
}
|
|
|
|
static void run_with_thp_swap(test_fn fn, const char *desc, size_t size)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with swapped-out THP (%zu kB)\n",
|
|
desc, size / 1024);
|
|
do_run_with_thp(fn, THP_RUN_PMD_SWAPOUT, size);
|
|
}
|
|
|
|
static void run_with_pte_mapped_thp(test_fn fn, const char *desc, size_t size)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with PTE-mapped THP (%zu kB)\n",
|
|
desc, size / 1024);
|
|
do_run_with_thp(fn, THP_RUN_PTE, size);
|
|
}
|
|
|
|
static void run_with_pte_mapped_thp_swap(test_fn fn, const char *desc, size_t size)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with swapped-out, PTE-mapped THP (%zu kB)\n",
|
|
desc, size / 1024);
|
|
do_run_with_thp(fn, THP_RUN_PTE_SWAPOUT, size);
|
|
}
|
|
|
|
static void run_with_single_pte_of_thp(test_fn fn, const char *desc, size_t size)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with single PTE of THP (%zu kB)\n",
|
|
desc, size / 1024);
|
|
do_run_with_thp(fn, THP_RUN_SINGLE_PTE, size);
|
|
}
|
|
|
|
static void run_with_single_pte_of_thp_swap(test_fn fn, const char *desc, size_t size)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with single PTE of swapped-out THP (%zu kB)\n",
|
|
desc, size / 1024);
|
|
do_run_with_thp(fn, THP_RUN_SINGLE_PTE_SWAPOUT, size);
|
|
}
|
|
|
|
static void run_with_partial_mremap_thp(test_fn fn, const char *desc, size_t size)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with partially mremap()'ed THP (%zu kB)\n",
|
|
desc, size / 1024);
|
|
do_run_with_thp(fn, THP_RUN_PARTIAL_MREMAP, size);
|
|
}
|
|
|
|
static void run_with_partial_shared_thp(test_fn fn, const char *desc, size_t size)
|
|
{
|
|
ksft_print_msg("[RUN] %s ... with partially shared THP (%zu kB)\n",
|
|
desc, size / 1024);
|
|
do_run_with_thp(fn, THP_RUN_PARTIAL_SHARED, size);
|
|
}
|
|
|
|
static void run_with_hugetlb(test_fn fn, const char *desc, size_t hugetlbsize)
|
|
{
|
|
int flags = MAP_PRIVATE | MAP_ANONYMOUS | MAP_HUGETLB;
|
|
char *mem, *dummy;
|
|
|
|
ksft_print_msg("[RUN] %s ... with hugetlb (%zu kB)\n", desc,
|
|
hugetlbsize / 1024);
|
|
|
|
flags |= __builtin_ctzll(hugetlbsize) << MAP_HUGE_SHIFT;
|
|
|
|
mem = mmap(NULL, hugetlbsize, PROT_READ | PROT_WRITE, flags, -1, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_skip("need more free huge pages\n");
|
|
return;
|
|
}
|
|
|
|
/* Populate an huge page. */
|
|
memset(mem, 0, hugetlbsize);
|
|
|
|
/*
|
|
* We need a total of two hugetlb pages to handle COW/unsharing
|
|
* properly, otherwise we might get zapped by a SIGBUS.
|
|
*/
|
|
dummy = mmap(NULL, hugetlbsize, PROT_READ | PROT_WRITE, flags, -1, 0);
|
|
if (dummy == MAP_FAILED) {
|
|
ksft_test_result_skip("need more free huge pages\n");
|
|
goto munmap;
|
|
}
|
|
munmap(dummy, hugetlbsize);
|
|
|
|
fn(mem, hugetlbsize, true);
|
|
munmap:
|
|
munmap(mem, hugetlbsize);
|
|
}
|
|
|
|
struct test_case {
|
|
const char *desc;
|
|
test_fn fn;
|
|
};
|
|
|
|
/*
|
|
* Test cases that are specific to anonymous pages: pages in private mappings
|
|
* that may get shared via COW during fork().
|
|
*/
|
|
static const struct test_case anon_test_cases[] = {
|
|
/*
|
|
* Basic COW tests for fork() without any GUP. If we miss to break COW,
|
|
* either the child can observe modifications by the parent or the
|
|
* other way around.
|
|
*/
|
|
{
|
|
"Basic COW after fork()",
|
|
test_cow_in_parent,
|
|
},
|
|
/*
|
|
* Basic test, but do an additional mprotect(PROT_READ)+
|
|
* mprotect(PROT_READ|PROT_WRITE) in the parent before write access.
|
|
*/
|
|
{
|
|
"Basic COW after fork() with mprotect() optimization",
|
|
test_cow_in_parent_mprotect,
|
|
},
|
|
/*
|
|
* vmsplice() [R/O GUP] + unmap in the child; modify in the parent. If
|
|
* we miss to break COW, the child observes modifications by the parent.
|
|
* This is CVE-2020-29374 reported by Jann Horn.
|
|
*/
|
|
{
|
|
"vmsplice() + unmap in child",
|
|
test_vmsplice_in_child,
|
|
},
|
|
/*
|
|
* vmsplice() test, but do an additional mprotect(PROT_READ)+
|
|
* mprotect(PROT_READ|PROT_WRITE) in the parent before write access.
|
|
*/
|
|
{
|
|
"vmsplice() + unmap in child with mprotect() optimization",
|
|
test_vmsplice_in_child_mprotect,
|
|
},
|
|
/*
|
|
* vmsplice() [R/O GUP] in parent before fork(), unmap in parent after
|
|
* fork(); modify in the child. If we miss to break COW, the parent
|
|
* observes modifications by the child.
|
|
*/
|
|
{
|
|
"vmsplice() before fork(), unmap in parent after fork()",
|
|
test_vmsplice_before_fork,
|
|
},
|
|
/*
|
|
* vmsplice() [R/O GUP] + unmap in parent after fork(); modify in the
|
|
* child. If we miss to break COW, the parent observes modifications by
|
|
* the child.
|
|
*/
|
|
{
|
|
"vmsplice() + unmap in parent after fork()",
|
|
test_vmsplice_after_fork,
|
|
},
|
|
#ifdef LOCAL_CONFIG_HAVE_LIBURING
|
|
/*
|
|
* Take a R/W longterm pin and then map the page R/O into the page
|
|
* table to trigger a write fault on next access. When modifying the
|
|
* page, the page content must be visible via the pin.
|
|
*/
|
|
{
|
|
"R/O-mapping a page registered as iouring fixed buffer",
|
|
test_iouring_ro,
|
|
},
|
|
/*
|
|
* Take a R/W longterm pin and then fork() a child. When modifying the
|
|
* page, the page content must be visible via the pin. We expect the
|
|
* pinned page to not get shared with the child.
|
|
*/
|
|
{
|
|
"fork() with an iouring fixed buffer",
|
|
test_iouring_fork,
|
|
},
|
|
|
|
#endif /* LOCAL_CONFIG_HAVE_LIBURING */
|
|
/*
|
|
* Take a R/O longterm pin on a R/O-mapped shared anonymous page.
|
|
* When modifying the page via the page table, the page content change
|
|
* must be visible via the pin.
|
|
*/
|
|
{
|
|
"R/O GUP pin on R/O-mapped shared page",
|
|
test_ro_pin_on_shared,
|
|
},
|
|
/* Same as above, but using GUP-fast. */
|
|
{
|
|
"R/O GUP-fast pin on R/O-mapped shared page",
|
|
test_ro_fast_pin_on_shared,
|
|
},
|
|
/*
|
|
* Take a R/O longterm pin on a R/O-mapped exclusive anonymous page that
|
|
* was previously shared. When modifying the page via the page table,
|
|
* the page content change must be visible via the pin.
|
|
*/
|
|
{
|
|
"R/O GUP pin on R/O-mapped previously-shared page",
|
|
test_ro_pin_on_ro_previously_shared,
|
|
},
|
|
/* Same as above, but using GUP-fast. */
|
|
{
|
|
"R/O GUP-fast pin on R/O-mapped previously-shared page",
|
|
test_ro_fast_pin_on_ro_previously_shared,
|
|
},
|
|
/*
|
|
* Take a R/O longterm pin on a R/O-mapped exclusive anonymous page.
|
|
* When modifying the page via the page table, the page content change
|
|
* must be visible via the pin.
|
|
*/
|
|
{
|
|
"R/O GUP pin on R/O-mapped exclusive page",
|
|
test_ro_pin_on_ro_exclusive,
|
|
},
|
|
/* Same as above, but using GUP-fast. */
|
|
{
|
|
"R/O GUP-fast pin on R/O-mapped exclusive page",
|
|
test_ro_fast_pin_on_ro_exclusive,
|
|
},
|
|
};
|
|
|
|
static void run_anon_test_case(struct test_case const *test_case)
|
|
{
|
|
int i;
|
|
|
|
run_with_base_page(test_case->fn, test_case->desc);
|
|
run_with_base_page_swap(test_case->fn, test_case->desc);
|
|
for (i = 0; i < nr_thpsizes; i++) {
|
|
size_t size = thpsizes[i];
|
|
struct thp_settings settings = *thp_current_settings();
|
|
|
|
settings.hugepages[sz2ord(pmdsize)].enabled = THP_NEVER;
|
|
settings.hugepages[sz2ord(size)].enabled = THP_ALWAYS;
|
|
thp_push_settings(&settings);
|
|
|
|
if (size == pmdsize) {
|
|
run_with_thp(test_case->fn, test_case->desc, size);
|
|
run_with_thp_swap(test_case->fn, test_case->desc, size);
|
|
}
|
|
|
|
run_with_pte_mapped_thp(test_case->fn, test_case->desc, size);
|
|
run_with_pte_mapped_thp_swap(test_case->fn, test_case->desc, size);
|
|
run_with_single_pte_of_thp(test_case->fn, test_case->desc, size);
|
|
run_with_single_pte_of_thp_swap(test_case->fn, test_case->desc, size);
|
|
run_with_partial_mremap_thp(test_case->fn, test_case->desc, size);
|
|
run_with_partial_shared_thp(test_case->fn, test_case->desc, size);
|
|
|
|
thp_pop_settings();
|
|
}
|
|
for (i = 0; i < nr_hugetlbsizes; i++)
|
|
run_with_hugetlb(test_case->fn, test_case->desc,
|
|
hugetlbsizes[i]);
|
|
}
|
|
|
|
static void run_anon_test_cases(void)
|
|
{
|
|
int i;
|
|
|
|
ksft_print_msg("[INFO] Anonymous memory tests in private mappings\n");
|
|
|
|
for (i = 0; i < ARRAY_SIZE(anon_test_cases); i++)
|
|
run_anon_test_case(&anon_test_cases[i]);
|
|
}
|
|
|
|
static int tests_per_anon_test_case(void)
|
|
{
|
|
int tests = 2 + nr_hugetlbsizes;
|
|
|
|
tests += 6 * nr_thpsizes;
|
|
if (pmdsize)
|
|
tests += 2;
|
|
return tests;
|
|
}
|
|
|
|
enum anon_thp_collapse_test {
|
|
ANON_THP_COLLAPSE_UNSHARED,
|
|
ANON_THP_COLLAPSE_FULLY_SHARED,
|
|
ANON_THP_COLLAPSE_LOWER_SHARED,
|
|
ANON_THP_COLLAPSE_UPPER_SHARED,
|
|
};
|
|
|
|
static void do_test_anon_thp_collapse(char *mem, size_t size,
|
|
enum anon_thp_collapse_test test)
|
|
{
|
|
struct comm_pipes comm_pipes;
|
|
char buf;
|
|
int ret;
|
|
|
|
ret = setup_comm_pipes(&comm_pipes);
|
|
if (ret) {
|
|
ksft_test_result_fail("pipe() failed\n");
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Trigger PTE-mapping the THP by temporarily mapping a single subpage
|
|
* R/O, such that we can try collapsing it later.
|
|
*/
|
|
ret = mprotect(mem + pagesize, pagesize, PROT_READ);
|
|
if (ret) {
|
|
ksft_test_result_fail("mprotect() failed\n");
|
|
goto close_comm_pipes;
|
|
}
|
|
ret = mprotect(mem + pagesize, pagesize, PROT_READ | PROT_WRITE);
|
|
if (ret) {
|
|
ksft_test_result_fail("mprotect() failed\n");
|
|
goto close_comm_pipes;
|
|
}
|
|
|
|
switch (test) {
|
|
case ANON_THP_COLLAPSE_UNSHARED:
|
|
/* Collapse before actually COW-sharing the page. */
|
|
ret = madvise(mem, size, MADV_COLLAPSE);
|
|
if (ret) {
|
|
ksft_test_result_skip("MADV_COLLAPSE failed: %s\n",
|
|
strerror(errno));
|
|
goto close_comm_pipes;
|
|
}
|
|
break;
|
|
case ANON_THP_COLLAPSE_FULLY_SHARED:
|
|
/* COW-share the full PTE-mapped THP. */
|
|
break;
|
|
case ANON_THP_COLLAPSE_LOWER_SHARED:
|
|
/* Don't COW-share the upper part of the THP. */
|
|
ret = madvise(mem + size / 2, size / 2, MADV_DONTFORK);
|
|
if (ret) {
|
|
ksft_test_result_fail("MADV_DONTFORK failed\n");
|
|
goto close_comm_pipes;
|
|
}
|
|
break;
|
|
case ANON_THP_COLLAPSE_UPPER_SHARED:
|
|
/* Don't COW-share the lower part of the THP. */
|
|
ret = madvise(mem, size / 2, MADV_DONTFORK);
|
|
if (ret) {
|
|
ksft_test_result_fail("MADV_DONTFORK failed\n");
|
|
goto close_comm_pipes;
|
|
}
|
|
break;
|
|
default:
|
|
assert(false);
|
|
}
|
|
|
|
ret = fork();
|
|
if (ret < 0) {
|
|
ksft_test_result_fail("fork() failed\n");
|
|
goto close_comm_pipes;
|
|
} else if (!ret) {
|
|
switch (test) {
|
|
case ANON_THP_COLLAPSE_UNSHARED:
|
|
case ANON_THP_COLLAPSE_FULLY_SHARED:
|
|
exit(child_memcmp_fn(mem, size, &comm_pipes));
|
|
break;
|
|
case ANON_THP_COLLAPSE_LOWER_SHARED:
|
|
exit(child_memcmp_fn(mem, size / 2, &comm_pipes));
|
|
break;
|
|
case ANON_THP_COLLAPSE_UPPER_SHARED:
|
|
exit(child_memcmp_fn(mem + size / 2, size / 2,
|
|
&comm_pipes));
|
|
break;
|
|
default:
|
|
assert(false);
|
|
}
|
|
}
|
|
|
|
while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
|
|
;
|
|
|
|
switch (test) {
|
|
case ANON_THP_COLLAPSE_UNSHARED:
|
|
break;
|
|
case ANON_THP_COLLAPSE_UPPER_SHARED:
|
|
case ANON_THP_COLLAPSE_LOWER_SHARED:
|
|
/*
|
|
* Revert MADV_DONTFORK such that we merge the VMAs and are
|
|
* able to actually collapse.
|
|
*/
|
|
ret = madvise(mem, size, MADV_DOFORK);
|
|
if (ret) {
|
|
ksft_test_result_fail("MADV_DOFORK failed\n");
|
|
write(comm_pipes.parent_ready[1], "0", 1);
|
|
wait(&ret);
|
|
goto close_comm_pipes;
|
|
}
|
|
/* FALLTHROUGH */
|
|
case ANON_THP_COLLAPSE_FULLY_SHARED:
|
|
/* Collapse before anyone modified the COW-shared page. */
|
|
ret = madvise(mem, size, MADV_COLLAPSE);
|
|
if (ret) {
|
|
ksft_test_result_skip("MADV_COLLAPSE failed: %s\n",
|
|
strerror(errno));
|
|
write(comm_pipes.parent_ready[1], "0", 1);
|
|
wait(&ret);
|
|
goto close_comm_pipes;
|
|
}
|
|
break;
|
|
default:
|
|
assert(false);
|
|
}
|
|
|
|
/* Modify the page. */
|
|
memset(mem, 0xff, size);
|
|
write(comm_pipes.parent_ready[1], "0", 1);
|
|
|
|
wait(&ret);
|
|
if (WIFEXITED(ret))
|
|
ret = WEXITSTATUS(ret);
|
|
else
|
|
ret = -EINVAL;
|
|
|
|
ksft_test_result(!ret, "No leak from parent into child\n");
|
|
close_comm_pipes:
|
|
close_comm_pipes(&comm_pipes);
|
|
}
|
|
|
|
static void test_anon_thp_collapse_unshared(char *mem, size_t size,
|
|
bool is_hugetlb)
|
|
{
|
|
assert(!is_hugetlb);
|
|
do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_UNSHARED);
|
|
}
|
|
|
|
static void test_anon_thp_collapse_fully_shared(char *mem, size_t size,
|
|
bool is_hugetlb)
|
|
{
|
|
assert(!is_hugetlb);
|
|
do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_FULLY_SHARED);
|
|
}
|
|
|
|
static void test_anon_thp_collapse_lower_shared(char *mem, size_t size,
|
|
bool is_hugetlb)
|
|
{
|
|
assert(!is_hugetlb);
|
|
do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_LOWER_SHARED);
|
|
}
|
|
|
|
static void test_anon_thp_collapse_upper_shared(char *mem, size_t size,
|
|
bool is_hugetlb)
|
|
{
|
|
assert(!is_hugetlb);
|
|
do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_UPPER_SHARED);
|
|
}
|
|
|
|
/*
|
|
* Test cases that are specific to anonymous THP: pages in private mappings
|
|
* that may get shared via COW during fork().
|
|
*/
|
|
static const struct test_case anon_thp_test_cases[] = {
|
|
/*
|
|
* Basic COW test for fork() without any GUP when collapsing a THP
|
|
* before fork().
|
|
*
|
|
* Re-mapping a PTE-mapped anon THP using a single PMD ("in-place
|
|
* collapse") might easily get COW handling wrong when not collapsing
|
|
* exclusivity information properly.
|
|
*/
|
|
{
|
|
"Basic COW after fork() when collapsing before fork()",
|
|
test_anon_thp_collapse_unshared,
|
|
},
|
|
/* Basic COW test, but collapse after COW-sharing a full THP. */
|
|
{
|
|
"Basic COW after fork() when collapsing after fork() (fully shared)",
|
|
test_anon_thp_collapse_fully_shared,
|
|
},
|
|
/*
|
|
* Basic COW test, but collapse after COW-sharing the lower half of a
|
|
* THP.
|
|
*/
|
|
{
|
|
"Basic COW after fork() when collapsing after fork() (lower shared)",
|
|
test_anon_thp_collapse_lower_shared,
|
|
},
|
|
/*
|
|
* Basic COW test, but collapse after COW-sharing the upper half of a
|
|
* THP.
|
|
*/
|
|
{
|
|
"Basic COW after fork() when collapsing after fork() (upper shared)",
|
|
test_anon_thp_collapse_upper_shared,
|
|
},
|
|
};
|
|
|
|
static void run_anon_thp_test_cases(void)
|
|
{
|
|
int i;
|
|
|
|
if (!pmdsize)
|
|
return;
|
|
|
|
ksft_print_msg("[INFO] Anonymous THP tests\n");
|
|
|
|
for (i = 0; i < ARRAY_SIZE(anon_thp_test_cases); i++) {
|
|
struct test_case const *test_case = &anon_thp_test_cases[i];
|
|
|
|
ksft_print_msg("[RUN] %s\n", test_case->desc);
|
|
do_run_with_thp(test_case->fn, THP_RUN_PMD, pmdsize);
|
|
}
|
|
}
|
|
|
|
static int tests_per_anon_thp_test_case(void)
|
|
{
|
|
return pmdsize ? 1 : 0;
|
|
}
|
|
|
|
typedef void (*non_anon_test_fn)(char *mem, const char *smem, size_t size);
|
|
|
|
static void test_cow(char *mem, const char *smem, size_t size)
|
|
{
|
|
char *old = malloc(size);
|
|
|
|
/* Backup the original content. */
|
|
memcpy(old, smem, size);
|
|
|
|
/* Modify the page. */
|
|
memset(mem, 0xff, size);
|
|
|
|
/* See if we still read the old values via the other mapping. */
|
|
ksft_test_result(!memcmp(smem, old, size),
|
|
"Other mapping not modified\n");
|
|
free(old);
|
|
}
|
|
|
|
static void test_ro_pin(char *mem, const char *smem, size_t size)
|
|
{
|
|
do_test_ro_pin(mem, size, RO_PIN_TEST, false);
|
|
}
|
|
|
|
static void test_ro_fast_pin(char *mem, const char *smem, size_t size)
|
|
{
|
|
do_test_ro_pin(mem, size, RO_PIN_TEST, true);
|
|
}
|
|
|
|
static void run_with_zeropage(non_anon_test_fn fn, const char *desc)
|
|
{
|
|
char *mem, *smem, tmp;
|
|
|
|
ksft_print_msg("[RUN] %s ... with shared zeropage\n", desc);
|
|
|
|
mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE,
|
|
MAP_PRIVATE | MAP_ANON, -1, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
return;
|
|
}
|
|
|
|
smem = mmap(NULL, pagesize, PROT_READ, MAP_PRIVATE | MAP_ANON, -1, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
goto munmap;
|
|
}
|
|
|
|
/* Read from the page to populate the shared zeropage. */
|
|
tmp = *mem + *smem;
|
|
asm volatile("" : "+r" (tmp));
|
|
|
|
fn(mem, smem, pagesize);
|
|
munmap:
|
|
munmap(mem, pagesize);
|
|
if (smem != MAP_FAILED)
|
|
munmap(smem, pagesize);
|
|
}
|
|
|
|
static void run_with_huge_zeropage(non_anon_test_fn fn, const char *desc)
|
|
{
|
|
char *mem, *smem, *mmap_mem, *mmap_smem, tmp;
|
|
size_t mmap_size;
|
|
int ret;
|
|
|
|
ksft_print_msg("[RUN] %s ... with huge zeropage\n", desc);
|
|
|
|
if (!has_huge_zeropage) {
|
|
ksft_test_result_skip("Huge zeropage not enabled\n");
|
|
return;
|
|
}
|
|
|
|
/* For alignment purposes, we need twice the thp size. */
|
|
mmap_size = 2 * pmdsize;
|
|
mmap_mem = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
|
|
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
|
|
if (mmap_mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
return;
|
|
}
|
|
mmap_smem = mmap(NULL, mmap_size, PROT_READ,
|
|
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
|
|
if (mmap_smem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
goto munmap;
|
|
}
|
|
|
|
/* We need a THP-aligned memory area. */
|
|
mem = (char *)(((uintptr_t)mmap_mem + pmdsize) & ~(pmdsize - 1));
|
|
smem = (char *)(((uintptr_t)mmap_smem + pmdsize) & ~(pmdsize - 1));
|
|
|
|
ret = madvise(mem, pmdsize, MADV_HUGEPAGE);
|
|
ret |= madvise(smem, pmdsize, MADV_HUGEPAGE);
|
|
if (ret) {
|
|
ksft_test_result_fail("MADV_HUGEPAGE failed\n");
|
|
goto munmap;
|
|
}
|
|
|
|
/*
|
|
* Read from the memory to populate the huge shared zeropage. Read from
|
|
* the first sub-page and test if we get another sub-page populated
|
|
* automatically.
|
|
*/
|
|
tmp = *mem + *smem;
|
|
asm volatile("" : "+r" (tmp));
|
|
if (!pagemap_is_populated(pagemap_fd, mem + pagesize) ||
|
|
!pagemap_is_populated(pagemap_fd, smem + pagesize)) {
|
|
ksft_test_result_skip("Did not get THPs populated\n");
|
|
goto munmap;
|
|
}
|
|
|
|
fn(mem, smem, pmdsize);
|
|
munmap:
|
|
munmap(mmap_mem, mmap_size);
|
|
if (mmap_smem != MAP_FAILED)
|
|
munmap(mmap_smem, mmap_size);
|
|
}
|
|
|
|
static void run_with_memfd(non_anon_test_fn fn, const char *desc)
|
|
{
|
|
char *mem, *smem, tmp;
|
|
int fd;
|
|
|
|
ksft_print_msg("[RUN] %s ... with memfd\n", desc);
|
|
|
|
fd = memfd_create("test", 0);
|
|
if (fd < 0) {
|
|
ksft_test_result_fail("memfd_create() failed\n");
|
|
return;
|
|
}
|
|
|
|
/* File consists of a single page filled with zeroes. */
|
|
if (fallocate(fd, 0, 0, pagesize)) {
|
|
ksft_test_result_fail("fallocate() failed\n");
|
|
goto close;
|
|
}
|
|
|
|
/* Create a private mapping of the memfd. */
|
|
mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
goto close;
|
|
}
|
|
smem = mmap(NULL, pagesize, PROT_READ, MAP_SHARED, fd, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
goto munmap;
|
|
}
|
|
|
|
/* Fault the page in. */
|
|
tmp = *mem + *smem;
|
|
asm volatile("" : "+r" (tmp));
|
|
|
|
fn(mem, smem, pagesize);
|
|
munmap:
|
|
munmap(mem, pagesize);
|
|
if (smem != MAP_FAILED)
|
|
munmap(smem, pagesize);
|
|
close:
|
|
close(fd);
|
|
}
|
|
|
|
static void run_with_tmpfile(non_anon_test_fn fn, const char *desc)
|
|
{
|
|
char *mem, *smem, tmp;
|
|
FILE *file;
|
|
int fd;
|
|
|
|
ksft_print_msg("[RUN] %s ... with tmpfile\n", desc);
|
|
|
|
file = tmpfile();
|
|
if (!file) {
|
|
ksft_test_result_fail("tmpfile() failed\n");
|
|
return;
|
|
}
|
|
|
|
fd = fileno(file);
|
|
if (fd < 0) {
|
|
ksft_test_result_skip("fileno() failed\n");
|
|
return;
|
|
}
|
|
|
|
/* File consists of a single page filled with zeroes. */
|
|
if (fallocate(fd, 0, 0, pagesize)) {
|
|
ksft_test_result_fail("fallocate() failed\n");
|
|
goto close;
|
|
}
|
|
|
|
/* Create a private mapping of the memfd. */
|
|
mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
goto close;
|
|
}
|
|
smem = mmap(NULL, pagesize, PROT_READ, MAP_SHARED, fd, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
goto munmap;
|
|
}
|
|
|
|
/* Fault the page in. */
|
|
tmp = *mem + *smem;
|
|
asm volatile("" : "+r" (tmp));
|
|
|
|
fn(mem, smem, pagesize);
|
|
munmap:
|
|
munmap(mem, pagesize);
|
|
if (smem != MAP_FAILED)
|
|
munmap(smem, pagesize);
|
|
close:
|
|
fclose(file);
|
|
}
|
|
|
|
static void run_with_memfd_hugetlb(non_anon_test_fn fn, const char *desc,
|
|
size_t hugetlbsize)
|
|
{
|
|
int flags = MFD_HUGETLB;
|
|
char *mem, *smem, tmp;
|
|
int fd;
|
|
|
|
ksft_print_msg("[RUN] %s ... with memfd hugetlb (%zu kB)\n", desc,
|
|
hugetlbsize / 1024);
|
|
|
|
flags |= __builtin_ctzll(hugetlbsize) << MFD_HUGE_SHIFT;
|
|
|
|
fd = memfd_create("test", flags);
|
|
if (fd < 0) {
|
|
ksft_test_result_skip("memfd_create() failed\n");
|
|
return;
|
|
}
|
|
|
|
/* File consists of a single page filled with zeroes. */
|
|
if (fallocate(fd, 0, 0, hugetlbsize)) {
|
|
ksft_test_result_skip("need more free huge pages\n");
|
|
goto close;
|
|
}
|
|
|
|
/* Create a private mapping of the memfd. */
|
|
mem = mmap(NULL, hugetlbsize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd,
|
|
0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_skip("need more free huge pages\n");
|
|
goto close;
|
|
}
|
|
smem = mmap(NULL, hugetlbsize, PROT_READ, MAP_SHARED, fd, 0);
|
|
if (mem == MAP_FAILED) {
|
|
ksft_test_result_fail("mmap() failed\n");
|
|
goto munmap;
|
|
}
|
|
|
|
/* Fault the page in. */
|
|
tmp = *mem + *smem;
|
|
asm volatile("" : "+r" (tmp));
|
|
|
|
fn(mem, smem, hugetlbsize);
|
|
munmap:
|
|
munmap(mem, hugetlbsize);
|
|
if (mem != MAP_FAILED)
|
|
munmap(smem, hugetlbsize);
|
|
close:
|
|
close(fd);
|
|
}
|
|
|
|
struct non_anon_test_case {
|
|
const char *desc;
|
|
non_anon_test_fn fn;
|
|
};
|
|
|
|
/*
|
|
* Test cases that target any pages in private mappings that are not anonymous:
|
|
* pages that may get shared via COW ndependent of fork(). This includes
|
|
* the shared zeropage(s), pagecache pages, ...
|
|
*/
|
|
static const struct non_anon_test_case non_anon_test_cases[] = {
|
|
/*
|
|
* Basic COW test without any GUP. If we miss to break COW, changes are
|
|
* visible via other private/shared mappings.
|
|
*/
|
|
{
|
|
"Basic COW",
|
|
test_cow,
|
|
},
|
|
/*
|
|
* Take a R/O longterm pin. When modifying the page via the page table,
|
|
* the page content change must be visible via the pin.
|
|
*/
|
|
{
|
|
"R/O longterm GUP pin",
|
|
test_ro_pin,
|
|
},
|
|
/* Same as above, but using GUP-fast. */
|
|
{
|
|
"R/O longterm GUP-fast pin",
|
|
test_ro_fast_pin,
|
|
},
|
|
};
|
|
|
|
static void run_non_anon_test_case(struct non_anon_test_case const *test_case)
|
|
{
|
|
int i;
|
|
|
|
run_with_zeropage(test_case->fn, test_case->desc);
|
|
run_with_memfd(test_case->fn, test_case->desc);
|
|
run_with_tmpfile(test_case->fn, test_case->desc);
|
|
if (pmdsize)
|
|
run_with_huge_zeropage(test_case->fn, test_case->desc);
|
|
for (i = 0; i < nr_hugetlbsizes; i++)
|
|
run_with_memfd_hugetlb(test_case->fn, test_case->desc,
|
|
hugetlbsizes[i]);
|
|
}
|
|
|
|
static void run_non_anon_test_cases(void)
|
|
{
|
|
int i;
|
|
|
|
ksft_print_msg("[RUN] Non-anonymous memory tests in private mappings\n");
|
|
|
|
for (i = 0; i < ARRAY_SIZE(non_anon_test_cases); i++)
|
|
run_non_anon_test_case(&non_anon_test_cases[i]);
|
|
}
|
|
|
|
static int tests_per_non_anon_test_case(void)
|
|
{
|
|
int tests = 3 + nr_hugetlbsizes;
|
|
|
|
if (pmdsize)
|
|
tests += 1;
|
|
return tests;
|
|
}
|
|
|
|
int main(int argc, char **argv)
|
|
{
|
|
int err;
|
|
struct thp_settings default_settings;
|
|
|
|
ksft_print_header();
|
|
|
|
pagesize = getpagesize();
|
|
pmdsize = read_pmd_pagesize();
|
|
if (pmdsize) {
|
|
/* Only if THP is supported. */
|
|
thp_read_settings(&default_settings);
|
|
default_settings.hugepages[sz2ord(pmdsize)].enabled = THP_INHERIT;
|
|
thp_save_settings();
|
|
thp_push_settings(&default_settings);
|
|
|
|
ksft_print_msg("[INFO] detected PMD size: %zu KiB\n",
|
|
pmdsize / 1024);
|
|
nr_thpsizes = detect_thp_sizes(thpsizes, ARRAY_SIZE(thpsizes));
|
|
}
|
|
nr_hugetlbsizes = detect_hugetlb_page_sizes(hugetlbsizes,
|
|
ARRAY_SIZE(hugetlbsizes));
|
|
detect_huge_zeropage();
|
|
|
|
ksft_set_plan(ARRAY_SIZE(anon_test_cases) * tests_per_anon_test_case() +
|
|
ARRAY_SIZE(anon_thp_test_cases) * tests_per_anon_thp_test_case() +
|
|
ARRAY_SIZE(non_anon_test_cases) * tests_per_non_anon_test_case());
|
|
|
|
gup_fd = open("/sys/kernel/debug/gup_test", O_RDWR);
|
|
pagemap_fd = open("/proc/self/pagemap", O_RDONLY);
|
|
if (pagemap_fd < 0)
|
|
ksft_exit_fail_msg("opening pagemap failed\n");
|
|
|
|
run_anon_test_cases();
|
|
run_anon_thp_test_cases();
|
|
run_non_anon_test_cases();
|
|
|
|
if (pmdsize) {
|
|
/* Only if THP is supported. */
|
|
thp_restore_settings();
|
|
}
|
|
|
|
err = ksft_get_fail_cnt();
|
|
if (err)
|
|
ksft_exit_fail_msg("%d out of %d tests failed\n",
|
|
err, ksft_test_num());
|
|
ksft_exit_pass();
|
|
}
|