root/tools/testing/selftests/mm/cow.c
// SPDX-License-Identifier: GPL-2.0-only
/*
 * COW (Copy On Write) tests.
 *
 * Copyright 2022, Red Hat, Inc.
 *
 * Author(s): David Hildenbrand <david@redhat.com>
 */
#define _GNU_SOURCE
#include <stdlib.h>
#include <string.h>
#include <stdbool.h>
#include <stdint.h>
#include <unistd.h>
#include <errno.h>
#include <fcntl.h>
#include <assert.h>
#include <linux/mman.h>
#include <sys/mman.h>
#include <sys/ioctl.h>
#include <sys/wait.h>
#include <linux/memfd.h>

#include "local_config.h"
#ifdef LOCAL_CONFIG_HAVE_LIBURING
#include <liburing.h>
#endif /* LOCAL_CONFIG_HAVE_LIBURING */

#include "../../../../mm/gup_test.h"
#include "kselftest.h"
#include "vm_util.h"
#include "thp_settings.h"

static size_t pagesize;
static int pagemap_fd;
static size_t pmdsize;
static int nr_thpsizes;
static size_t thpsizes[20];
static int nr_hugetlbsizes;
static size_t hugetlbsizes[10];
static int gup_fd;
static bool has_huge_zeropage;

static int detect_thp_sizes(size_t sizes[], int max)
{
        int count = 0;
        unsigned long orders;
        size_t kb;
        int i;

        /* thp not supported at all. */
        if (!pmdsize)
                return 0;

        orders = 1UL << sz2ord(pmdsize, pagesize);
        orders |= thp_supported_orders();

        for (i = 0; orders && count < max; i++) {
                if (!(orders & (1UL << i)))
                        continue;
                orders &= ~(1UL << i);
                kb = (pagesize >> 10) << i;
                sizes[count++] = kb * 1024;
                ksft_print_msg("[INFO] detected THP size: %zu KiB\n", kb);
        }

        return count;
}

static bool range_is_swapped(void *addr, size_t size)
{
        for (; size; addr += pagesize, size -= pagesize)
                if (!pagemap_is_swapped(pagemap_fd, addr))
                        return false;
        return true;
}

static bool populate_page_checked(char *addr)
{
        bool ret;

        FORCE_READ(*addr);
        ret = pagemap_is_populated(pagemap_fd, addr);
        if (!ret)
                ksft_print_msg("Failed to populate page\n");

        return ret;
}

struct comm_pipes {
        int child_ready[2];
        int parent_ready[2];
};

static int setup_comm_pipes(struct comm_pipes *comm_pipes)
{
        if (pipe(comm_pipes->child_ready) < 0) {
                ksft_perror("pipe() failed");
                return -errno;
        }
        if (pipe(comm_pipes->parent_ready) < 0) {
                ksft_perror("pipe() failed");
                close(comm_pipes->child_ready[0]);
                close(comm_pipes->child_ready[1]);
                return -errno;
        }

        return 0;
}

static void close_comm_pipes(struct comm_pipes *comm_pipes)
{
        close(comm_pipes->child_ready[0]);
        close(comm_pipes->child_ready[1]);
        close(comm_pipes->parent_ready[0]);
        close(comm_pipes->parent_ready[1]);
}

static int child_memcmp_fn(char *mem, size_t size,
                           struct comm_pipes *comm_pipes)
{
        char *old = malloc(size);
        char buf;

        /* Backup the original content. */
        memcpy(old, mem, size);

        /* Wait until the parent modified the page. */
        write(comm_pipes->child_ready[1], "0", 1);
        while (read(comm_pipes->parent_ready[0], &buf, 1) != 1)
                ;

        /* See if we still read the old values. */
        return memcmp(old, mem, size);
}

static int child_vmsplice_memcmp_fn(char *mem, size_t size,
                                    struct comm_pipes *comm_pipes)
{
        struct iovec iov = {
                .iov_base = mem,
                .iov_len = size,
        };
        ssize_t cur, total, transferred;
        char *old, *new;
        int fds[2];
        char buf;

        old = malloc(size);
        new = malloc(size);

        /* Backup the original content. */
        memcpy(old, mem, size);

        if (pipe(fds) < 0)
                return -errno;

        /* Trigger a read-only pin. */
        transferred = vmsplice(fds[1], &iov, 1, 0);
        if (transferred < 0)
                return -errno;
        if (transferred == 0)
                return -EINVAL;

        /* Unmap it from our page tables. */
        if (munmap(mem, size) < 0)
                return -errno;

        /* Wait until the parent modified it. */
        write(comm_pipes->child_ready[1], "0", 1);
        while (read(comm_pipes->parent_ready[0], &buf, 1) != 1)
                ;

        /* See if we still read the old values via the pipe. */
        for (total = 0; total < transferred; total += cur) {
                cur = read(fds[0], new + total, transferred - total);
                if (cur < 0)
                        return -errno;
        }

        return memcmp(old, new, transferred);
}

typedef int (*child_fn)(char *mem, size_t size, struct comm_pipes *comm_pipes);

static void do_test_cow_in_parent(char *mem, size_t size, bool do_mprotect,
                child_fn fn, bool xfail)
{
        struct comm_pipes comm_pipes;
        char buf;
        int ret;

        ret = setup_comm_pipes(&comm_pipes);
        if (ret) {
                log_test_result(KSFT_FAIL);
                return;
        }

        ret = fork();
        if (ret < 0) {
                ksft_perror("fork() failed");
                log_test_result(KSFT_FAIL);
                goto close_comm_pipes;
        } else if (!ret) {
                exit(fn(mem, size, &comm_pipes));
        }

        while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
                ;

        if (do_mprotect) {
                /*
                 * mprotect() optimizations might try avoiding
                 * write-faults by directly mapping pages writable.
                 */
                ret = mprotect(mem, size, PROT_READ);
                if (ret) {
                        ksft_perror("mprotect() failed");
                        log_test_result(KSFT_FAIL);
                        write(comm_pipes.parent_ready[1], "0", 1);
                        wait(&ret);
                        goto close_comm_pipes;
                }

                ret = mprotect(mem, size, PROT_READ|PROT_WRITE);
                if (ret) {
                        ksft_perror("mprotect() failed");
                        log_test_result(KSFT_FAIL);
                        write(comm_pipes.parent_ready[1], "0", 1);
                        wait(&ret);
                        goto close_comm_pipes;
                }
        }

        /* Modify the page. */
        memset(mem, 0xff, size);
        write(comm_pipes.parent_ready[1], "0", 1);

        wait(&ret);
        if (WIFEXITED(ret))
                ret = WEXITSTATUS(ret);
        else
                ret = -EINVAL;

        if (!ret) {
                log_test_result(KSFT_PASS);
        } else if (xfail) {
                /*
                 * With hugetlb, some vmsplice() tests are currently expected to
                 * fail because (a) harder to fix and (b) nobody really cares.
                 * Flag them as expected failure for now.
                 */
                ksft_print_msg("Leak from parent into child\n");
                log_test_result(KSFT_XFAIL);
        } else {
                ksft_print_msg("Leak from parent into child\n");
                log_test_result(KSFT_FAIL);
        }
close_comm_pipes:
        close_comm_pipes(&comm_pipes);
}

static void test_cow_in_parent(char *mem, size_t size, bool is_hugetlb)
{
        do_test_cow_in_parent(mem, size, false, child_memcmp_fn, false);
}

static void test_cow_in_parent_mprotect(char *mem, size_t size, bool is_hugetlb)
{
        do_test_cow_in_parent(mem, size, true, child_memcmp_fn, false);
}

static void test_vmsplice_in_child(char *mem, size_t size, bool is_hugetlb)
{
        do_test_cow_in_parent(mem, size, false, child_vmsplice_memcmp_fn,
                              is_hugetlb);
}

static void test_vmsplice_in_child_mprotect(char *mem, size_t size,
                bool is_hugetlb)
{
        do_test_cow_in_parent(mem, size, true, child_vmsplice_memcmp_fn,
                              is_hugetlb);
}

static void do_test_vmsplice_in_parent(char *mem, size_t size,
                                       bool before_fork, bool xfail)
{
        struct iovec iov = {
                .iov_base = mem,
                .iov_len = size,
        };
        ssize_t cur, total, transferred = 0;
        struct comm_pipes comm_pipes;
        char *old, *new;
        int ret, fds[2];
        char buf;

        old = malloc(size);
        new = malloc(size);

        memcpy(old, mem, size);

        ret = setup_comm_pipes(&comm_pipes);
        if (ret) {
                log_test_result(KSFT_FAIL);
                goto free;
        }

        if (pipe(fds) < 0) {
                ksft_perror("pipe() failed");
                log_test_result(KSFT_FAIL);
                goto close_comm_pipes;
        }

        if (before_fork) {
                transferred = vmsplice(fds[1], &iov, 1, 0);
                if (transferred <= 0) {
                        ksft_perror("vmsplice() failed\n");
                        log_test_result(KSFT_FAIL);
                        goto close_pipe;
                }
        }

        ret = fork();
        if (ret < 0) {
                ksft_perror("fork() failed\n");
                log_test_result(KSFT_FAIL);
                goto close_pipe;
        } else if (!ret) {
                write(comm_pipes.child_ready[1], "0", 1);
                while (read(comm_pipes.parent_ready[0], &buf, 1) != 1)
                        ;
                /* Modify page content in the child. */
                memset(mem, 0xff, size);
                exit(0);
        }

        if (!before_fork) {
                transferred = vmsplice(fds[1], &iov, 1, 0);
                if (transferred <= 0) {
                        ksft_perror("vmsplice() failed");
                        log_test_result(KSFT_FAIL);
                        wait(&ret);
                        goto close_pipe;
                }
        }

        while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
                ;
        if (munmap(mem, size) < 0) {
                ksft_perror("munmap() failed");
                log_test_result(KSFT_FAIL);
                goto close_pipe;
        }
        write(comm_pipes.parent_ready[1], "0", 1);

        /* Wait until the child is done writing. */
        wait(&ret);
        if (!WIFEXITED(ret)) {
                ksft_perror("wait() failed");
                log_test_result(KSFT_FAIL);
                goto close_pipe;
        }

        /* See if we still read the old values. */
        for (total = 0; total < transferred; total += cur) {
                cur = read(fds[0], new + total, transferred - total);
                if (cur < 0) {
                        ksft_perror("read() failed");
                        log_test_result(KSFT_FAIL);
                        goto close_pipe;
                }
        }

        if (!memcmp(old, new, transferred)) {
                log_test_result(KSFT_PASS);
        } else if (xfail) {
                /*
                 * With hugetlb, some vmsplice() tests are currently expected to
                 * fail because (a) harder to fix and (b) nobody really cares.
                 * Flag them as expected failure for now.
                 */
                ksft_print_msg("Leak from child into parent\n");
                log_test_result(KSFT_XFAIL);
        } else {
                ksft_print_msg("Leak from child into parent\n");
                log_test_result(KSFT_FAIL);
        }
close_pipe:
        close(fds[0]);
        close(fds[1]);
close_comm_pipes:
        close_comm_pipes(&comm_pipes);
free:
        free(old);
        free(new);
}

static void test_vmsplice_before_fork(char *mem, size_t size, bool is_hugetlb)
{
        do_test_vmsplice_in_parent(mem, size, true, is_hugetlb);
}

static void test_vmsplice_after_fork(char *mem, size_t size, bool is_hugetlb)
{
        do_test_vmsplice_in_parent(mem, size, false, is_hugetlb);
}

#ifdef LOCAL_CONFIG_HAVE_LIBURING
static void do_test_iouring(char *mem, size_t size, bool use_fork)
{
        struct comm_pipes comm_pipes;
        struct io_uring_cqe *cqe;
        struct io_uring_sqe *sqe;
        struct io_uring ring;
        ssize_t cur, total;
        struct iovec iov;
        char *buf, *tmp;
        int ret, fd;
        FILE *file;

        ret = setup_comm_pipes(&comm_pipes);
        if (ret) {
                log_test_result(KSFT_FAIL);
                return;
        }

        file = tmpfile();
        if (!file) {
                ksft_perror("tmpfile() failed");
                log_test_result(KSFT_FAIL);
                goto close_comm_pipes;
        }
        fd = fileno(file);
        assert(fd);

        tmp = malloc(size);
        if (!tmp) {
                ksft_print_msg("malloc() failed\n");
                log_test_result(KSFT_FAIL);
                goto close_file;
        }

        /* Skip on errors, as we might just lack kernel support. */
        ret = io_uring_queue_init(1, &ring, 0);
        if (ret < 0) {
                ksft_print_msg("io_uring_queue_init() failed\n");
                log_test_result(KSFT_SKIP);
                goto free_tmp;
        }

        /*
         * Register the range as a fixed buffer. This will FOLL_WRITE | FOLL_PIN
         * | FOLL_LONGTERM the range.
         *
         * Skip on errors, as we might just lack kernel support or might not
         * have sufficient MEMLOCK permissions.
         */
        iov.iov_base = mem;
        iov.iov_len = size;
        ret = io_uring_register_buffers(&ring, &iov, 1);
        if (ret) {
                ksft_print_msg("io_uring_register_buffers() failed\n");
                log_test_result(KSFT_SKIP);
                goto queue_exit;
        }

        if (use_fork) {
                /*
                 * fork() and keep the child alive until we're done. Note that
                 * we expect the pinned page to not get shared with the child.
                 */
                ret = fork();
                if (ret < 0) {
                        ksft_perror("fork() failed");
                        log_test_result(KSFT_FAIL);
                        goto unregister_buffers;
                } else if (!ret) {
                        write(comm_pipes.child_ready[1], "0", 1);
                        while (read(comm_pipes.parent_ready[0], &buf, 1) != 1)
                                ;
                        exit(0);
                }

                while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
                        ;
        } else {
                /*
                 * Map the page R/O into the page table. Enable softdirty
                 * tracking to stop the page from getting mapped R/W immediately
                 * again by mprotect() optimizations. Note that we don't have an
                 * easy way to test if that worked (the pagemap does not export
                 * if the page is mapped R/O vs. R/W).
                 */
                ret = mprotect(mem, size, PROT_READ);
                if (ret) {
                        ksft_perror("mprotect() failed");
                        log_test_result(KSFT_FAIL);
                        goto unregister_buffers;
                }

                clear_softdirty();
                ret = mprotect(mem, size, PROT_READ | PROT_WRITE);
                if (ret) {
                        ksft_perror("mprotect() failed");
                        log_test_result(KSFT_FAIL);
                        goto unregister_buffers;
                }
        }

        /*
         * Modify the page and write page content as observed by the fixed
         * buffer pin to the file so we can verify it.
         */
        memset(mem, 0xff, size);
        sqe = io_uring_get_sqe(&ring);
        if (!sqe) {
                ksft_print_msg("io_uring_get_sqe() failed\n");
                log_test_result(KSFT_FAIL);
                goto quit_child;
        }
        io_uring_prep_write_fixed(sqe, fd, mem, size, 0, 0);

        ret = io_uring_submit(&ring);
        if (ret < 0) {
                ksft_print_msg("io_uring_submit() failed\n");
                log_test_result(KSFT_FAIL);
                goto quit_child;
        }

        ret = io_uring_wait_cqe(&ring, &cqe);
        if (ret < 0) {
                ksft_print_msg("io_uring_wait_cqe() failed\n");
                log_test_result(KSFT_FAIL);
                goto quit_child;
        }

        if (cqe->res != size) {
                ksft_print_msg("write_fixed failed\n");
                log_test_result(KSFT_FAIL);
                goto quit_child;
        }
        io_uring_cqe_seen(&ring, cqe);

        /* Read back the file content to the temporary buffer. */
        total = 0;
        while (total < size) {
                cur = pread(fd, tmp + total, size - total, total);
                if (cur < 0) {
                        ksft_perror("pread() failed\n");
                        log_test_result(KSFT_FAIL);
                        goto quit_child;
                }
                total += cur;
        }

        /* Finally, check if we read what we expected. */
        if (!memcmp(mem, tmp, size)) {
                log_test_result(KSFT_PASS);
        } else {
                ksft_print_msg("Longtom R/W pin is not reliable\n");
                log_test_result(KSFT_FAIL);
        }

quit_child:
        if (use_fork) {
                write(comm_pipes.parent_ready[1], "0", 1);
                wait(&ret);
        }
unregister_buffers:
        io_uring_unregister_buffers(&ring);
queue_exit:
        io_uring_queue_exit(&ring);
free_tmp:
        free(tmp);
close_file:
        fclose(file);
close_comm_pipes:
        close_comm_pipes(&comm_pipes);
}

static void test_iouring_ro(char *mem, size_t size, bool is_hugetlb)
{
        do_test_iouring(mem, size, false);
}

static void test_iouring_fork(char *mem, size_t size, bool is_hugetlb)
{
        do_test_iouring(mem, size, true);
}

#endif /* LOCAL_CONFIG_HAVE_LIBURING */

enum ro_pin_test {
        RO_PIN_TEST,
        RO_PIN_TEST_SHARED,
        RO_PIN_TEST_PREVIOUSLY_SHARED,
        RO_PIN_TEST_RO_EXCLUSIVE,
};

static void do_test_ro_pin(char *mem, size_t size, enum ro_pin_test test,
                           bool fast)
{
        struct pin_longterm_test args;
        struct comm_pipes comm_pipes;
        char *tmp, buf;
        __u64 tmp_val;
        int ret;

        if (gup_fd < 0) {
                ksft_print_msg("gup_test not available\n");
                log_test_result(KSFT_SKIP);
                return;
        }

        tmp = malloc(size);
        if (!tmp) {
                ksft_perror("malloc() failed\n");
                log_test_result(KSFT_FAIL);
                return;
        }

        ret = setup_comm_pipes(&comm_pipes);
        if (ret) {
                log_test_result(KSFT_FAIL);
                goto free_tmp;
        }

        switch (test) {
        case RO_PIN_TEST:
                break;
        case RO_PIN_TEST_SHARED:
        case RO_PIN_TEST_PREVIOUSLY_SHARED:
                /*
                 * Share the pages with our child. As the pages are not pinned,
                 * this should just work.
                 */
                ret = fork();
                if (ret < 0) {
                        ksft_perror("fork() failed");
                        log_test_result(KSFT_FAIL);
                        goto close_comm_pipes;
                } else if (!ret) {
                        write(comm_pipes.child_ready[1], "0", 1);
                        while (read(comm_pipes.parent_ready[0], &buf, 1) != 1)
                                ;
                        exit(0);
                }

                /* Wait until our child is ready. */
                while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
                        ;

                if (test == RO_PIN_TEST_PREVIOUSLY_SHARED) {
                        /*
                         * Tell the child to quit now and wait until it quit.
                         * The pages should now be mapped R/O into our page
                         * tables, but they are no longer shared.
                         */
                        write(comm_pipes.parent_ready[1], "0", 1);
                        wait(&ret);
                        if (!WIFEXITED(ret))
                                ksft_print_msg("[INFO] wait() failed\n");
                }
                break;
        case RO_PIN_TEST_RO_EXCLUSIVE:
                /*
                 * Map the page R/O into the page table. Enable softdirty
                 * tracking to stop the page from getting mapped R/W immediately
                 * again by mprotect() optimizations. Note that we don't have an
                 * easy way to test if that worked (the pagemap does not export
                 * if the page is mapped R/O vs. R/W).
                 */
                ret = mprotect(mem, size, PROT_READ);
                clear_softdirty();
                ret |= mprotect(mem, size, PROT_READ | PROT_WRITE);
                if (ret) {
                        ksft_perror("mprotect() failed");
                        log_test_result(KSFT_FAIL);
                        goto close_comm_pipes;
                }
                break;
        default:
                assert(false);
        }

        /* Take a R/O pin. This should trigger unsharing. */
        args.addr = (__u64)(uintptr_t)mem;
        args.size = size;
        args.flags = fast ? PIN_LONGTERM_TEST_FLAG_USE_FAST : 0;
        ret = ioctl(gup_fd, PIN_LONGTERM_TEST_START, &args);
        if (ret) {
                if (errno == EINVAL)
                        ret = KSFT_SKIP;
                else
                        ret = KSFT_FAIL;
                ksft_perror("PIN_LONGTERM_TEST_START failed");
                log_test_result(ret);
                goto wait;
        }

        /* Modify the page. */
        memset(mem, 0xff, size);

        /*
         * Read back the content via the pin to the temporary buffer and
         * test if we observed the modification.
         */
        tmp_val = (__u64)(uintptr_t)tmp;
        ret = ioctl(gup_fd, PIN_LONGTERM_TEST_READ, &tmp_val);
        if (ret) {
                ksft_perror("PIN_LONGTERM_TEST_READ failed");
                log_test_result(KSFT_FAIL);
        } else {
                if (!memcmp(mem, tmp, size)) {
                        log_test_result(KSFT_PASS);
                } else {
                        ksft_print_msg("Longterm R/O pin is not reliable\n");
                        log_test_result(KSFT_FAIL);
                }
        }

        ret = ioctl(gup_fd, PIN_LONGTERM_TEST_STOP);
        if (ret)
                ksft_perror("PIN_LONGTERM_TEST_STOP failed");
wait:
        switch (test) {
        case RO_PIN_TEST_SHARED:
                write(comm_pipes.parent_ready[1], "0", 1);
                wait(&ret);
                if (!WIFEXITED(ret))
                        ksft_perror("wait() failed");
                break;
        default:
                break;
        }
close_comm_pipes:
        close_comm_pipes(&comm_pipes);
free_tmp:
        free(tmp);
}

static void test_ro_pin_on_shared(char *mem, size_t size, bool is_hugetlb)
{
        do_test_ro_pin(mem, size, RO_PIN_TEST_SHARED, false);
}

static void test_ro_fast_pin_on_shared(char *mem, size_t size, bool is_hugetlb)
{
        do_test_ro_pin(mem, size, RO_PIN_TEST_SHARED, true);
}

static void test_ro_pin_on_ro_previously_shared(char *mem, size_t size,
                bool is_hugetlb)
{
        do_test_ro_pin(mem, size, RO_PIN_TEST_PREVIOUSLY_SHARED, false);
}

static void test_ro_fast_pin_on_ro_previously_shared(char *mem, size_t size,
                bool is_hugetlb)
{
        do_test_ro_pin(mem, size, RO_PIN_TEST_PREVIOUSLY_SHARED, true);
}

static void test_ro_pin_on_ro_exclusive(char *mem, size_t size,
                bool is_hugetlb)
{
        do_test_ro_pin(mem, size, RO_PIN_TEST_RO_EXCLUSIVE, false);
}

static void test_ro_fast_pin_on_ro_exclusive(char *mem, size_t size,
                bool is_hugetlb)
{
        do_test_ro_pin(mem, size, RO_PIN_TEST_RO_EXCLUSIVE, true);
}

typedef void (*test_fn)(char *mem, size_t size, bool hugetlb);

static void do_run_with_base_page(test_fn fn, bool swapout)
{
        char *mem;
        int ret;

        mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE,
                   MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
        if (mem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                return;
        }

        ret = madvise(mem, pagesize, MADV_NOHUGEPAGE);
        /* Ignore if not around on a kernel. */
        if (ret && errno != EINVAL) {
                ksft_perror("MADV_NOHUGEPAGE failed");
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        /* Populate a base page. */
        memset(mem, 1, pagesize);

        if (swapout) {
                madvise(mem, pagesize, MADV_PAGEOUT);
                if (!pagemap_is_swapped(pagemap_fd, mem)) {
                        ksft_print_msg("MADV_PAGEOUT did not work, is swap enabled?\n");
                        log_test_result(KSFT_SKIP);
                        goto munmap;
                }
        }

        fn(mem, pagesize, false);
munmap:
        munmap(mem, pagesize);
}

static void run_with_base_page(test_fn fn, const char *desc)
{
        log_test_start("%s ... with base page", desc);
        do_run_with_base_page(fn, false);
}

static void run_with_base_page_swap(test_fn fn, const char *desc)
{
        log_test_start("%s ... with swapped out base page", desc);
        do_run_with_base_page(fn, true);
}

enum thp_run {
        THP_RUN_PMD,
        THP_RUN_PMD_SWAPOUT,
        THP_RUN_PTE,
        THP_RUN_PTE_SWAPOUT,
        THP_RUN_SINGLE_PTE,
        THP_RUN_SINGLE_PTE_SWAPOUT,
        THP_RUN_PARTIAL_MREMAP,
        THP_RUN_PARTIAL_SHARED,
};

static void do_run_with_thp(test_fn fn, enum thp_run thp_run, size_t thpsize)
{
        char *mem, *mmap_mem, *tmp, *mremap_mem = MAP_FAILED;
        size_t size, mmap_size, mremap_size;
        int ret;

        /* For alignment purposes, we need twice the thp size. */
        mmap_size = 2 * thpsize;
        mmap_mem = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
                        MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
        if (mmap_mem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                return;
        }

        /* We need a THP-aligned memory area. */
        mem = (char *)(((uintptr_t)mmap_mem + thpsize) & ~(thpsize - 1));

        ret = madvise(mem, thpsize, MADV_HUGEPAGE);
        if (ret) {
                ksft_perror("MADV_HUGEPAGE failed");
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        /*
         * Try to populate a THP. Touch the first sub-page and test if
         * we get the last sub-page populated automatically.
         */
        mem[0] = 1;
        if (!pagemap_is_populated(pagemap_fd, mem + thpsize - pagesize)) {
                ksft_print_msg("Did not get a THP populated\n");
                log_test_result(KSFT_SKIP);
                goto munmap;
        }
        memset(mem, 1, thpsize);

        size = thpsize;
        switch (thp_run) {
        case THP_RUN_PMD:
        case THP_RUN_PMD_SWAPOUT:
                assert(thpsize == pmdsize);
                break;
        case THP_RUN_PTE:
        case THP_RUN_PTE_SWAPOUT:
                /*
                 * Trigger PTE-mapping the THP by temporarily mapping a single
                 * subpage R/O. This is a noop if the THP is not pmdsize (and
                 * therefore already PTE-mapped).
                 */
                ret = mprotect(mem + pagesize, pagesize, PROT_READ);
                if (ret) {
                        ksft_perror("mprotect() failed");
                        log_test_result(KSFT_FAIL);
                        goto munmap;
                }
                ret = mprotect(mem + pagesize, pagesize, PROT_READ | PROT_WRITE);
                if (ret) {
                        ksft_perror("mprotect() failed");
                        log_test_result(KSFT_FAIL);
                        goto munmap;
                }
                break;
        case THP_RUN_SINGLE_PTE:
        case THP_RUN_SINGLE_PTE_SWAPOUT:
                /*
                 * Discard all but a single subpage of that PTE-mapped THP. What
                 * remains is a single PTE mapping a single subpage.
                 */
                ret = madvise(mem + pagesize, thpsize - pagesize, MADV_DONTNEED);
                if (ret) {
                        ksft_perror("MADV_DONTNEED failed");
                        log_test_result(KSFT_FAIL);
                        goto munmap;
                }
                size = pagesize;
                break;
        case THP_RUN_PARTIAL_MREMAP:
                /*
                 * Remap half of the THP. We need some new memory location
                 * for that.
                 */
                mremap_size = thpsize / 2;
                mremap_mem = mmap(NULL, mremap_size, PROT_NONE,
                                  MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
                if (mremap_mem == MAP_FAILED) {
                        ksft_perror("mmap() failed");
                        log_test_result(KSFT_FAIL);
                        goto munmap;
                }
                tmp = mremap(mem + mremap_size, mremap_size, mremap_size,
                             MREMAP_MAYMOVE | MREMAP_FIXED, mremap_mem);
                if (tmp != mremap_mem) {
                        ksft_perror("mremap() failed");
                        log_test_result(KSFT_FAIL);
                        goto munmap;
                }
                size = mremap_size;
                break;
        case THP_RUN_PARTIAL_SHARED:
                /*
                 * Share the first page of the THP with a child and quit the
                 * child. This will result in some parts of the THP never
                 * have been shared.
                 */
                ret = madvise(mem + pagesize, thpsize - pagesize, MADV_DONTFORK);
                if (ret) {
                        ksft_perror("MADV_DONTFORK failed");
                        log_test_result(KSFT_FAIL);
                        goto munmap;
                }
                ret = fork();
                if (ret < 0) {
                        ksft_perror("fork() failed");
                        log_test_result(KSFT_FAIL);
                        goto munmap;
                } else if (!ret) {
                        exit(0);
                }
                wait(&ret);
                /* Allow for sharing all pages again. */
                ret = madvise(mem + pagesize, thpsize - pagesize, MADV_DOFORK);
                if (ret) {
                        ksft_perror("MADV_DOFORK failed");
                        log_test_result(KSFT_FAIL);
                        goto munmap;
                }
                break;
        default:
                assert(false);
        }

        switch (thp_run) {
        case THP_RUN_PMD_SWAPOUT:
        case THP_RUN_PTE_SWAPOUT:
        case THP_RUN_SINGLE_PTE_SWAPOUT:
                madvise(mem, size, MADV_PAGEOUT);
                if (!range_is_swapped(mem, size)) {
                        ksft_print_msg("MADV_PAGEOUT did not work, is swap enabled?\n");
                        log_test_result(KSFT_SKIP);
                        goto munmap;
                }
                break;
        default:
                break;
        }

        fn(mem, size, false);
munmap:
        munmap(mmap_mem, mmap_size);
        if (mremap_mem != MAP_FAILED)
                munmap(mremap_mem, mremap_size);
}

static void run_with_thp(test_fn fn, const char *desc, size_t size)
{
        log_test_start("%s ... with THP (%zu kB)",
                desc, size / 1024);
        do_run_with_thp(fn, THP_RUN_PMD, size);
}

static void run_with_thp_swap(test_fn fn, const char *desc, size_t size)
{
        log_test_start("%s ... with swapped-out THP (%zu kB)",
                desc, size / 1024);
        do_run_with_thp(fn, THP_RUN_PMD_SWAPOUT, size);
}

static void run_with_pte_mapped_thp(test_fn fn, const char *desc, size_t size)
{
        log_test_start("%s ... with PTE-mapped THP (%zu kB)",
                desc, size / 1024);
        do_run_with_thp(fn, THP_RUN_PTE, size);
}

static void run_with_pte_mapped_thp_swap(test_fn fn, const char *desc, size_t size)
{
        log_test_start("%s ... with swapped-out, PTE-mapped THP (%zu kB)",
                desc, size / 1024);
        do_run_with_thp(fn, THP_RUN_PTE_SWAPOUT, size);
}

static void run_with_single_pte_of_thp(test_fn fn, const char *desc, size_t size)
{
        log_test_start("%s ... with single PTE of THP (%zu kB)",
                desc, size / 1024);
        do_run_with_thp(fn, THP_RUN_SINGLE_PTE, size);
}

static void run_with_single_pte_of_thp_swap(test_fn fn, const char *desc, size_t size)
{
        log_test_start("%s ... with single PTE of swapped-out THP (%zu kB)",
                desc, size / 1024);
        do_run_with_thp(fn, THP_RUN_SINGLE_PTE_SWAPOUT, size);
}

static void run_with_partial_mremap_thp(test_fn fn, const char *desc, size_t size)
{
        log_test_start("%s ... with partially mremap()'ed THP (%zu kB)",
                desc, size / 1024);
        do_run_with_thp(fn, THP_RUN_PARTIAL_MREMAP, size);
}

static void run_with_partial_shared_thp(test_fn fn, const char *desc, size_t size)
{
        log_test_start("%s ... with partially shared THP (%zu kB)",
                desc, size / 1024);
        do_run_with_thp(fn, THP_RUN_PARTIAL_SHARED, size);
}

static void run_with_hugetlb(test_fn fn, const char *desc, size_t hugetlbsize)
{
        int flags = MAP_PRIVATE | MAP_ANONYMOUS | MAP_HUGETLB;
        char *mem, *dummy;

        log_test_start("%s ... with hugetlb (%zu kB)", desc,
                       hugetlbsize / 1024);

        flags |= __builtin_ctzll(hugetlbsize) << MAP_HUGE_SHIFT;

        mem = mmap(NULL, hugetlbsize, PROT_READ | PROT_WRITE, flags, -1, 0);
        if (mem == MAP_FAILED) {
                ksft_perror("need more free huge pages");
                log_test_result(KSFT_SKIP);
                return;
        }

        /* Populate an huge page. */
        memset(mem, 1, hugetlbsize);

        /*
         * We need a total of two hugetlb pages to handle COW/unsharing
         * properly, otherwise we might get zapped by a SIGBUS.
         */
        dummy = mmap(NULL, hugetlbsize, PROT_READ | PROT_WRITE, flags, -1, 0);
        if (dummy == MAP_FAILED) {
                ksft_perror("need more free huge pages");
                log_test_result(KSFT_SKIP);
                goto munmap;
        }
        munmap(dummy, hugetlbsize);

        fn(mem, hugetlbsize, true);
munmap:
        munmap(mem, hugetlbsize);
}

struct test_case {
        const char *desc;
        test_fn fn;
};

/*
 * Test cases that are specific to anonymous pages: pages in private mappings
 * that may get shared via COW during fork().
 */
static const struct test_case anon_test_cases[] = {
        /*
         * Basic COW tests for fork() without any GUP. If we miss to break COW,
         * either the child can observe modifications by the parent or the
         * other way around.
         */
        {
                "Basic COW after fork()",
                test_cow_in_parent,
        },
        /*
         * Basic test, but do an additional mprotect(PROT_READ)+
         * mprotect(PROT_READ|PROT_WRITE) in the parent before write access.
         */
        {
                "Basic COW after fork() with mprotect() optimization",
                test_cow_in_parent_mprotect,
        },
        /*
         * vmsplice() [R/O GUP] + unmap in the child; modify in the parent. If
         * we miss to break COW, the child observes modifications by the parent.
         * This is CVE-2020-29374 reported by Jann Horn.
         */
        {
                "vmsplice() + unmap in child",
                test_vmsplice_in_child,
        },
        /*
         * vmsplice() test, but do an additional mprotect(PROT_READ)+
         * mprotect(PROT_READ|PROT_WRITE) in the parent before write access.
         */
        {
                "vmsplice() + unmap in child with mprotect() optimization",
                test_vmsplice_in_child_mprotect,
        },
        /*
         * vmsplice() [R/O GUP] in parent before fork(), unmap in parent after
         * fork(); modify in the child. If we miss to break COW, the parent
         * observes modifications by the child.
         */
        {
                "vmsplice() before fork(), unmap in parent after fork()",
                test_vmsplice_before_fork,
        },
        /*
         * vmsplice() [R/O GUP] + unmap in parent after fork(); modify in the
         * child. If we miss to break COW, the parent observes modifications by
         * the child.
         */
        {
                "vmsplice() + unmap in parent after fork()",
                test_vmsplice_after_fork,
        },
#ifdef LOCAL_CONFIG_HAVE_LIBURING
        /*
         * Take a R/W longterm pin and then map the page R/O into the page
         * table to trigger a write fault on next access. When modifying the
         * page, the page content must be visible via the pin.
         */
        {
                "R/O-mapping a page registered as iouring fixed buffer",
                test_iouring_ro,
        },
        /*
         * Take a R/W longterm pin and then fork() a child. When modifying the
         * page, the page content must be visible via the pin. We expect the
         * pinned page to not get shared with the child.
         */
        {
                "fork() with an iouring fixed buffer",
                test_iouring_fork,
        },

#endif /* LOCAL_CONFIG_HAVE_LIBURING */
        /*
         * Take a R/O longterm pin on a R/O-mapped shared anonymous page.
         * When modifying the page via the page table, the page content change
         * must be visible via the pin.
         */
        {
                "R/O GUP pin on R/O-mapped shared page",
                test_ro_pin_on_shared,
        },
        /* Same as above, but using GUP-fast. */
        {
                "R/O GUP-fast pin on R/O-mapped shared page",
                test_ro_fast_pin_on_shared,
        },
        /*
         * Take a R/O longterm pin on a R/O-mapped exclusive anonymous page that
         * was previously shared. When modifying the page via the page table,
         * the page content change must be visible via the pin.
         */
        {
                "R/O GUP pin on R/O-mapped previously-shared page",
                test_ro_pin_on_ro_previously_shared,
        },
        /* Same as above, but using GUP-fast. */
        {
                "R/O GUP-fast pin on R/O-mapped previously-shared page",
                test_ro_fast_pin_on_ro_previously_shared,
        },
        /*
         * Take a R/O longterm pin on a R/O-mapped exclusive anonymous page.
         * When modifying the page via the page table, the page content change
         * must be visible via the pin.
         */
        {
                "R/O GUP pin on R/O-mapped exclusive page",
                test_ro_pin_on_ro_exclusive,
        },
        /* Same as above, but using GUP-fast. */
        {
                "R/O GUP-fast pin on R/O-mapped exclusive page",
                test_ro_fast_pin_on_ro_exclusive,
        },
};

static void run_anon_test_case(struct test_case const *test_case)
{
        int i;

        run_with_base_page(test_case->fn, test_case->desc);
        run_with_base_page_swap(test_case->fn, test_case->desc);
        for (i = 0; i < nr_thpsizes; i++) {
                size_t size = thpsizes[i];
                struct thp_settings settings = *thp_current_settings();

                settings.hugepages[sz2ord(pmdsize, pagesize)].enabled = THP_NEVER;
                settings.hugepages[sz2ord(size, pagesize)].enabled = THP_ALWAYS;
                thp_push_settings(&settings);

                if (size == pmdsize) {
                        run_with_thp(test_case->fn, test_case->desc, size);
                        run_with_thp_swap(test_case->fn, test_case->desc, size);
                }

                run_with_pte_mapped_thp(test_case->fn, test_case->desc, size);
                run_with_pte_mapped_thp_swap(test_case->fn, test_case->desc, size);
                run_with_single_pte_of_thp(test_case->fn, test_case->desc, size);
                run_with_single_pte_of_thp_swap(test_case->fn, test_case->desc, size);
                run_with_partial_mremap_thp(test_case->fn, test_case->desc, size);
                run_with_partial_shared_thp(test_case->fn, test_case->desc, size);

                thp_pop_settings();
        }
        for (i = 0; i < nr_hugetlbsizes; i++)
                run_with_hugetlb(test_case->fn, test_case->desc,
                                 hugetlbsizes[i]);
}

static void run_anon_test_cases(void)
{
        int i;

        ksft_print_msg("[INFO] Anonymous memory tests in private mappings\n");

        for (i = 0; i < ARRAY_SIZE(anon_test_cases); i++)
                run_anon_test_case(&anon_test_cases[i]);
}

static int tests_per_anon_test_case(void)
{
        int tests = 2 + nr_hugetlbsizes;

        tests += 6 * nr_thpsizes;
        if (pmdsize)
                tests += 2;
        return tests;
}

enum anon_thp_collapse_test {
        ANON_THP_COLLAPSE_UNSHARED,
        ANON_THP_COLLAPSE_FULLY_SHARED,
        ANON_THP_COLLAPSE_LOWER_SHARED,
        ANON_THP_COLLAPSE_UPPER_SHARED,
};

static void do_test_anon_thp_collapse(char *mem, size_t size,
                                      enum anon_thp_collapse_test test)
{
        struct comm_pipes comm_pipes;
        char buf;
        int ret;

        ret = setup_comm_pipes(&comm_pipes);
        if (ret) {
                log_test_result(KSFT_FAIL);
                return;
        }

        /*
         * Trigger PTE-mapping the THP by temporarily mapping a single subpage
         * R/O, such that we can try collapsing it later.
         */
        ret = mprotect(mem + pagesize, pagesize, PROT_READ);
        if (ret) {
                ksft_perror("mprotect() failed");
                log_test_result(KSFT_FAIL);
                goto close_comm_pipes;
        }
        ret = mprotect(mem + pagesize, pagesize, PROT_READ | PROT_WRITE);
        if (ret) {
                ksft_perror("mprotect() failed");
                log_test_result(KSFT_FAIL);
                goto close_comm_pipes;
        }

        switch (test) {
        case ANON_THP_COLLAPSE_UNSHARED:
                /* Collapse before actually COW-sharing the page. */
                ret = madvise(mem, size, MADV_COLLAPSE);
                if (ret) {
                        ksft_perror("MADV_COLLAPSE failed");
                        log_test_result(KSFT_SKIP);
                        goto close_comm_pipes;
                }
                break;
        case ANON_THP_COLLAPSE_FULLY_SHARED:
                /* COW-share the full PTE-mapped THP. */
                break;
        case ANON_THP_COLLAPSE_LOWER_SHARED:
                /* Don't COW-share the upper part of the THP. */
                ret = madvise(mem + size / 2, size / 2, MADV_DONTFORK);
                if (ret) {
                        ksft_perror("MADV_DONTFORK failed");
                        log_test_result(KSFT_FAIL);
                        goto close_comm_pipes;
                }
                break;
        case ANON_THP_COLLAPSE_UPPER_SHARED:
                /* Don't COW-share the lower part of the THP. */
                ret = madvise(mem, size / 2, MADV_DONTFORK);
                if (ret) {
                        ksft_perror("MADV_DONTFORK failed");
                        log_test_result(KSFT_FAIL);
                        goto close_comm_pipes;
                }
                break;
        default:
                assert(false);
        }

        ret = fork();
        if (ret < 0) {
                ksft_perror("fork() failed");
                log_test_result(KSFT_FAIL);
                goto close_comm_pipes;
        } else if (!ret) {
                switch (test) {
                case ANON_THP_COLLAPSE_UNSHARED:
                case ANON_THP_COLLAPSE_FULLY_SHARED:
                        exit(child_memcmp_fn(mem, size, &comm_pipes));
                        break;
                case ANON_THP_COLLAPSE_LOWER_SHARED:
                        exit(child_memcmp_fn(mem, size / 2, &comm_pipes));
                        break;
                case ANON_THP_COLLAPSE_UPPER_SHARED:
                        exit(child_memcmp_fn(mem + size / 2, size / 2,
                                             &comm_pipes));
                        break;
                default:
                        assert(false);
                }
        }

        while (read(comm_pipes.child_ready[0], &buf, 1) != 1)
                ;

        switch (test) {
        case ANON_THP_COLLAPSE_UNSHARED:
                break;
        case ANON_THP_COLLAPSE_UPPER_SHARED:
        case ANON_THP_COLLAPSE_LOWER_SHARED:
                /*
                 * Revert MADV_DONTFORK such that we merge the VMAs and are
                 * able to actually collapse.
                 */
                ret = madvise(mem, size, MADV_DOFORK);
                if (ret) {
                        ksft_perror("MADV_DOFORK failed");
                        log_test_result(KSFT_FAIL);
                        write(comm_pipes.parent_ready[1], "0", 1);
                        wait(&ret);
                        goto close_comm_pipes;
                }
                /* FALLTHROUGH */
        case ANON_THP_COLLAPSE_FULLY_SHARED:
                /* Collapse before anyone modified the COW-shared page. */
                ret = madvise(mem, size, MADV_COLLAPSE);
                if (ret) {
                        ksft_perror("MADV_COLLAPSE failed");
                        log_test_result(KSFT_SKIP);
                        write(comm_pipes.parent_ready[1], "0", 1);
                        wait(&ret);
                        goto close_comm_pipes;
                }
                break;
        default:
                assert(false);
        }

        /* Modify the page. */
        memset(mem, 0xff, size);
        write(comm_pipes.parent_ready[1], "0", 1);

        wait(&ret);
        if (WIFEXITED(ret))
                ret = WEXITSTATUS(ret);
        else
                ret = -EINVAL;

        if (!ret) {
                log_test_result(KSFT_PASS);
        } else {
                ksft_print_msg("Leak from parent into child\n");
                log_test_result(KSFT_FAIL);
        }
close_comm_pipes:
        close_comm_pipes(&comm_pipes);
}

static void test_anon_thp_collapse_unshared(char *mem, size_t size,
                bool is_hugetlb)
{
        assert(!is_hugetlb);
        do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_UNSHARED);
}

static void test_anon_thp_collapse_fully_shared(char *mem, size_t size,
                bool is_hugetlb)
{
        assert(!is_hugetlb);
        do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_FULLY_SHARED);
}

static void test_anon_thp_collapse_lower_shared(char *mem, size_t size,
                bool is_hugetlb)
{
        assert(!is_hugetlb);
        do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_LOWER_SHARED);
}

static void test_anon_thp_collapse_upper_shared(char *mem, size_t size,
                bool is_hugetlb)
{
        assert(!is_hugetlb);
        do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_UPPER_SHARED);
}

/*
 * Test cases that are specific to anonymous THP: pages in private mappings
 * that may get shared via COW during fork().
 */
static const struct test_case anon_thp_test_cases[] = {
        /*
         * Basic COW test for fork() without any GUP when collapsing a THP
         * before fork().
         *
         * Re-mapping a PTE-mapped anon THP using a single PMD ("in-place
         * collapse") might easily get COW handling wrong when not collapsing
         * exclusivity information properly.
         */
        {
                "Basic COW after fork() when collapsing before fork()",
                test_anon_thp_collapse_unshared,
        },
        /* Basic COW test, but collapse after COW-sharing a full THP. */
        {
                "Basic COW after fork() when collapsing after fork() (fully shared)",
                test_anon_thp_collapse_fully_shared,
        },
        /*
         * Basic COW test, but collapse after COW-sharing the lower half of a
         * THP.
         */
        {
                "Basic COW after fork() when collapsing after fork() (lower shared)",
                test_anon_thp_collapse_lower_shared,
        },
        /*
         * Basic COW test, but collapse after COW-sharing the upper half of a
         * THP.
         */
        {
                "Basic COW after fork() when collapsing after fork() (upper shared)",
                test_anon_thp_collapse_upper_shared,
        },
};

static void run_anon_thp_test_cases(void)
{
        int i;

        if (!pmdsize)
                return;

        ksft_print_msg("[INFO] Anonymous THP tests\n");

        for (i = 0; i < ARRAY_SIZE(anon_thp_test_cases); i++) {
                struct test_case const *test_case = &anon_thp_test_cases[i];

                log_test_start("%s", test_case->desc);
                do_run_with_thp(test_case->fn, THP_RUN_PMD, pmdsize);
        }
}

static int tests_per_anon_thp_test_case(void)
{
        return pmdsize ? 1 : 0;
}

typedef void (*non_anon_test_fn)(char *mem, const char *smem, size_t size);

static void test_cow(char *mem, const char *smem, size_t size)
{
        char *old = malloc(size);

        /* Backup the original content. */
        memcpy(old, smem, size);

        /* Modify the page. */
        memset(mem, 0xff, size);

        /* See if we still read the old values via the other mapping. */
        if (!memcmp(smem, old, size)) {
                log_test_result(KSFT_PASS);
        } else {
                ksft_print_msg("Other mapping modified\n");
                log_test_result(KSFT_FAIL);
        }
        free(old);
}

static void test_ro_pin(char *mem, const char *smem, size_t size)
{
        do_test_ro_pin(mem, size, RO_PIN_TEST, false);
}

static void test_ro_fast_pin(char *mem, const char *smem, size_t size)
{
        do_test_ro_pin(mem, size, RO_PIN_TEST, true);
}

static void run_with_zeropage(non_anon_test_fn fn, const char *desc)
{
        char *mem, *smem;

        log_test_start("%s ... with shared zeropage", desc);

        mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE,
                   MAP_PRIVATE | MAP_ANON, -1, 0);
        if (mem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                return;
        }

        smem = mmap(NULL, pagesize, PROT_READ, MAP_PRIVATE | MAP_ANON, -1, 0);
        if (smem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        /* Read from the page to populate the shared zeropage. */
        if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        fn(mem, smem, pagesize);
munmap:
        munmap(mem, pagesize);
        if (smem != MAP_FAILED)
                munmap(smem, pagesize);
}

static void run_with_huge_zeropage(non_anon_test_fn fn, const char *desc)
{
        char *mem, *smem, *mmap_mem, *mmap_smem;
        size_t mmap_size;
        int ret;

        log_test_start("%s ... with huge zeropage", desc);

        if (!has_huge_zeropage) {
                ksft_print_msg("Huge zeropage not enabled\n");
                log_test_result(KSFT_SKIP);
                return;
        }

        /* For alignment purposes, we need twice the thp size. */
        mmap_size = 2 * pmdsize;
        mmap_mem = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
                        MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
        if (mmap_mem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                return;
        }
        mmap_smem = mmap(NULL, mmap_size, PROT_READ,
                         MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
        if (mmap_smem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        /* We need a THP-aligned memory area. */
        mem = (char *)(((uintptr_t)mmap_mem + pmdsize) & ~(pmdsize - 1));
        smem = (char *)(((uintptr_t)mmap_smem + pmdsize) & ~(pmdsize - 1));

        ret = madvise(mem, pmdsize, MADV_HUGEPAGE);
        if (ret) {
                ksft_perror("madvise()");
                log_test_result(KSFT_FAIL);
                goto munmap;
        }
        ret = madvise(smem, pmdsize, MADV_HUGEPAGE);
        if (ret) {
                ksft_perror("madvise()");
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        /*
         * Read from the memory to populate the huge shared zeropage. Read from
         * the first sub-page and test if we get another sub-page populated
         * automatically.
         */
        if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        if (!pagemap_is_populated(pagemap_fd, mem + pagesize) ||
            !pagemap_is_populated(pagemap_fd, smem + pagesize)) {
                ksft_test_result_skip("Did not get THPs populated\n");
                goto munmap;
        }

        fn(mem, smem, pmdsize);
munmap:
        munmap(mmap_mem, mmap_size);
        if (mmap_smem != MAP_FAILED)
                munmap(mmap_smem, mmap_size);
}

static void run_with_memfd(non_anon_test_fn fn, const char *desc)
{
        char *mem, *smem;
        int fd;

        log_test_start("%s ... with memfd", desc);

        fd = memfd_create("test", 0);
        if (fd < 0) {
                ksft_perror("memfd_create() failed");
                log_test_result(KSFT_FAIL);
                return;
        }

        /* File consists of a single page filled with zeroes. */
        if (fallocate(fd, 0, 0, pagesize)) {
                ksft_perror("fallocate() failed");
                log_test_result(KSFT_FAIL);
                goto close;
        }

        /* Create a private mapping of the memfd. */
        mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0);
        if (mem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                goto close;
        }
        smem = mmap(NULL, pagesize, PROT_READ, MAP_SHARED, fd, 0);
        if (smem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        /* Fault the page in. */
        if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        fn(mem, smem, pagesize);
munmap:
        munmap(mem, pagesize);
        if (smem != MAP_FAILED)
                munmap(smem, pagesize);
close:
        close(fd);
}

static void run_with_tmpfile(non_anon_test_fn fn, const char *desc)
{
        char *mem, *smem;
        FILE *file;
        int fd;

        log_test_start("%s ... with tmpfile", desc);

        file = tmpfile();
        if (!file) {
                ksft_perror("tmpfile() failed");
                log_test_result(KSFT_FAIL);
                return;
        }

        fd = fileno(file);
        if (fd < 0) {
                ksft_perror("fileno() failed");
                log_test_result(KSFT_SKIP);
                return;
        }

        /* File consists of a single page filled with zeroes. */
        if (fallocate(fd, 0, 0, pagesize)) {
                ksft_perror("fallocate() failed");
                log_test_result(KSFT_FAIL);
                goto close;
        }

        /* Create a private mapping of the memfd. */
        mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0);
        if (mem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                goto close;
        }
        smem = mmap(NULL, pagesize, PROT_READ, MAP_SHARED, fd, 0);
        if (smem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        /* Fault the page in. */
        if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        fn(mem, smem, pagesize);
munmap:
        munmap(mem, pagesize);
        if (smem != MAP_FAILED)
                munmap(smem, pagesize);
close:
        fclose(file);
}

static void run_with_memfd_hugetlb(non_anon_test_fn fn, const char *desc,
                                   size_t hugetlbsize)
{
        int flags = MFD_HUGETLB;
        char *mem, *smem;
        int fd;

        log_test_start("%s ... with memfd hugetlb (%zu kB)", desc,
                       hugetlbsize / 1024);

        flags |= __builtin_ctzll(hugetlbsize) << MFD_HUGE_SHIFT;

        fd = memfd_create("test", flags);
        if (fd < 0) {
                ksft_perror("memfd_create() failed");
                log_test_result(KSFT_SKIP);
                return;
        }

        /* File consists of a single page filled with zeroes. */
        if (fallocate(fd, 0, 0, hugetlbsize)) {
                ksft_perror("need more free huge pages");
                log_test_result(KSFT_SKIP);
                goto close;
        }

        /* Create a private mapping of the memfd. */
        mem = mmap(NULL, hugetlbsize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd,
                   0);
        if (mem == MAP_FAILED) {
                ksft_perror("need more free huge pages");
                log_test_result(KSFT_SKIP);
                goto close;
        }
        smem = mmap(NULL, hugetlbsize, PROT_READ, MAP_SHARED, fd, 0);
        if (smem == MAP_FAILED) {
                ksft_perror("mmap() failed");
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        /* Fault the page in. */
        if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
                log_test_result(KSFT_FAIL);
                goto munmap;
        }

        fn(mem, smem, hugetlbsize);
munmap:
        munmap(mem, hugetlbsize);
        if (smem != MAP_FAILED)
                munmap(smem, hugetlbsize);
close:
        close(fd);
}

struct non_anon_test_case {
        const char *desc;
        non_anon_test_fn fn;
};

/*
 * Test cases that target any pages in private mappings that are not anonymous:
 * pages that may get shared via COW ndependent of fork(). This includes
 * the shared zeropage(s), pagecache pages, ...
 */
static const struct non_anon_test_case non_anon_test_cases[] = {
        /*
         * Basic COW test without any GUP. If we miss to break COW, changes are
         * visible via other private/shared mappings.
         */
        {
                "Basic COW",
                test_cow,
        },
        /*
         * Take a R/O longterm pin. When modifying the page via the page table,
         * the page content change must be visible via the pin.
         */
        {
                "R/O longterm GUP pin",
                test_ro_pin,
        },
        /* Same as above, but using GUP-fast. */
        {
                "R/O longterm GUP-fast pin",
                test_ro_fast_pin,
        },
};

static void run_non_anon_test_case(struct non_anon_test_case const *test_case)
{
        int i;

        run_with_zeropage(test_case->fn, test_case->desc);
        run_with_memfd(test_case->fn, test_case->desc);
        run_with_tmpfile(test_case->fn, test_case->desc);
        if (pmdsize)
                run_with_huge_zeropage(test_case->fn, test_case->desc);
        for (i = 0; i < nr_hugetlbsizes; i++)
                run_with_memfd_hugetlb(test_case->fn, test_case->desc,
                                       hugetlbsizes[i]);
}

static void run_non_anon_test_cases(void)
{
        int i;

        ksft_print_msg("[RUN] Non-anonymous memory tests in private mappings\n");

        for (i = 0; i < ARRAY_SIZE(non_anon_test_cases); i++)
                run_non_anon_test_case(&non_anon_test_cases[i]);
}

static int tests_per_non_anon_test_case(void)
{
        int tests = 3 + nr_hugetlbsizes;

        if (pmdsize)
                tests += 1;
        return tests;
}

int main(int argc, char **argv)
{
        struct thp_settings default_settings;

        ksft_print_header();

        pagesize = getpagesize();
        pmdsize = read_pmd_pagesize();
        if (pmdsize) {
                /* Only if THP is supported. */
                thp_read_settings(&default_settings);
                default_settings.hugepages[sz2ord(pmdsize, pagesize)].enabled = THP_INHERIT;
                thp_save_settings();
                thp_push_settings(&default_settings);

                ksft_print_msg("[INFO] detected PMD size: %zu KiB\n",
                               pmdsize / 1024);
                nr_thpsizes = detect_thp_sizes(thpsizes, ARRAY_SIZE(thpsizes));
        }
        nr_hugetlbsizes = detect_hugetlb_page_sizes(hugetlbsizes,
                                                    ARRAY_SIZE(hugetlbsizes));
        has_huge_zeropage = detect_huge_zeropage();

        ksft_set_plan(ARRAY_SIZE(anon_test_cases) * tests_per_anon_test_case() +
                      ARRAY_SIZE(anon_thp_test_cases) * tests_per_anon_thp_test_case() +
                      ARRAY_SIZE(non_anon_test_cases) * tests_per_non_anon_test_case());

        gup_fd = open("/sys/kernel/debug/gup_test", O_RDWR);
        pagemap_fd = open("/proc/self/pagemap", O_RDONLY);
        if (pagemap_fd < 0)
                ksft_exit_fail_msg("opening pagemap failed\n");

        run_anon_test_cases();
        run_anon_thp_test_cases();
        run_non_anon_test_cases();

        if (pmdsize) {
                /* Only if THP is supported. */
                thp_restore_settings();
        }

        ksft_finished();
}