1#!/bin/bash 2# SPDX-License-Identifier: GPL-2.0 3# Please run as root 4 5# Kselftest framework requirement - SKIP code is 4. 6ksft_skip=4 7 8count_total=0 9count_pass=0 10count_fail=0 11count_skip=0 12exitcode=0 13 14usage() { 15 cat <<EOF 16usage: ${BASH_SOURCE[0]:-$0} [ options ] 17 18 -a: run all tests, including extra ones (other than destructive ones) 19 -t: specify specific categories to tests to run 20 -h: display this message 21 -n: disable TAP output 22 -d: run destructive tests 23 24The default behavior is to run required tests only. If -a is specified, 25will run all tests. 26 27Alternatively, specific groups tests can be run by passing a string 28to the -t argument containing one or more of the following categories 29separated by spaces: 30- mmap 31 tests for mmap(2) 32- gup_test 33 tests for gup 34- userfaultfd 35 tests for userfaultfd(2) 36- compaction 37 a test for the patch "Allow compaction of unevictable pages" 38- mlock 39 tests for mlock(2) 40- mremap 41 tests for mremap(2) 42- hugevm 43 tests for very large virtual address space 44- vmalloc 45 vmalloc smoke tests 46- hmm 47 hmm smoke tests 48- madv_populate 49 test memadvise(2) MADV_POPULATE_{READ,WRITE} options 50- memfd_secret 51 test memfd_secret(2) 52- process_mrelease 53 test process_mrelease(2) 54- ksm 55 ksm tests that do not require >=2 NUMA nodes 56- ksm_numa 57 ksm tests that require >=2 NUMA nodes 58- pkey 59 memory protection key tests 60- soft_dirty 61 test soft dirty page bit semantics 62- pagemap 63 test pagemap_scan IOCTL 64- cow 65 test copy-on-write semantics 66- thp 67 test transparent huge pages 68- hugetlb 69 test hugetlbfs huge pages 70- migration 71 invoke move_pages(2) to exercise the migration entry code 72 paths in the kernel 73- mkdirty 74 test handling of code that might set PTE/PMD dirty in 75 read-only VMAs 76- mdwe 77 test prctl(PR_SET_MDWE, ...) 78 79example: ./run_vmtests.sh -t "hmm mmap ksm" 80EOF 81 exit 0 82} 83 84RUN_ALL=false 85RUN_DESTRUCTIVE=false 86TAP_PREFIX="# " 87 88while getopts "aht:n" OPT; do 89 case ${OPT} in 90 "a") RUN_ALL=true ;; 91 "h") usage ;; 92 "t") VM_SELFTEST_ITEMS=${OPTARG} ;; 93 "n") TAP_PREFIX= ;; 94 "d") RUN_DESTRUCTIVE=true ;; 95 esac 96done 97shift $((OPTIND -1)) 98 99# default behavior: run all tests 100VM_SELFTEST_ITEMS=${VM_SELFTEST_ITEMS:-default} 101 102test_selected() { 103 if [ "$VM_SELFTEST_ITEMS" == "default" ]; then 104 # If no VM_SELFTEST_ITEMS are specified, run all tests 105 return 0 106 fi 107 # If test selected argument is one of the test items 108 if [[ " ${VM_SELFTEST_ITEMS[*]} " =~ " ${1} " ]]; then 109 return 0 110 else 111 return 1 112 fi 113} 114 115run_gup_matrix() { 116 # -t: thp=on, -T: thp=off, -H: hugetlb=on 117 local hugetlb_mb=$(( needmem_KB / 1024 )) 118 119 for huge in -t -T "-H -m $hugetlb_mb"; do 120 # -u: gup-fast, -U: gup-basic, -a: pin-fast, -b: pin-basic, -L: pin-longterm 121 for test_cmd in -u -U -a -b -L; do 122 # -w: write=1, -W: write=0 123 for write in -w -W; do 124 # -S: shared 125 for share in -S " "; do 126 # -n: How many pages to fetch together? 512 is special 127 # because it's default thp size (or 2M on x86), 123 to 128 # just test partial gup when hit a huge in whatever form 129 for num in "-n 1" "-n 512" "-n 123"; do 130 CATEGORY="gup_test" run_test ./gup_test \ 131 $huge $test_cmd $write $share $num 132 done 133 done 134 done 135 done 136 done 137} 138 139# get huge pagesize and freepages from /proc/meminfo 140while read -r name size unit; do 141 if [ "$name" = "HugePages_Free:" ]; then 142 freepgs="$size" 143 fi 144 if [ "$name" = "Hugepagesize:" ]; then 145 hpgsize_KB="$size" 146 fi 147done < /proc/meminfo 148 149# Simple hugetlbfs tests have a hardcoded minimum requirement of 150# huge pages totaling 256MB (262144KB) in size. The userfaultfd 151# hugetlb test requires a minimum of 2 * nr_cpus huge pages. Take 152# both of these requirements into account and attempt to increase 153# number of huge pages available. 154nr_cpus=$(nproc) 155uffd_min_KB=$((hpgsize_KB * nr_cpus * 2)) 156hugetlb_min_KB=$((256 * 1024)) 157if [[ $uffd_min_KB -gt $hugetlb_min_KB ]]; then 158 needmem_KB=$uffd_min_KB 159else 160 needmem_KB=$hugetlb_min_KB 161fi 162 163# set proper nr_hugepages 164if [ -n "$freepgs" ] && [ -n "$hpgsize_KB" ]; then 165 nr_hugepgs=$(cat /proc/sys/vm/nr_hugepages) 166 needpgs=$((needmem_KB / hpgsize_KB)) 167 tries=2 168 while [ "$tries" -gt 0 ] && [ "$freepgs" -lt "$needpgs" ]; do 169 lackpgs=$((needpgs - freepgs)) 170 echo 3 > /proc/sys/vm/drop_caches 171 if ! echo $((lackpgs + nr_hugepgs)) > /proc/sys/vm/nr_hugepages; then 172 echo "Please run this test as root" 173 exit $ksft_skip 174 fi 175 while read -r name size unit; do 176 if [ "$name" = "HugePages_Free:" ]; then 177 freepgs=$size 178 fi 179 done < /proc/meminfo 180 tries=$((tries - 1)) 181 done 182 if [ "$freepgs" -lt "$needpgs" ]; then 183 printf "Not enough huge pages available (%d < %d)\n" \ 184 "$freepgs" "$needpgs" 185 fi 186else 187 echo "no hugetlbfs support in kernel?" 188 exit 1 189fi 190 191# filter 64bit architectures 192ARCH64STR="arm64 ia64 mips64 parisc64 ppc64 ppc64le riscv64 s390x sparc64 x86_64" 193if [ -z "$ARCH" ]; then 194 ARCH=$(uname -m 2>/dev/null | sed -e 's/aarch64.*/arm64/') 195fi 196VADDR64=0 197echo "$ARCH64STR" | grep "$ARCH" &>/dev/null && VADDR64=1 198 199tap_prefix() { 200 sed -e "s/^/${TAP_PREFIX}/" 201} 202 203tap_output() { 204 if [[ ! -z "$TAP_PREFIX" ]]; then 205 read str 206 echo $str 207 fi 208} 209 210pretty_name() { 211 echo "$*" | sed -e 's/^\(bash \)\?\.\///' 212} 213 214# Usage: run_test [test binary] [arbitrary test arguments...] 215run_test() { 216 if test_selected ${CATEGORY}; then 217 # On memory constrainted systems some tests can fail to allocate hugepages. 218 # perform some cleanup before the test for a higher success rate. 219 if [ ${CATEGORY} == "thp" ] | [ ${CATEGORY} == "hugetlb" ]; then 220 echo 3 > /proc/sys/vm/drop_caches 221 sleep 2 222 echo 1 > /proc/sys/vm/compact_memory 223 sleep 2 224 fi 225 226 local test=$(pretty_name "$*") 227 local title="running $*" 228 local sep=$(echo -n "$title" | tr "[:graph:][:space:]" -) 229 printf "%s\n%s\n%s\n" "$sep" "$title" "$sep" | tap_prefix 230 231 ("$@" 2>&1) | tap_prefix 232 local ret=${PIPESTATUS[0]} 233 count_total=$(( count_total + 1 )) 234 if [ $ret -eq 0 ]; then 235 count_pass=$(( count_pass + 1 )) 236 echo "[PASS]" | tap_prefix 237 echo "ok ${count_total} ${test}" | tap_output 238 elif [ $ret -eq $ksft_skip ]; then 239 count_skip=$(( count_skip + 1 )) 240 echo "[SKIP]" | tap_prefix 241 echo "ok ${count_total} ${test} # SKIP" | tap_output 242 exitcode=$ksft_skip 243 else 244 count_fail=$(( count_fail + 1 )) 245 echo "[FAIL]" | tap_prefix 246 echo "not ok ${count_total} ${test} # exit=$ret" | tap_output 247 exitcode=1 248 fi 249 fi # test_selected 250} 251 252echo "TAP version 13" | tap_output 253 254CATEGORY="hugetlb" run_test ./hugepage-mmap 255 256shmmax=$(cat /proc/sys/kernel/shmmax) 257shmall=$(cat /proc/sys/kernel/shmall) 258echo 268435456 > /proc/sys/kernel/shmmax 259echo 4194304 > /proc/sys/kernel/shmall 260CATEGORY="hugetlb" run_test ./hugepage-shm 261echo "$shmmax" > /proc/sys/kernel/shmmax 262echo "$shmall" > /proc/sys/kernel/shmall 263 264CATEGORY="hugetlb" run_test ./map_hugetlb 265CATEGORY="hugetlb" run_test ./hugepage-mremap 266CATEGORY="hugetlb" run_test ./hugepage-vmemmap 267CATEGORY="hugetlb" run_test ./hugetlb-madvise 268CATEGORY="hugetlb" run_test ./hugetlb_dio 269 270nr_hugepages_tmp=$(cat /proc/sys/vm/nr_hugepages) 271# For this test, we need one and just one huge page 272echo 1 > /proc/sys/vm/nr_hugepages 273CATEGORY="hugetlb" run_test ./hugetlb_fault_after_madv 274CATEGORY="hugetlb" run_test ./hugetlb_madv_vs_map 275# Restore the previous number of huge pages, since further tests rely on it 276echo "$nr_hugepages_tmp" > /proc/sys/vm/nr_hugepages 277 278if test_selected "hugetlb"; then 279 echo "NOTE: These hugetlb tests provide minimal coverage. Use" | tap_prefix 280 echo " https://github.com/libhugetlbfs/libhugetlbfs.git for" | tap_prefix 281 echo " hugetlb regression testing." | tap_prefix 282fi 283 284CATEGORY="mmap" run_test ./map_fixed_noreplace 285 286if $RUN_ALL; then 287 run_gup_matrix 288else 289 # get_user_pages_fast() benchmark 290 CATEGORY="gup_test" run_test ./gup_test -u 291 # pin_user_pages_fast() benchmark 292 CATEGORY="gup_test" run_test ./gup_test -a 293fi 294# Dump pages 0, 19, and 4096, using pin_user_pages: 295CATEGORY="gup_test" run_test ./gup_test -ct -F 0x1 0 19 0x1000 296CATEGORY="gup_test" run_test ./gup_longterm 297 298CATEGORY="userfaultfd" run_test ./uffd-unit-tests 299uffd_stress_bin=./uffd-stress 300CATEGORY="userfaultfd" run_test ${uffd_stress_bin} anon 20 16 301# Hugetlb tests require source and destination huge pages. Pass in half 302# the size of the free pages we have, which is used for *each*. 303half_ufd_size_MB=$((freepgs / 2)) 304CATEGORY="userfaultfd" run_test ${uffd_stress_bin} hugetlb "$half_ufd_size_MB" 32 305CATEGORY="userfaultfd" run_test ${uffd_stress_bin} hugetlb-private "$half_ufd_size_MB" 32 306CATEGORY="userfaultfd" run_test ${uffd_stress_bin} shmem 20 16 307CATEGORY="userfaultfd" run_test ${uffd_stress_bin} shmem-private 20 16 308 309#cleanup 310echo "$nr_hugepgs" > /proc/sys/vm/nr_hugepages 311 312CATEGORY="compaction" run_test ./compaction_test 313 314if command -v sudo &> /dev/null; 315then 316 CATEGORY="mlock" run_test sudo -u nobody ./on-fault-limit 317else 318 echo "# SKIP ./on-fault-limit" 319fi 320 321CATEGORY="mmap" run_test ./map_populate 322 323CATEGORY="mlock" run_test ./mlock-random-test 324 325CATEGORY="mlock" run_test ./mlock2-tests 326 327CATEGORY="process_mrelease" run_test ./mrelease_test 328 329CATEGORY="mremap" run_test ./mremap_test 330 331CATEGORY="hugetlb" run_test ./thuge-gen 332CATEGORY="hugetlb" run_test ./charge_reserved_hugetlb.sh -cgroup-v2 333CATEGORY="hugetlb" run_test ./hugetlb_reparenting_test.sh -cgroup-v2 334if $RUN_DESTRUCTIVE; then 335nr_hugepages_tmp=$(cat /proc/sys/vm/nr_hugepages) 336enable_soft_offline=$(cat /proc/sys/vm/enable_soft_offline) 337echo 8 > /proc/sys/vm/nr_hugepages 338CATEGORY="hugetlb" run_test ./hugetlb-soft-offline 339echo "$nr_hugepages_tmp" > /proc/sys/vm/nr_hugepages 340echo "$enable_soft_offline" > /proc/sys/vm/enable_soft_offline 341CATEGORY="hugetlb" run_test ./hugetlb-read-hwpoison 342fi 343 344if [ $VADDR64 -ne 0 ]; then 345 346 # set overcommit_policy as OVERCOMMIT_ALWAYS so that kernel 347 # allows high virtual address allocation requests independent 348 # of platform's physical memory. 349 350 prev_policy=$(cat /proc/sys/vm/overcommit_memory) 351 echo 1 > /proc/sys/vm/overcommit_memory 352 CATEGORY="hugevm" run_test ./virtual_address_range 353 echo $prev_policy > /proc/sys/vm/overcommit_memory 354 355 # va high address boundary switch test 356 ARCH_ARM64="arm64" 357 prev_nr_hugepages=$(cat /proc/sys/vm/nr_hugepages) 358 if [ "$ARCH" == "$ARCH_ARM64" ]; then 359 echo 6 > /proc/sys/vm/nr_hugepages 360 fi 361 CATEGORY="hugevm" run_test bash ./va_high_addr_switch.sh 362 if [ "$ARCH" == "$ARCH_ARM64" ]; then 363 echo $prev_nr_hugepages > /proc/sys/vm/nr_hugepages 364 fi 365fi # VADDR64 366 367# vmalloc stability smoke test 368CATEGORY="vmalloc" run_test bash ./test_vmalloc.sh smoke 369 370CATEGORY="mremap" run_test ./mremap_dontunmap 371 372CATEGORY="hmm" run_test bash ./test_hmm.sh smoke 373 374# MADV_POPULATE_READ and MADV_POPULATE_WRITE tests 375CATEGORY="madv_populate" run_test ./madv_populate 376 377(echo 0 | sudo tee /proc/sys/kernel/yama/ptrace_scope 2>&1) | tap_prefix 378CATEGORY="memfd_secret" run_test ./memfd_secret 379 380# KSM KSM_MERGE_TIME_HUGE_PAGES test with size of 100 381CATEGORY="ksm" run_test ./ksm_tests -H -s 100 382# KSM KSM_MERGE_TIME test with size of 100 383CATEGORY="ksm" run_test ./ksm_tests -P -s 100 384# KSM MADV_MERGEABLE test with 10 identical pages 385CATEGORY="ksm" run_test ./ksm_tests -M -p 10 386# KSM unmerge test 387CATEGORY="ksm" run_test ./ksm_tests -U 388# KSM test with 10 zero pages and use_zero_pages = 0 389CATEGORY="ksm" run_test ./ksm_tests -Z -p 10 -z 0 390# KSM test with 10 zero pages and use_zero_pages = 1 391CATEGORY="ksm" run_test ./ksm_tests -Z -p 10 -z 1 392# KSM test with 2 NUMA nodes and merge_across_nodes = 1 393CATEGORY="ksm_numa" run_test ./ksm_tests -N -m 1 394# KSM test with 2 NUMA nodes and merge_across_nodes = 0 395CATEGORY="ksm_numa" run_test ./ksm_tests -N -m 0 396 397CATEGORY="ksm" run_test ./ksm_functional_tests 398 399# protection_keys tests 400nr_hugepgs=$(cat /proc/sys/vm/nr_hugepages) 401if [ -x ./protection_keys_32 ] 402then 403 CATEGORY="pkey" run_test ./protection_keys_32 404fi 405 406if [ -x ./protection_keys_64 ] 407then 408 CATEGORY="pkey" run_test ./protection_keys_64 409fi 410echo "$nr_hugepgs" > /proc/sys/vm/nr_hugepages 411 412if [ -x ./soft-dirty ] 413then 414 CATEGORY="soft_dirty" run_test ./soft-dirty 415fi 416 417CATEGORY="pagemap" run_test ./pagemap_ioctl 418 419# COW tests 420CATEGORY="cow" run_test ./cow 421 422CATEGORY="thp" run_test ./khugepaged 423 424CATEGORY="thp" run_test ./khugepaged -s 2 425 426CATEGORY="thp" run_test ./transhuge-stress -d 20 427 428# Try to create XFS if not provided 429if [ -z "${SPLIT_HUGE_PAGE_TEST_XFS_PATH}" ]; then 430 if test_selected "thp"; then 431 if grep xfs /proc/filesystems &>/dev/null; then 432 XFS_IMG=$(mktemp /tmp/xfs_img_XXXXXX) 433 SPLIT_HUGE_PAGE_TEST_XFS_PATH=$(mktemp -d /tmp/xfs_dir_XXXXXX) 434 truncate -s 314572800 ${XFS_IMG} 435 mkfs.xfs -q ${XFS_IMG} 436 mount -o loop ${XFS_IMG} ${SPLIT_HUGE_PAGE_TEST_XFS_PATH} 437 MOUNTED_XFS=1 438 fi 439 fi 440fi 441 442CATEGORY="thp" run_test ./split_huge_page_test ${SPLIT_HUGE_PAGE_TEST_XFS_PATH} 443 444if [ -n "${MOUNTED_XFS}" ]; then 445 umount ${SPLIT_HUGE_PAGE_TEST_XFS_PATH} 446 rmdir ${SPLIT_HUGE_PAGE_TEST_XFS_PATH} 447 rm -f ${XFS_IMG} 448fi 449 450CATEGORY="migration" run_test ./migration 451 452CATEGORY="mkdirty" run_test ./mkdirty 453 454CATEGORY="mdwe" run_test ./mdwe_test 455 456echo "SUMMARY: PASS=${count_pass} SKIP=${count_skip} FAIL=${count_fail}" | tap_prefix 457echo "1..${count_total}" | tap_output 458 459exit $exitcode 460