1 /* 2 * Copyright (c) 2016 Fujitsu Ltd. 3 * Author: Naoya Horiguchi <n-horiguchi (at) ah.jp.nec.com> 4 * Ported: Guangwen Feng <fenggw-fnst (at) cn.fujitsu.com> 5 * 6 * This program is free software: you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License as published by 8 * the Free Software Foundation, either version 2 of the License, or 9 * (at your option) any later version. 10 * 11 * This program is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 14 * GNU General Public License for more details. 15 * 16 * You should have received a copy of the GNU General Public License 17 * along with this program, if not, see <http://www.gnu.org/licenses/>. 18 */ 19 20 /* 21 * This is a regression test for the race condition between move_pages() 22 * and freeing hugepages, where move_pages() calls follow_page(FOLL_GET) 23 * for hugepages internally and tries to get its refcount without 24 * preventing concurrent freeing. 25 * 26 * This test can crash the buggy kernel, and the bug was fixed in: 27 * 28 * commit e66f17ff71772b209eed39de35aaa99ba819c93d 29 * Author: Naoya Horiguchi <n-horiguchi (at) ah.jp.nec.com> 30 * Date: Wed Feb 11 15:25:22 2015 -0800 31 * 32 * mm/hugetlb: take page table lock in follow_huge_pmd() 33 */ 34 35 #include <errno.h> 36 #include <unistd.h> 37 #include <string.h> 38 #include <stdio.h> 39 #include <sys/types.h> 40 #include <sys/wait.h> 41 42 #include "tst_test.h" 43 #include "move_pages_support.h" 44 #include "lapi/mmap.h" 45 46 #ifdef HAVE_NUMA_V2 47 48 #define LOOPS 1000 49 #define PATH_MEMINFO "/proc/meminfo" 50 #define PATH_NR_HUGEPAGES "/proc/sys/vm/nr_hugepages" 51 #define PATH_HUGEPAGES "/sys/kernel/mm/hugepages/" 52 #define TEST_PAGES 2 53 #define TEST_NODES 2 54 55 static int pgsz, hpsz; 56 static long orig_hugepages = -1; 57 static char path_hugepages_node1[PATH_MAX]; 58 static char path_hugepages_node2[PATH_MAX]; 59 static long orig_hugepages_node1 = -1; 60 static long orig_hugepages_node2 = -1; 61 static unsigned int node1, node2; 62 static void *addr; 63 64 static void do_child(void) 65 { 66 int test_pages = TEST_PAGES * hpsz / pgsz; 67 int i, j; 68 int *nodes, *status; 69 void **pages; 70 pid_t ppid = getppid(); 71 72 pages = SAFE_MALLOC(sizeof(char *) * test_pages); 73 nodes = SAFE_MALLOC(sizeof(int) * test_pages); 74 status = SAFE_MALLOC(sizeof(int) * test_pages); 75 76 for (i = 0; i < test_pages; i++) 77 pages[i] = addr + i * pgsz; 78 79 for (i = 0; ; i++) { 80 for (j = 0; j < test_pages; j++) { 81 if (i % 2 == 0) 82 nodes[j] = node1; 83 else 84 nodes[j] = node2; 85 status[j] = 0; 86 } 87 88 TEST(numa_move_pages(ppid, test_pages, 89 pages, nodes, status, MPOL_MF_MOVE_ALL)); 90 if (TST_RET < 0) { 91 tst_res(TFAIL | TTERRNO, "move_pages failed"); 92 break; 93 } 94 } 95 96 exit(0); 97 } 98 99 static void do_test(void) 100 { 101 int i; 102 pid_t cpid = -1; 103 int status; 104 unsigned int twenty_percent = (tst_timeout_remaining() / 5); 105 106 addr = SAFE_MMAP(NULL, TEST_PAGES * hpsz, PROT_READ | PROT_WRITE, 107 MAP_PRIVATE | MAP_ANONYMOUS | MAP_HUGETLB, -1, 0); 108 109 SAFE_MUNMAP(addr, TEST_PAGES * hpsz); 110 111 cpid = SAFE_FORK(); 112 if (cpid == 0) 113 do_child(); 114 115 for (i = 0; i < LOOPS; i++) { 116 void *ptr; 117 118 ptr = SAFE_MMAP(NULL, TEST_PAGES * hpsz, 119 PROT_READ | PROT_WRITE, 120 MAP_PRIVATE | MAP_ANONYMOUS | MAP_HUGETLB, -1, 0); 121 if (ptr != addr) 122 tst_brk(TBROK, "Failed to mmap at desired addr"); 123 124 memset(addr, 0, TEST_PAGES * hpsz); 125 126 SAFE_MUNMAP(addr, TEST_PAGES * hpsz); 127 128 if (tst_timeout_remaining() < twenty_percent) 129 break; 130 } 131 132 SAFE_KILL(cpid, SIGKILL); 133 SAFE_WAITPID(cpid, &status, 0); 134 if (!WIFEXITED(status)) 135 tst_res(TPASS, "Bug not reproduced"); 136 } 137 138 static void alloc_free_huge_on_node(unsigned int node, size_t size) 139 { 140 char *mem; 141 long ret; 142 struct bitmask *bm; 143 144 tst_res(TINFO, "Allocating and freeing %zu hugepages on node %u", 145 size / hpsz, node); 146 147 mem = mmap(NULL, size, PROT_READ | PROT_WRITE, 148 MAP_PRIVATE | MAP_ANONYMOUS | MAP_HUGETLB, -1, 0); 149 if (mem == MAP_FAILED) { 150 if (errno == ENOMEM) 151 tst_brk(TCONF, "Cannot allocate huge pages"); 152 153 tst_brk(TBROK | TERRNO, "mmap(..., MAP_HUGETLB, ...) failed"); 154 } 155 156 bm = numa_bitmask_alloc(numa_max_possible_node() + 1); 157 if (!bm) 158 tst_brk(TBROK | TERRNO, "numa_bitmask_alloc() failed"); 159 160 numa_bitmask_setbit(bm, node); 161 162 ret = mbind(mem, size, MPOL_BIND, bm->maskp, bm->size + 1, 0); 163 if (ret) { 164 if (errno == ENOMEM) 165 tst_brk(TCONF, "Cannot mbind huge pages"); 166 167 tst_brk(TBROK | TERRNO, "mbind() failed"); 168 } 169 170 TEST(mlock(mem, size)); 171 if (TST_RET) { 172 SAFE_MUNMAP(mem, size); 173 if (TST_ERR == ENOMEM || TST_ERR == EAGAIN) 174 tst_brk(TCONF, "Cannot lock huge pages"); 175 tst_brk(TBROK | TTERRNO, "mlock failed"); 176 } 177 178 numa_bitmask_free(bm); 179 180 SAFE_MUNMAP(mem, size); 181 } 182 183 static void setup(void) 184 { 185 int ret; 186 long memfree; 187 188 check_config(TEST_NODES); 189 190 if (access(PATH_HUGEPAGES, F_OK)) 191 tst_brk(TCONF, "Huge page not supported"); 192 193 ret = get_allowed_nodes(NH_MEMS, TEST_NODES, &node1, &node2); 194 if (ret < 0) 195 tst_brk(TBROK | TERRNO, "get_allowed_nodes: %d", ret); 196 197 pgsz = (int)get_page_size(); 198 SAFE_FILE_LINES_SCANF(PATH_MEMINFO, "Hugepagesize: %d", &hpsz); 199 200 SAFE_FILE_LINES_SCANF(PATH_MEMINFO, "MemFree: %ld", &memfree); 201 tst_res(TINFO, "Free RAM %ld kB", memfree); 202 203 if (4 * hpsz > memfree) 204 tst_brk(TBROK, "Not enough free RAM"); 205 206 snprintf(path_hugepages_node1, sizeof(path_hugepages_node1), 207 "/sys/devices/system/node/node%u/hugepages/hugepages-%dkB/nr_hugepages", 208 node1, hpsz); 209 210 snprintf(path_hugepages_node2, sizeof(path_hugepages_node2), 211 "/sys/devices/system/node/node%u/hugepages/hugepages-%dkB/nr_hugepages", 212 node2, hpsz); 213 214 if (!access(path_hugepages_node1, F_OK)) { 215 SAFE_FILE_SCANF(path_hugepages_node1, 216 "%ld", &orig_hugepages_node1); 217 tst_res(TINFO, 218 "Increasing %dkB hugepages pool on node %u to %ld", 219 hpsz, node1, orig_hugepages_node1 + 4); 220 SAFE_FILE_PRINTF(path_hugepages_node1, 221 "%ld", orig_hugepages_node1 + 4); 222 } 223 224 if (!access(path_hugepages_node2, F_OK)) { 225 SAFE_FILE_SCANF(path_hugepages_node2, 226 "%ld", &orig_hugepages_node2); 227 tst_res(TINFO, 228 "Increasing %dkB hugepages pool on node %u to %ld", 229 hpsz, node2, orig_hugepages_node2 + 4); 230 SAFE_FILE_PRINTF(path_hugepages_node2, 231 "%ld", orig_hugepages_node2 + 4); 232 } 233 234 hpsz *= 1024; 235 236 if (orig_hugepages_node1 == -1 || orig_hugepages_node2 == -1) { 237 SAFE_FILE_SCANF(PATH_NR_HUGEPAGES, "%ld", &orig_hugepages); 238 tst_res(TINFO, "Increasing global hugepages pool to %ld", 239 orig_hugepages + 8); 240 SAFE_FILE_PRINTF(PATH_NR_HUGEPAGES, "%ld", orig_hugepages + 8); 241 } 242 243 alloc_free_huge_on_node(node1, 4L * hpsz); 244 alloc_free_huge_on_node(node2, 4L * hpsz); 245 } 246 247 static void cleanup(void) 248 { 249 if (orig_hugepages != -1) 250 SAFE_FILE_PRINTF(PATH_NR_HUGEPAGES, "%ld", orig_hugepages); 251 252 if (orig_hugepages_node1 != -1) { 253 SAFE_FILE_PRINTF(path_hugepages_node1, 254 "%ld", orig_hugepages_node1); 255 } 256 257 if (orig_hugepages_node2 != -1) { 258 SAFE_FILE_PRINTF(path_hugepages_node2, 259 "%ld", orig_hugepages_node2); 260 } 261 } 262 263 static struct tst_test test = { 264 .min_kver = "2.6.32", 265 .needs_root = 1, 266 .forks_child = 1, 267 .setup = setup, 268 .cleanup = cleanup, 269 .test_all = do_test, 270 }; 271 272 #else 273 TST_TEST_TCONF(NUMA_ERROR_MSG); 274 #endif 275