Lines Matching full:map

166 static int unlock_lock_check(char *map)  in unlock_lock_check()  argument
168 if (is_vmflag_set((unsigned long)map, LOCKED)) { in unlock_lock_check()
178 char *map; in test_mlock_lock() local
181 map = mmap(NULL, 2 * page_size, PROT_READ | PROT_WRITE, in test_mlock_lock()
183 if (map == MAP_FAILED) in test_mlock_lock()
186 if (mlock2_(map, 2 * page_size, 0)) { in test_mlock_lock()
187 munmap(map, 2 * page_size); in test_mlock_lock()
191 ksft_test_result(lock_check((unsigned long)map), "%s: Locked\n", __func__); in test_mlock_lock()
194 if (munlock(map, 2 * page_size)) { in test_mlock_lock()
195 munmap(map, 2 * page_size); in test_mlock_lock()
199 ksft_test_result(!unlock_lock_check(map), "%s: Locked\n", __func__); in test_mlock_lock()
200 munmap(map, 2 * page_size); in test_mlock_lock()
203 static int onfault_check(char *map) in onfault_check() argument
205 *map = 'a'; in onfault_check()
206 if (!is_vma_lock_on_fault((unsigned long)map)) { in onfault_check()
214 static int unlock_onfault_check(char *map) in unlock_onfault_check() argument
218 if (is_vma_lock_on_fault((unsigned long)map) || in unlock_onfault_check()
219 is_vma_lock_on_fault((unsigned long)map + page_size)) { in unlock_onfault_check()
229 char *map; in test_mlock_onfault() local
232 map = mmap(NULL, 2 * page_size, PROT_READ | PROT_WRITE, in test_mlock_onfault()
234 if (map == MAP_FAILED) in test_mlock_onfault()
237 if (mlock2_(map, 2 * page_size, MLOCK_ONFAULT)) { in test_mlock_onfault()
238 munmap(map, 2 * page_size); in test_mlock_onfault()
242 ksft_test_result(!onfault_check(map), "%s: VMA marked for lock on fault\n", __func__); in test_mlock_onfault()
245 if (munlock(map, 2 * page_size)) { in test_mlock_onfault()
246 munmap(map, 2 * page_size); in test_mlock_onfault()
250 ksft_test_result(!unlock_onfault_check(map), "VMA open lock after fault\n"); in test_mlock_onfault()
251 munmap(map, 2 * page_size); in test_mlock_onfault()
256 char *map; in test_lock_onfault_of_present() local
259 map = mmap(NULL, 2 * page_size, PROT_READ | PROT_WRITE, in test_lock_onfault_of_present()
261 if (map == MAP_FAILED) in test_lock_onfault_of_present()
264 *map = 'a'; in test_lock_onfault_of_present()
266 if (mlock2_(map, 2 * page_size, MLOCK_ONFAULT)) { in test_lock_onfault_of_present()
267 munmap(map, 2 * page_size); in test_lock_onfault_of_present()
271 ksft_test_result(is_vma_lock_on_fault((unsigned long)map) || in test_lock_onfault_of_present()
272 is_vma_lock_on_fault((unsigned long)map + page_size), in test_lock_onfault_of_present()
274 munmap(map, 2 * page_size); in test_lock_onfault_of_present()
279 char *map; in test_munlockall0() local
282 map = mmap(NULL, 2 * page_size, PROT_READ | PROT_WRITE, in test_munlockall0()
284 if (map == MAP_FAILED) in test_munlockall0()
288 munmap(map, 2 * page_size); in test_munlockall0()
292 ksft_test_result(lock_check((unsigned long)map), "%s: Locked memory area\n", __func__); in test_munlockall0()
295 munmap(map, 2 * page_size); in test_munlockall0()
299 ksft_test_result(!unlock_lock_check(map), "%s: No locked memory\n", __func__); in test_munlockall0()
300 munmap(map, 2 * page_size); in test_munlockall0()
305 char *map; in test_munlockall1() local
308 map = mmap(NULL, 2 * page_size, PROT_READ | PROT_WRITE, in test_munlockall1()
310 if (map == MAP_FAILED) in test_munlockall1()
314 munmap(map, 2 * page_size); in test_munlockall1()
318 ksft_test_result(!onfault_check(map), "%s: VMA marked for lock on fault\n", __func__); in test_munlockall1()
321 munmap(map, 2 * page_size); in test_munlockall1()
325 ksft_test_result(!unlock_onfault_check(map), "%s: Unlocked\n", __func__); in test_munlockall1()
328 munmap(map, 2 * page_size); in test_munlockall1()
332 ksft_test_result(lock_check((unsigned long)map), "%s: Locked\n", __func__); in test_munlockall1()
335 munmap(map, 2 * page_size); in test_munlockall1()
339 ksft_test_result(!unlock_lock_check(map), "%s: No locked memory\n", __func__); in test_munlockall1()
340 munmap(map, 2 * page_size); in test_munlockall1()
345 void *map; in test_vma_management() local
351 map = mmap(NULL, 3 * page_size, PROT_READ | PROT_WRITE, in test_vma_management()
353 if (map == MAP_FAILED) in test_vma_management()
356 if (call_mlock && mlock2_(map, 3 * page_size, MLOCK_ONFAULT)) { in test_vma_management()
357 munmap(map, 3 * page_size); in test_vma_management()
361 if (get_vm_area((unsigned long)map, &page1) || in test_vma_management()
362 get_vm_area((unsigned long)map + page_size, &page2) || in test_vma_management()
363 get_vm_area((unsigned long)map + page_size * 2, &page3)) { in test_vma_management()
364 munmap(map, 3 * page_size); in test_vma_management()
374 munmap(map, 3 * page_size); in test_vma_management()
378 if (munlock(map + page_size, page_size)) { in test_vma_management()
379 munmap(map, 3 * page_size); in test_vma_management()
383 if (get_vm_area((unsigned long)map, &page1) || in test_vma_management()
384 get_vm_area((unsigned long)map + page_size, &page2) || in test_vma_management()
385 get_vm_area((unsigned long)map + page_size * 2, &page3)) { in test_vma_management()
386 munmap(map, 3 * page_size); in test_vma_management()
392 munmap(map, 3 * page_size); in test_vma_management()
397 if (munlock(map, page_size * 3)) { in test_vma_management()
398 munmap(map, 3 * page_size); in test_vma_management()
402 if (get_vm_area((unsigned long)map, &page1) || in test_vma_management()
403 get_vm_area((unsigned long)map + page_size, &page2) || in test_vma_management()
404 get_vm_area((unsigned long)map + page_size * 2, &page3)) { in test_vma_management()
405 munmap(map, 3 * page_size); in test_vma_management()
411 munmap(map, 3 * page_size); in test_vma_management()
416 munmap(map, 3 * page_size); in test_vma_management()
431 void *map; in main() local
435 map = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_PRIVATE, -1, 0); in main()
436 if (map == MAP_FAILED) in main()
439 ret = mlock2_(map, size, MLOCK_ONFAULT); in main()
443 munmap(map, size); in main()