Lines Matching refs:user_addr

51 	unsigned long user_addr;  in do_usercopy_stack()  local
69 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, in do_usercopy_stack()
72 if (user_addr >= TASK_SIZE) { in do_usercopy_stack()
79 if (copy_to_user((void __user *)user_addr, good_stack, in do_usercopy_stack()
86 if (copy_to_user((void __user *)user_addr, bad_stack, in do_usercopy_stack()
100 if (copy_from_user(good_stack, (void __user *)user_addr, in do_usercopy_stack()
107 if (copy_from_user(bad_stack, (void __user *)user_addr, in do_usercopy_stack()
115 vm_munmap(user_addr, PAGE_SIZE); in do_usercopy_stack()
124 unsigned long user_addr; in do_usercopy_heap_size() local
137 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, in do_usercopy_heap_size()
140 if (user_addr >= TASK_SIZE) { in do_usercopy_heap_size()
148 test_user_addr = (void __user *)(user_addr + 16); in do_usercopy_heap_size()
178 vm_munmap(user_addr, PAGE_SIZE); in do_usercopy_heap_size()
192 unsigned char __user *user_addr; in do_usercopy_heap_whitelist() local
218 user_addr = (void __user *)user_alloc; in do_usercopy_heap_whitelist()
228 if (copy_to_user(user_addr, buf + offset, size)) { in do_usercopy_heap_whitelist()
234 if (copy_to_user(user_addr, buf + offset - 1, size)) { in do_usercopy_heap_whitelist()
240 if (copy_from_user(buf + offset, user_addr, size)) { in do_usercopy_heap_whitelist()
246 if (copy_from_user(buf + offset - 1, user_addr, size)) { in do_usercopy_heap_whitelist()
297 unsigned long user_addr; in lkdtm_USERCOPY_KERNEL() local
299 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, in lkdtm_USERCOPY_KERNEL()
302 if (user_addr >= TASK_SIZE) { in lkdtm_USERCOPY_KERNEL()
308 if (copy_to_user((void __user *)user_addr, test_text, in lkdtm_USERCOPY_KERNEL()
315 if (copy_to_user((void __user *)user_addr, vm_mmap, in lkdtm_USERCOPY_KERNEL()
322 vm_munmap(user_addr, PAGE_SIZE); in lkdtm_USERCOPY_KERNEL()