Lines Matching refs:user_entry
2628 vm_named_entry_t user_entry; in mach_make_memory_entry_internal() local
2650 user_entry = NULL; in mach_make_memory_entry_internal()
2656 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_VA… in mach_make_memory_entry_internal()
2663 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_AR… in mach_make_memory_entry_internal()
2675 user_entry = NULL; in mach_make_memory_entry_internal()
2686 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_AR… in mach_make_memory_entry_internal()
2692 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_AR… in mach_make_memory_entry_internal()
2703 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_RI… in mach_make_memory_entry_internal()
2722 …prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_SUCCESS); in mach_make_memory_entry_internal()
2732 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_AR… in mach_make_memory_entry_internal()
2742 kr = mach_memory_entry_allocate(&user_entry, &user_handle); in mach_make_memory_entry_internal()
2744 …prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_FAILURE); in mach_make_memory_entry_internal()
2871 user_entry, in mach_make_memory_entry_internal()
2880 user_entry->internal = TRUE; in mach_make_memory_entry_internal()
2881 user_entry->is_sub_map = FALSE; in mach_make_memory_entry_internal()
2882 user_entry->offset = 0; in mach_make_memory_entry_internal()
2883 user_entry->data_offset = 0; in mach_make_memory_entry_internal()
2884 user_entry->protection = protections; in mach_make_memory_entry_internal()
2885 SET_MAP_MEM(access, user_entry->protection); in mach_make_memory_entry_internal()
2886 user_entry->size = map_size; in mach_make_memory_entry_internal()
2891 *size = CAST_DOWN(vm_size_t, (user_entry->size - in mach_make_memory_entry_internal()
2892 user_entry->data_offset)); in mach_make_memory_entry_internal()
2894 …prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_SUCCESS); in mach_make_memory_entry_internal()
2902 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_TA… in mach_make_memory_entry_internal()
2909 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_AR… in mach_make_memory_entry_internal()
2928 …ze 0x%llx prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, kr); in mach_make_memory_entry_internal()
2933 kr = mach_memory_entry_allocate(&user_entry, &user_handle); in mach_make_memory_entry_internal()
2936 …prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_FAILURE); in mach_make_memory_entry_internal()
2940 user_entry->backing.copy = copy; in mach_make_memory_entry_internal()
2941 user_entry->internal = FALSE; in mach_make_memory_entry_internal()
2942 user_entry->is_sub_map = FALSE; in mach_make_memory_entry_internal()
2943 user_entry->is_copy = TRUE; in mach_make_memory_entry_internal()
2944 user_entry->offset = 0; in mach_make_memory_entry_internal()
2945 user_entry->protection = protections; in mach_make_memory_entry_internal()
2946 user_entry->size = map_size; in mach_make_memory_entry_internal()
2947 user_entry->data_offset = offset_in_page; in mach_make_memory_entry_internal()
2949 *size = CAST_DOWN(vm_size_t, (user_entry->size - in mach_make_memory_entry_internal()
2950 user_entry->data_offset)); in mach_make_memory_entry_internal()
2952 …prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_SUCCESS); in mach_make_memory_entry_internal()
2965 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_TA… in mach_make_memory_entry_internal()
3024 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_INVALID_AR… in mach_make_memory_entry_internal()
3075 …ze 0x%llx prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, kr); in mach_make_memory_entry_internal()
3091 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_PROTECTION… in mach_make_memory_entry_internal()
3111 …ot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_PROTECTION… in mach_make_memory_entry_internal()
3142 …prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_SUCCESS); in mach_make_memory_entry_internal()
3159 kr = mach_memory_entry_allocate(&user_entry, &user_handle); in mach_make_memory_entry_internal()
3165 …prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_FAILURE); in mach_make_memory_entry_internal()
3169 user_entry->backing.copy = copy; in mach_make_memory_entry_internal()
3170 user_entry->is_sub_map = FALSE; in mach_make_memory_entry_internal()
3171 user_entry->is_object = FALSE; in mach_make_memory_entry_internal()
3172 user_entry->internal = FALSE; in mach_make_memory_entry_internal()
3173 user_entry->protection = protections; in mach_make_memory_entry_internal()
3174 user_entry->size = map_size; in mach_make_memory_entry_internal()
3175 user_entry->data_offset = offset_in_page; in mach_make_memory_entry_internal()
3178 user_entry->is_copy = TRUE; in mach_make_memory_entry_internal()
3179 user_entry->offset = 0; in mach_make_memory_entry_internal()
3181 user_entry->is_object = TRUE; in mach_make_memory_entry_internal()
3182 user_entry->internal = object->internal; in mach_make_memory_entry_internal()
3183 user_entry->offset = VME_OFFSET(vm_map_copy_first_entry(copy)); in mach_make_memory_entry_internal()
3184 SET_MAP_MEM(GET_MAP_MEM(permission), user_entry->protection); in mach_make_memory_entry_internal()
3187 *size = CAST_DOWN(vm_size_t, (user_entry->size - in mach_make_memory_entry_internal()
3188 user_entry->data_offset)); in mach_make_memory_entry_internal()
3190 …prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_SUCCESS); in mach_make_memory_entry_internal()
3254 if (mach_memory_entry_allocate(&user_entry, &user_handle) in mach_make_memory_entry_internal()
3260 user_entry->size = map_size; in mach_make_memory_entry_internal()
3261 user_entry->offset = parent_entry->offset + map_start; in mach_make_memory_entry_internal()
3262 user_entry->data_offset = offset_in_page; in mach_make_memory_entry_internal()
3263 user_entry->is_sub_map = parent_entry->is_sub_map; in mach_make_memory_entry_internal()
3264 user_entry->is_copy = parent_entry->is_copy; in mach_make_memory_entry_internal()
3265 user_entry->internal = parent_entry->internal; in mach_make_memory_entry_internal()
3266 user_entry->protection = protections; in mach_make_memory_entry_internal()
3269 SET_MAP_MEM(access, user_entry->protection); in mach_make_memory_entry_internal()
3275 user_entry->backing.map = map; in mach_make_memory_entry_internal()
3281 user_entry, in mach_make_memory_entry_internal()
3283 user_entry->offset, in mach_make_memory_entry_internal()
3284 user_entry->size, in mach_make_memory_entry_internal()
3285 (user_entry->protection & VM_PROT_ALL)); in mach_make_memory_entry_internal()
3289 assert(user_entry->is_object); in mach_make_memory_entry_internal()
3315 *size = CAST_DOWN(vm_size_t, (user_entry->size - in mach_make_memory_entry_internal()
3316 user_entry->data_offset)); in mach_make_memory_entry_internal()
3318 …prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, KERN_SUCCESS); in mach_make_memory_entry_internal()
3330 …ze 0x%llx prot 0x%x -> entry %p kr 0x%x\n", target_map, offset, *size, permission, user_entry, kr); in mach_make_memory_entry_internal()
3425 vm_named_entry_t user_entry; in mach_memory_entry_allocate() local
3428 user_entry = kalloc_type(struct vm_named_entry, in mach_memory_entry_allocate()
3431 named_entry_lock_init(user_entry); in mach_memory_entry_allocate()
3433 user_entry->backing.copy = NULL; in mach_memory_entry_allocate()
3434 user_entry->is_object = FALSE; in mach_memory_entry_allocate()
3435 user_entry->is_sub_map = FALSE; in mach_memory_entry_allocate()
3436 user_entry->is_copy = FALSE; in mach_memory_entry_allocate()
3437 user_entry->internal = FALSE; in mach_memory_entry_allocate()
3438 user_entry->size = 0; in mach_memory_entry_allocate()
3439 user_entry->offset = 0; in mach_memory_entry_allocate()
3440 user_entry->data_offset = 0; in mach_memory_entry_allocate()
3441 user_entry->protection = VM_PROT_NONE; in mach_memory_entry_allocate()
3442 user_entry->ref_count = 1; in mach_memory_entry_allocate()
3444 user_handle = ipc_kobject_alloc_port((ipc_kobject_t)user_entry, in mach_memory_entry_allocate()
3448 *user_entry_p = user_entry; in mach_memory_entry_allocate()
3453 user_entry->named_entry_port = user_handle; in mach_memory_entry_allocate()
3455 OSBacktrace(&user_entry->named_entry_bt[0], in mach_memory_entry_allocate()
3460 queue_enter(&vm_named_entry_list, user_entry, in mach_memory_entry_allocate()
3491 vm_named_entry_t user_entry; in mach_memory_object_memory_entry_64() local
3515 if (mach_memory_entry_allocate(&user_entry, &user_handle) in mach_memory_object_memory_entry_64()
3521 user_entry->size = size; in mach_memory_object_memory_entry_64()
3522 user_entry->offset = 0; in mach_memory_object_memory_entry_64()
3523 user_entry->protection = permission & VM_PROT_ALL; in mach_memory_object_memory_entry_64()
3525 SET_MAP_MEM(access, user_entry->protection); in mach_memory_object_memory_entry_64()
3526 user_entry->is_sub_map = FALSE; in mach_memory_object_memory_entry_64()
3527 assert(user_entry->ref_count == 1); in mach_memory_object_memory_entry_64()
3529 kr = vm_named_entry_from_vm_object(user_entry, object, 0, size, in mach_memory_object_memory_entry_64()
3530 (user_entry->protection & VM_PROT_ALL)); in mach_memory_object_memory_entry_64()
3534 user_entry->internal = object->internal; in mach_memory_object_memory_entry_64()
4315 vm_named_entry_t user_entry; in vm_region_object_create() local
4320 if (mach_memory_entry_allocate(&user_entry, &user_handle) in vm_region_object_create()
4333 user_entry->backing.map = new_map; in vm_region_object_create()
4334 user_entry->internal = TRUE; in vm_region_object_create()
4335 user_entry->is_sub_map = TRUE; in vm_region_object_create()
4336 user_entry->offset = 0; in vm_region_object_create()
4337 user_entry->protection = VM_PROT_ALL; in vm_region_object_create()
4338 user_entry->size = size; in vm_region_object_create()
4339 assert(user_entry->ref_count == 1); in vm_region_object_create()