Lines Matching refs:uccontext

39 	struct kern_userspace_coredump_context * uccontext = refcon;  in user_dump_init()  local
41 if (!uccontext) { in user_dump_init()
46 if (!uccontext->task) { in user_dump_init()
52 if (!uccontext->task->active) { in user_dump_init()
58 if (kdp_task_is_locked(uccontext->task)) { in user_dump_init()
64 if (kdp_vm_map_is_acquired_exclusive(uccontext->task->map)) { in user_dump_init()
76 struct kern_userspace_coredump_context * uccontext; member
97 struct kern_userspace_coredump_context * uccontext = refcon; in user_dump_save_summary() local
98 …struct user_dump_preflight_context udc_preflight = {.uccontext = uccontext, .region_count = 0, .du… in user_dump_save_summary()
102 ret = kdp_traverse_mappings(uccontext->task, in user_dump_save_summary()
112 kern_collect_userth_state_size(uccontext->task, &thread_count, &thread_state_size); in user_dump_save_summary()
138 struct kern_userspace_coredump_context * uccontext = refcon; in user_dump_save_seg_descriptions() local
142 kern_return_t ret = kdp_traverse_mappings(uccontext->task, in user_dump_save_seg_descriptions()
157 struct kern_userspace_coredump_context * uccontext = refcon; in user_dump_save_thread_state() local
161 kern_collect_userth_state_size(uccontext->task, &thread_count, &thread_state_size); in user_dump_save_thread_state()
162 queue_iterate(&uccontext->task->threads, thread, thread_t, task_threads) { in user_dump_save_thread_state()
163 kern_collect_userth_state(uccontext->task, thread, buf, thread_state_size); in user_dump_save_thread_state()
177 struct kern_userspace_coredump_context * uccontext = refcon; in user_dump_save_sw_vers_detail() local
185 …kern_return_t ret = kdp_task_dyld_info(uccontext->task, KDP_FAULT_FLAGS_ENABLE_FAULTING, &dyld_loa… in user_dump_save_sw_vers_detail()
211 struct kern_userspace_coredump_context * uccontext = refcon; in user_dump_save_segment_data() local
213 kern_return_t ret = kdp_traverse_mappings(uccontext->task, in user_dump_save_segment_data()