1 /*
2 * Copyright (c) 2000-2018 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28
29 #include <mach/mach_types.h>
30 #include <mach/machine/vm_param.h>
31 #include <mach/task.h>
32
33 #include <kern/kern_types.h>
34 #include <kern/ledger.h>
35 #include <kern/processor.h>
36 #include <kern/thread.h>
37 #include <kern/task.h>
38 #include <kern/spl.h>
39 #include <kern/ast.h>
40 #include <kern/monotonic.h>
41 #include <machine/monotonic.h>
42 #include <ipc/ipc_port.h>
43 #include <ipc/ipc_space.h>
44 #include <vm/vm_map_xnu.h>
45 #include <vm/vm_kern.h>
46 #include <vm/pmap.h>
47 #include <vm/vm_protos.h> /* last */
48 #include <sys/resource.h>
49 #include <sys/signal.h>
50 #include <sys/errno.h>
51 #include <sys/proc_require.h>
52
53 #include <machine/limits.h>
54 #include <sys/codesign.h> /* CS_CDHASH_LEN */
55
56 #undef thread_should_halt
57
58 /* BSD KERN COMPONENT INTERFACE */
59
60 extern unsigned int not_in_kdp; /* Skip acquiring locks if we're in kdp */
61
62 thread_t get_firstthread(task_t);
63 int get_task_userstop(task_t);
64 int get_thread_userstop(thread_t);
65 boolean_t current_thread_aborted(void);
66 void task_act_iterate_wth_args(task_t, void (*)(thread_t, void *), void *);
67 kern_return_t get_signalact(task_t, thread_t *, int);
68 int fill_task_rusage(task_t task, rusage_info_current *ri);
69 int fill_task_io_rusage(task_t task, rusage_info_current *ri);
70 int fill_task_qos_rusage(task_t task, rusage_info_current *ri);
71 uint64_t get_task_logical_writes(task_t task, bool external);
72 void fill_task_billed_usage(task_t task, rusage_info_current *ri);
73 void task_bsdtask_kill(task_t);
74
75 extern uint64_t get_dispatchqueue_serialno_offset_from_proc(void *p);
76 extern uint64_t get_dispatchqueue_label_offset_from_proc(void *p);
77 extern uint64_t proc_uniqueid_task(void *p, void *t);
78 extern int proc_pidversion(void *p);
79 extern int proc_getcdhash(void *p, char *cdhash);
80
81 int mach_to_bsd_errno(kern_return_t mach_err);
82 kern_return_t kern_return_for_errno(int bsd_errno);
83
84 #if MACH_BSD
85 extern void psignal(void *, int);
86 #endif
87
88 /*
89 *
90 */
91 void *
get_bsdtask_info(task_t t)92 get_bsdtask_info(task_t t)
93 {
94 void *proc_from_task = task_get_proc_raw(t);
95 proc_require(proc_from_task, PROC_REQUIRE_ALLOW_NULL | PROC_REQUIRE_ALLOW_ALL);
96 return task_has_proc(t) ? proc_from_task : NULL;
97 }
98
99 void
task_bsdtask_kill(task_t t)100 task_bsdtask_kill(task_t t)
101 {
102 void * bsd_info = get_bsdtask_info(t);
103 if (bsd_info != NULL) {
104 psignal(bsd_info, SIGKILL);
105 }
106 }
107 /*
108 *
109 */
110 void *
get_bsdthreadtask_info(thread_t th)111 get_bsdthreadtask_info(thread_t th)
112 {
113 return get_thread_ro(th)->tro_proc;
114 }
115
116 /*
117 *
118 */
119 void
set_bsdtask_info(task_t t,void * v)120 set_bsdtask_info(task_t t, void * v)
121 {
122 void *proc_from_task = task_get_proc_raw(t);
123 if (v == NULL) {
124 task_clear_has_proc(t);
125 } else {
126 if (v != proc_from_task) {
127 panic("set_bsdtask_info trying to set random bsd_info %p", v);
128 }
129 task_set_has_proc(t);
130 }
131 }
132
133 __abortlike
134 static void
__thread_ro_circularity_panic(thread_t th,thread_ro_t tro)135 __thread_ro_circularity_panic(thread_t th, thread_ro_t tro)
136 {
137 panic("tro %p points back to %p instead of %p", tro, tro->tro_owner, th);
138 }
139
140 __attribute__((always_inline))
141 thread_ro_t
get_thread_ro_unchecked(thread_t th)142 get_thread_ro_unchecked(thread_t th)
143 {
144 return th->t_tro;
145 }
146
147 thread_ro_t
get_thread_ro(thread_t th)148 get_thread_ro(thread_t th)
149 {
150 thread_ro_t tro = th->t_tro;
151
152 zone_require_ro(ZONE_ID_THREAD_RO, sizeof(struct thread_ro), tro);
153 if (tro->tro_owner != th) {
154 __thread_ro_circularity_panic(th, tro);
155 }
156 return tro;
157 }
158
159 __attribute__((always_inline))
160 thread_ro_t
current_thread_ro_unchecked(void)161 current_thread_ro_unchecked(void)
162 {
163 return get_thread_ro_unchecked(current_thread());
164 }
165
166 thread_ro_t
current_thread_ro(void)167 current_thread_ro(void)
168 {
169 return get_thread_ro(current_thread());
170 }
171
172 void
clear_thread_ro_proc(thread_t th)173 clear_thread_ro_proc(thread_t th)
174 {
175 thread_ro_t tro = get_thread_ro(th);
176
177 zalloc_ro_clear_field(ZONE_ID_THREAD_RO, tro, tro_proc);
178 }
179
180 struct uthread *
get_bsdthread_info(thread_t th)181 get_bsdthread_info(thread_t th)
182 {
183 return (struct uthread *)((uintptr_t)th + sizeof(struct thread));
184 }
185
186 thread_t
get_machthread(struct uthread * uth)187 get_machthread(struct uthread *uth)
188 {
189 return (struct thread *)((uintptr_t)uth - sizeof(struct thread));
190 }
191
192 /*
193 * This is used to remember any FS error from VNOP_PAGEIN code when
194 * invoked under vm_fault(). The value is an errno style value. It can
195 * be retrieved by exception handlers using thread_get_state().
196 */
197 void
set_thread_pagein_error(thread_t th,int error)198 set_thread_pagein_error(thread_t th, int error)
199 {
200 assert(th == current_thread());
201 if (error == 0 || th->t_pagein_error == 0) {
202 th->t_pagein_error = error;
203 }
204 }
205
206 #if defined(__x86_64__)
207 /*
208 * Returns non-zero if the thread has a non-NULL task
209 * and that task has an LDT.
210 */
211 int
thread_task_has_ldt(thread_t th)212 thread_task_has_ldt(thread_t th)
213 {
214 task_t task = get_threadtask(th);
215 return task && task->i386_ldt != 0;
216 }
217 #endif /* __x86_64__ */
218
219 /*
220 * XXX
221 */
222 int get_thread_lock_count(thread_t th); /* forced forward */
223 int
get_thread_lock_count(thread_t th __unused)224 get_thread_lock_count(thread_t th __unused)
225 {
226 /*
227 * TODO: one day: resurect counting locks held to disallow
228 * holding locks across upcalls.
229 *
230 * never worked on arm.
231 */
232 return 0;
233 }
234
235 /*
236 * Returns a thread reference.
237 */
238 thread_t
get_firstthread(task_t task)239 get_firstthread(task_t task)
240 {
241 thread_t thread = THREAD_NULL;
242 task_lock(task);
243
244 if (!task->active) {
245 task_unlock(task);
246 return THREAD_NULL;
247 }
248
249 thread = (thread_t)(void *)queue_first(&task->threads);
250
251 if (queue_end(&task->threads, (queue_entry_t)thread)) {
252 task_unlock(task);
253 return THREAD_NULL;
254 }
255
256 thread_reference(thread);
257 task_unlock(task);
258 return thread;
259 }
260
261 kern_return_t
get_signalact(task_t task,thread_t * result_out,int setast)262 get_signalact(
263 task_t task,
264 thread_t *result_out,
265 int setast)
266 {
267 kern_return_t result = KERN_SUCCESS;
268 thread_t inc, thread = THREAD_NULL;
269
270 task_lock(task);
271
272 if (!task->active) {
273 task_unlock(task);
274
275 return KERN_FAILURE;
276 }
277
278 for (inc = (thread_t)(void *)queue_first(&task->threads);
279 !queue_end(&task->threads, (queue_entry_t)inc);) {
280 thread_mtx_lock(inc);
281 if (inc->active &&
282 (inc->sched_flags & TH_SFLAG_ABORTED_MASK) != TH_SFLAG_ABORT) {
283 thread = inc;
284 break;
285 }
286 thread_mtx_unlock(inc);
287
288 inc = (thread_t)(void *)queue_next(&inc->task_threads);
289 }
290
291 if (result_out) {
292 *result_out = thread;
293 }
294
295 if (thread) {
296 if (setast) {
297 act_set_astbsd(thread);
298 }
299
300 thread_mtx_unlock(thread);
301 } else {
302 result = KERN_FAILURE;
303 }
304
305 task_unlock(task);
306
307 return result;
308 }
309
310
311 kern_return_t
check_actforsig(task_t task,thread_t thread,int setast)312 check_actforsig(
313 task_t task,
314 thread_t thread,
315 int setast)
316 {
317 kern_return_t result = KERN_FAILURE;
318 thread_t inc;
319
320 task_lock(task);
321
322 if (!task->active) {
323 task_unlock(task);
324
325 return KERN_FAILURE;
326 }
327
328 for (inc = (thread_t)(void *)queue_first(&task->threads);
329 !queue_end(&task->threads, (queue_entry_t)inc);) {
330 if (inc == thread) {
331 thread_mtx_lock(inc);
332
333 if (inc->active &&
334 (inc->sched_flags & TH_SFLAG_ABORTED_MASK) != TH_SFLAG_ABORT) {
335 result = KERN_SUCCESS;
336 break;
337 }
338
339 thread_mtx_unlock(inc);
340 break;
341 }
342
343 inc = (thread_t)(void *)queue_next(&inc->task_threads);
344 }
345
346 if (result == KERN_SUCCESS) {
347 if (setast) {
348 act_set_astbsd(thread);
349 }
350
351 thread_mtx_unlock(thread);
352 }
353
354 task_unlock(task);
355
356 return result;
357 }
358
359 ledger_t
get_task_ledger(task_t t)360 get_task_ledger(task_t t)
361 {
362 return t->ledger;
363 }
364
365 /*
366 * This is only safe to call from a thread executing in
367 * in the task's context or if the task is locked. Otherwise,
368 * the map could be switched for the task (and freed) before
369 * we go to return it here.
370 */
371 vm_map_t
get_task_map(task_t t)372 get_task_map(task_t t)
373 {
374 return t->map;
375 }
376
377 vm_map_t
get_task_map_reference(task_t t)378 get_task_map_reference(task_t t)
379 {
380 vm_map_t m;
381
382 if (t == NULL) {
383 return VM_MAP_NULL;
384 }
385
386 task_lock(t);
387 if (!t->active) {
388 task_unlock(t);
389 return VM_MAP_NULL;
390 }
391 m = t->map;
392 vm_map_reference(m);
393 task_unlock(t);
394 return m;
395 }
396
397 /*
398 *
399 */
400 ipc_space_t
get_task_ipcspace(task_t t)401 get_task_ipcspace(task_t t)
402 {
403 return t->itk_space;
404 }
405
406 int
get_task_numacts(task_t t)407 get_task_numacts(task_t t)
408 {
409 return t->thread_count;
410 }
411
412 /* does this machine need 64bit register set for signal handler */
413 int
is_64signalregset(void)414 is_64signalregset(void)
415 {
416 if (task_has_64Bit_data(current_task())) {
417 return 1;
418 }
419
420 return 0;
421 }
422
423 /*
424 * Swap in a new map for the task/thread pair; the old map reference is
425 * returned. Also does a pmap switch if thread provided is current thread.
426 */
427 vm_map_t
swap_task_map(task_t task,thread_t thread,vm_map_t map)428 swap_task_map(task_t task, thread_t thread, vm_map_t map)
429 {
430 vm_map_t old_map;
431 boolean_t doswitch = (thread == current_thread()) ? TRUE : FALSE;
432
433 if (task != get_threadtask(thread)) {
434 panic("swap_task_map");
435 }
436
437 task_lock(task);
438 mp_disable_preemption();
439
440 /* verify that the map has been activated if the task is enabled for IPC access and is not a corpse */
441 assert(!task->ipc_active || task_is_a_corpse(task) || (map->owning_task == task));
442
443 old_map = task->map;
444 thread->map = task->map = map;
445 vm_commit_pagezero_status(map);
446
447 if (doswitch) {
448 PMAP_SWITCH_USER(thread, map, cpu_number());
449 }
450 mp_enable_preemption();
451 task_unlock(task);
452
453 return old_map;
454 }
455
456 /*
457 *
458 * This is only safe to call from a thread executing in
459 * in the task's context or if the task is locked. Otherwise,
460 * the map could be switched for the task (and freed) before
461 * we go to return it here.
462 */
463 pmap_t
get_task_pmap(task_t t)464 get_task_pmap(task_t t)
465 {
466 return t->map->pmap;
467 }
468
469 /*
470 *
471 */
472 uint64_t
get_task_resident_size(task_t task)473 get_task_resident_size(task_t task)
474 {
475 uint64_t val;
476
477 ledger_get_balance(task->ledger, task_ledgers.phys_mem, (ledger_amount_t *) &val);
478 return val;
479 }
480
481 uint64_t
get_task_compressed(task_t task)482 get_task_compressed(task_t task)
483 {
484 uint64_t val;
485
486 ledger_get_balance(task->ledger, task_ledgers.internal_compressed, (ledger_amount_t *) &val);
487 return val;
488 }
489
490 uint64_t
get_task_resident_max(task_t task)491 get_task_resident_max(task_t task)
492 {
493 uint64_t val;
494
495 ledger_get_lifetime_max(task->ledger, task_ledgers.phys_mem, (ledger_amount_t *) &val);
496 return val;
497 }
498
499 /*
500 * Get the balance for a given field in the task ledger.
501 * Returns 0 if the entry is invalid.
502 */
503 static uint64_t
get_task_ledger_balance(task_t task,int entry)504 get_task_ledger_balance(task_t task, int entry)
505 {
506 ledger_amount_t balance = 0;
507
508 ledger_get_balance(task->ledger, entry, &balance);
509 return balance;
510 }
511
512 uint64_t
get_task_purgeable_size(task_t task)513 get_task_purgeable_size(task_t task)
514 {
515 kern_return_t ret;
516 ledger_amount_t balance = 0;
517 uint64_t volatile_size = 0;
518
519 ret = ledger_get_balance(task->ledger, task_ledgers.purgeable_volatile, &balance);
520 if (ret != KERN_SUCCESS) {
521 return 0;
522 }
523
524 volatile_size += balance;
525
526 ret = ledger_get_balance(task->ledger, task_ledgers.purgeable_volatile_compressed, &balance);
527 if (ret != KERN_SUCCESS) {
528 return 0;
529 }
530
531 volatile_size += balance;
532
533 return volatile_size;
534 }
535
536 /*
537 *
538 */
539 uint64_t
get_task_phys_footprint(task_t task)540 get_task_phys_footprint(task_t task)
541 {
542 return get_task_ledger_balance(task, task_ledgers.phys_footprint);
543 }
544
545 #if CONFIG_LEDGER_INTERVAL_MAX
546 /*
547 *
548 */
549 uint64_t
get_task_phys_footprint_interval_max(task_t task,int reset)550 get_task_phys_footprint_interval_max(task_t task, int reset)
551 {
552 kern_return_t ret;
553 ledger_amount_t max;
554
555 ret = ledger_get_interval_max(task->ledger, task_ledgers.phys_footprint, &max, reset);
556
557 if (KERN_SUCCESS == ret) {
558 return max;
559 }
560
561 return 0;
562 }
563 #endif /* CONFIG_LEDGER_INTERVAL_MAX */
564
565 /*
566 *
567 */
568 uint64_t
get_task_phys_footprint_lifetime_max(task_t task)569 get_task_phys_footprint_lifetime_max(task_t task)
570 {
571 kern_return_t ret;
572 ledger_amount_t max;
573
574 ret = ledger_get_lifetime_max(task->ledger, task_ledgers.phys_footprint, &max);
575
576 if (KERN_SUCCESS == ret) {
577 return max;
578 }
579
580 return 0;
581 }
582
583 /*
584 *
585 */
586 uint64_t
get_task_phys_footprint_limit(task_t task)587 get_task_phys_footprint_limit(task_t task)
588 {
589 kern_return_t ret;
590 ledger_amount_t max;
591
592 ret = ledger_get_limit(task->ledger, task_ledgers.phys_footprint, &max);
593 if (KERN_SUCCESS == ret) {
594 return max;
595 }
596
597 return 0;
598 }
599
600 uint64_t
get_task_internal(task_t task)601 get_task_internal(task_t task)
602 {
603 return get_task_ledger_balance(task, task_ledgers.internal);
604 }
605
606 uint64_t
get_task_internal_compressed(task_t task)607 get_task_internal_compressed(task_t task)
608 {
609 return get_task_ledger_balance(task, task_ledgers.internal_compressed);
610 }
611
612 uint64_t
get_task_purgeable_nonvolatile(task_t task)613 get_task_purgeable_nonvolatile(task_t task)
614 {
615 return get_task_ledger_balance(task, task_ledgers.purgeable_nonvolatile);
616 }
617
618 uint64_t
get_task_purgeable_nonvolatile_compressed(task_t task)619 get_task_purgeable_nonvolatile_compressed(task_t task)
620 {
621 return get_task_ledger_balance(task, task_ledgers.purgeable_nonvolatile_compressed);
622 }
623
624 uint64_t
get_task_alternate_accounting(task_t task)625 get_task_alternate_accounting(task_t task)
626 {
627 return get_task_ledger_balance(task, task_ledgers.alternate_accounting);
628 }
629
630 uint64_t
get_task_alternate_accounting_compressed(task_t task)631 get_task_alternate_accounting_compressed(task_t task)
632 {
633 return get_task_ledger_balance(task, task_ledgers.alternate_accounting_compressed);
634 }
635
636 uint64_t
get_task_page_table(task_t task)637 get_task_page_table(task_t task)
638 {
639 return get_task_ledger_balance(task, task_ledgers.page_table);
640 }
641
642 #if CONFIG_FREEZE
643 uint64_t
get_task_frozen_to_swap(task_t task)644 get_task_frozen_to_swap(task_t task)
645 {
646 return get_task_ledger_balance(task, task_ledgers.frozen_to_swap);
647 }
648 #endif /* CONFIG_FREEZE */
649
650 uint64_t
get_task_iokit_mapped(task_t task)651 get_task_iokit_mapped(task_t task)
652 {
653 return get_task_ledger_balance(task, task_ledgers.iokit_mapped);
654 }
655
656 uint64_t
get_task_network_nonvolatile(task_t task)657 get_task_network_nonvolatile(task_t task)
658 {
659 return get_task_ledger_balance(task, task_ledgers.network_nonvolatile);
660 }
661
662 uint64_t
get_task_network_nonvolatile_compressed(task_t task)663 get_task_network_nonvolatile_compressed(task_t task)
664 {
665 return get_task_ledger_balance(task, task_ledgers.network_nonvolatile_compressed);
666 }
667
668 uint64_t
get_task_wired_mem(task_t task)669 get_task_wired_mem(task_t task)
670 {
671 return get_task_ledger_balance(task, task_ledgers.wired_mem);
672 }
673
674 uint64_t
get_task_tagged_footprint(task_t task)675 get_task_tagged_footprint(task_t task)
676 {
677 kern_return_t ret;
678 ledger_amount_t credit, debit;
679
680 ret = ledger_get_entries(task->ledger, task_ledgers.tagged_footprint, &credit, &debit);
681 if (KERN_SUCCESS == ret) {
682 return credit - debit;
683 }
684
685 return 0;
686 }
687
688 uint64_t
get_task_tagged_footprint_compressed(task_t task)689 get_task_tagged_footprint_compressed(task_t task)
690 {
691 kern_return_t ret;
692 ledger_amount_t credit, debit;
693
694 ret = ledger_get_entries(task->ledger, task_ledgers.tagged_footprint_compressed, &credit, &debit);
695 if (KERN_SUCCESS == ret) {
696 return credit - debit;
697 }
698
699 return 0;
700 }
701
702 uint64_t
get_task_media_footprint(task_t task)703 get_task_media_footprint(task_t task)
704 {
705 kern_return_t ret;
706 ledger_amount_t credit, debit;
707
708 ret = ledger_get_entries(task->ledger, task_ledgers.media_footprint, &credit, &debit);
709 if (KERN_SUCCESS == ret) {
710 return credit - debit;
711 }
712
713 return 0;
714 }
715
716 uint64_t
get_task_media_footprint_compressed(task_t task)717 get_task_media_footprint_compressed(task_t task)
718 {
719 kern_return_t ret;
720 ledger_amount_t credit, debit;
721
722 ret = ledger_get_entries(task->ledger, task_ledgers.media_footprint_compressed, &credit, &debit);
723 if (KERN_SUCCESS == ret) {
724 return credit - debit;
725 }
726
727 return 0;
728 }
729
730 uint64_t
get_task_graphics_footprint(task_t task)731 get_task_graphics_footprint(task_t task)
732 {
733 kern_return_t ret;
734 ledger_amount_t credit, debit;
735
736 ret = ledger_get_entries(task->ledger, task_ledgers.graphics_footprint, &credit, &debit);
737 if (KERN_SUCCESS == ret) {
738 return credit - debit;
739 }
740
741 return 0;
742 }
743
744
745 uint64_t
get_task_graphics_footprint_compressed(task_t task)746 get_task_graphics_footprint_compressed(task_t task)
747 {
748 kern_return_t ret;
749 ledger_amount_t credit, debit;
750
751 ret = ledger_get_entries(task->ledger, task_ledgers.graphics_footprint_compressed, &credit, &debit);
752 if (KERN_SUCCESS == ret) {
753 return credit - debit;
754 }
755
756 return 0;
757 }
758
759 uint64_t
get_task_neural_footprint(task_t task)760 get_task_neural_footprint(task_t task)
761 {
762 kern_return_t ret;
763 ledger_amount_t credit, debit;
764
765 ret = ledger_get_entries(task->ledger, task_ledgers.neural_footprint, &credit, &debit);
766 if (KERN_SUCCESS == ret) {
767 return credit - debit;
768 }
769
770 return 0;
771 }
772
773 uint64_t
get_task_neural_footprint_compressed(task_t task)774 get_task_neural_footprint_compressed(task_t task)
775 {
776 kern_return_t ret;
777 ledger_amount_t credit, debit;
778
779 ret = ledger_get_entries(task->ledger, task_ledgers.neural_footprint_compressed, &credit, &debit);
780 if (KERN_SUCCESS == ret) {
781 return credit - debit;
782 }
783
784 return 0;
785 }
786
787 uint64_t
get_task_neural_nofootprint_total(task_t task)788 get_task_neural_nofootprint_total(task_t task)
789 {
790 kern_return_t ret;
791 ledger_amount_t credit, debit;
792
793 ret = ledger_get_entries(task->ledger, task_ledgers.neural_nofootprint_total, &credit, &debit);
794 if (KERN_SUCCESS == ret) {
795 return credit - debit;
796 }
797
798 return 0;
799 }
800
801 #if CONFIG_LEDGER_INTERVAL_MAX
802 uint64_t
get_task_neural_nofootprint_total_interval_max(task_t task,int reset)803 get_task_neural_nofootprint_total_interval_max(task_t task, int reset)
804 {
805 kern_return_t ret;
806 ledger_amount_t max;
807
808 ret = ledger_get_interval_max(task->ledger, task_ledgers.neural_nofootprint_total, &max, reset);
809
810 if (KERN_SUCCESS == ret) {
811 return max;
812 }
813
814 return 0;
815 }
816 #endif /* CONFIG_LEDGER_INTERVAL_MAX */
817
818 uint64_t
get_task_neural_nofootprint_total_lifetime_max(task_t task)819 get_task_neural_nofootprint_total_lifetime_max(task_t task)
820 {
821 kern_return_t ret;
822 ledger_amount_t max;
823
824 ret = ledger_get_lifetime_max(task->ledger, task_ledgers.neural_nofootprint_total, &max);
825
826 if (KERN_SUCCESS == ret) {
827 return max;
828 }
829
830 return 0;
831 }
832
833 uint64_t
get_task_cpu_time(task_t task)834 get_task_cpu_time(task_t task)
835 {
836 return get_task_ledger_balance(task, task_ledgers.cpu_time);
837 }
838
839 uint32_t
get_task_loadTag(task_t task)840 get_task_loadTag(task_t task)
841 {
842 return os_atomic_load(&task->loadTag, relaxed);
843 }
844
845 uint32_t
set_task_loadTag(task_t task,uint32_t loadTag)846 set_task_loadTag(task_t task, uint32_t loadTag)
847 {
848 return os_atomic_xchg(&task->loadTag, loadTag, relaxed);
849 }
850
851
852 task_t
get_threadtask(thread_t th)853 get_threadtask(thread_t th)
854 {
855 return get_thread_ro(th)->tro_task;
856 }
857
858 task_t
get_threadtask_early(thread_t th)859 get_threadtask_early(thread_t th)
860 {
861 if (__improbable(startup_phase < STARTUP_SUB_EARLY_BOOT)) {
862 if (th == THREAD_NULL || th->t_tro == NULL) {
863 return TASK_NULL;
864 }
865 }
866 return get_threadtask(th);
867 }
868
869 /*
870 *
871 */
872 vm_map_offset_t
get_map_min(vm_map_t map)873 get_map_min(
874 vm_map_t map)
875 {
876 return vm_map_min(map);
877 }
878
879 /*
880 *
881 */
882 vm_map_offset_t
get_map_max(vm_map_t map)883 get_map_max(
884 vm_map_t map)
885 {
886 return vm_map_max(map);
887 }
888 vm_map_size_t
get_vmmap_size(vm_map_t map)889 get_vmmap_size(
890 vm_map_t map)
891 {
892 return vm_map_adjusted_size(map);
893 }
894 int
get_task_page_size(task_t task)895 get_task_page_size(
896 task_t task)
897 {
898 return vm_map_page_size(task->map);
899 }
900
901 #if CONFIG_COREDUMP
902
903 static int
get_vmsubmap_entries(vm_map_t map,vm_object_offset_t start,vm_object_offset_t end)904 get_vmsubmap_entries(
905 vm_map_t map,
906 vm_object_offset_t start,
907 vm_object_offset_t end)
908 {
909 int total_entries = 0;
910 vm_map_entry_t entry;
911
912 vmlp_api_start(GET_VMSUBMAP_ENTRIES);
913
914 if (not_in_kdp) {
915 vm_map_lock_read(map);
916 }
917 vmlp_range_event(map, start, end - start);
918 entry = vm_map_first_entry(map);
919 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < start)) {
920 entry = entry->vme_next;
921 }
922
923 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < end)) {
924 if (entry->is_sub_map) {
925 total_entries +=
926 get_vmsubmap_entries(VME_SUBMAP(entry),
927 VME_OFFSET(entry),
928 (VME_OFFSET(entry) +
929 entry->vme_end -
930 entry->vme_start));
931 } else {
932 total_entries += 1;
933 }
934 entry = entry->vme_next;
935 }
936 if (not_in_kdp) {
937 vm_map_unlock_read(map);
938 }
939 vmlp_api_end(GET_VMSUBMAP_ENTRIES, total_entries);
940 return total_entries;
941 }
942
943 int
get_vmmap_entries(vm_map_t map)944 get_vmmap_entries(
945 vm_map_t map)
946 {
947 int total_entries = 0;
948 vm_map_entry_t entry;
949
950 vmlp_api_start(GET_VMMAP_ENTRIES);
951
952 if (not_in_kdp) {
953 vm_map_lock_read(map);
954 }
955 entry = vm_map_first_entry(map);
956
957 while (entry != vm_map_to_entry(map)) {
958 vmlp_range_event_entry(map, entry);
959 if (entry->is_sub_map) {
960 total_entries +=
961 get_vmsubmap_entries(VME_SUBMAP(entry),
962 VME_OFFSET(entry),
963 (VME_OFFSET(entry) +
964 entry->vme_end -
965 entry->vme_start));
966 } else {
967 total_entries += 1;
968 }
969 entry = entry->vme_next;
970 }
971 if (not_in_kdp) {
972 vm_map_unlock_read(map);
973 }
974 vmlp_api_end(GET_VMMAP_ENTRIES, total_entries);
975 return total_entries;
976 }
977 #endif /* CONFIG_COREDUMP */
978
979 int
get_task_userstop(task_t task)980 get_task_userstop(
981 task_t task)
982 {
983 return task->user_stop_count;
984 }
985
986 int
get_thread_userstop(thread_t th)987 get_thread_userstop(
988 thread_t th)
989 {
990 return th->user_stop_count;
991 }
992
993 boolean_t
get_task_pidsuspended(task_t task)994 get_task_pidsuspended(
995 task_t task)
996 {
997 return task->pidsuspended;
998 }
999
1000 boolean_t
get_task_frozen(task_t task)1001 get_task_frozen(
1002 task_t task)
1003 {
1004 return task->frozen;
1005 }
1006
1007 boolean_t
thread_should_abort(thread_t th)1008 thread_should_abort(
1009 thread_t th)
1010 {
1011 return (th->sched_flags & TH_SFLAG_ABORTED_MASK) == TH_SFLAG_ABORT;
1012 }
1013
1014 /*
1015 * This routine is like thread_should_abort() above. It checks to
1016 * see if the current thread is aborted. But unlike above, it also
1017 * checks to see if thread is safely aborted. If so, it returns
1018 * that fact, and clears the condition (safe aborts only should
1019 * have a single effect, and a poll of the abort status
1020 * qualifies.
1021 */
1022 boolean_t
current_thread_aborted(void)1023 current_thread_aborted(
1024 void)
1025 {
1026 thread_t th = current_thread();
1027 spl_t s;
1028
1029 if ((th->sched_flags & TH_SFLAG_ABORTED_MASK) == TH_SFLAG_ABORT &&
1030 (th->options & TH_OPT_INTMASK) != THREAD_UNINT) {
1031 return TRUE;
1032 }
1033 if (th->sched_flags & TH_SFLAG_ABORTSAFELY) {
1034 s = splsched();
1035 thread_lock(th);
1036 if (th->sched_flags & TH_SFLAG_ABORTSAFELY) {
1037 th->sched_flags &= ~TH_SFLAG_ABORTED_MASK;
1038 }
1039 thread_unlock(th);
1040 splx(s);
1041 }
1042 return FALSE;
1043 }
1044
1045 void
task_act_iterate_wth_args(task_t task,void (* func_callback)(thread_t,void *),void * func_arg)1046 task_act_iterate_wth_args(
1047 task_t task,
1048 void (*func_callback)(thread_t, void *),
1049 void *func_arg)
1050 {
1051 thread_t inc;
1052
1053 task_lock(task);
1054
1055 for (inc = (thread_t)(void *)queue_first(&task->threads);
1056 !queue_end(&task->threads, (queue_entry_t)inc);) {
1057 (void) (*func_callback)(inc, func_arg);
1058 inc = (thread_t)(void *)queue_next(&inc->task_threads);
1059 }
1060
1061 task_unlock(task);
1062 }
1063
1064 #include <sys/bsdtask_info.h>
1065
1066 void
fill_taskprocinfo(task_t task,struct proc_taskinfo_internal * ptinfo)1067 fill_taskprocinfo(task_t task, struct proc_taskinfo_internal * ptinfo)
1068 {
1069 vm_map_t map;
1070 task_absolutetime_info_data_t tinfo;
1071 thread_t thread;
1072 uint32_t cswitch = 0, numrunning = 0;
1073 uint32_t syscalls_unix = 0;
1074 uint32_t syscalls_mach = 0;
1075
1076 task_lock(task);
1077
1078 map = (task == kernel_task)? kernel_map: task->map;
1079
1080 ptinfo->pti_virtual_size = vm_map_adjusted_size(map);
1081 ledger_get_balance(task->ledger, task_ledgers.phys_mem, (ledger_amount_t *) &ptinfo->pti_resident_size);
1082
1083 ptinfo->pti_policy = ((task != kernel_task)?
1084 POLICY_TIMESHARE: POLICY_RR);
1085
1086 queue_iterate(&task->threads, thread, thread_t, task_threads) {
1087 spl_t x;
1088
1089 if (thread->options & TH_OPT_IDLE_THREAD) {
1090 continue;
1091 }
1092
1093 x = splsched();
1094 thread_lock(thread);
1095
1096 if ((thread->state & TH_RUN) == TH_RUN) {
1097 numrunning++;
1098 }
1099 cswitch += thread->c_switch;
1100
1101 syscalls_unix += thread->syscalls_unix;
1102 syscalls_mach += thread->syscalls_mach;
1103
1104 thread_unlock(thread);
1105 splx(x);
1106 }
1107
1108 struct recount_times_mach term_times = recount_task_terminated_times(task);
1109 struct recount_times_mach total_times = recount_task_times(task);
1110
1111 tinfo.threads_user = total_times.rtm_user - term_times.rtm_user;
1112 tinfo.threads_system = total_times.rtm_system - term_times.rtm_system;
1113 ptinfo->pti_threads_system = tinfo.threads_system;
1114 ptinfo->pti_threads_user = tinfo.threads_user;
1115
1116 ptinfo->pti_total_system = total_times.rtm_system;
1117 ptinfo->pti_total_user = total_times.rtm_user;
1118
1119 ptinfo->pti_faults = (int32_t) MIN(counter_load(&task->faults), INT32_MAX);
1120 ptinfo->pti_pageins = (int32_t) MIN(counter_load(&task->pageins), INT32_MAX);
1121 ptinfo->pti_cow_faults = (int32_t) MIN(counter_load(&task->cow_faults), INT32_MAX);
1122 ptinfo->pti_messages_sent = (int32_t) MIN(counter_load(&task->messages_sent), INT32_MAX);
1123 ptinfo->pti_messages_received = (int32_t) MIN(counter_load(&task->messages_received), INT32_MAX);
1124 ptinfo->pti_syscalls_mach = (int32_t) MIN(task->syscalls_mach + syscalls_mach, INT32_MAX);
1125 ptinfo->pti_syscalls_unix = (int32_t) MIN(task->syscalls_unix + syscalls_unix, INT32_MAX);
1126 ptinfo->pti_csw = (int32_t) MIN(task->c_switch + cswitch, INT32_MAX);
1127 ptinfo->pti_threadnum = task->thread_count;
1128 ptinfo->pti_numrunning = numrunning;
1129 ptinfo->pti_priority = task->priority;
1130
1131 task_unlock(task);
1132 }
1133
1134 int
fill_taskthreadinfo(task_t task,uint64_t thaddr,bool thuniqueid,struct proc_threadinfo_internal * ptinfo,void * vpp,int * vidp)1135 fill_taskthreadinfo(task_t task, uint64_t thaddr, bool thuniqueid, struct proc_threadinfo_internal * ptinfo, void * vpp, int *vidp)
1136 {
1137 thread_t thact;
1138 int err = 0;
1139 mach_msg_type_number_t count;
1140 thread_basic_info_data_t basic_info;
1141 kern_return_t kret;
1142 uint64_t addr = 0;
1143
1144 task_lock(task);
1145
1146 for (thact = (thread_t)(void *)queue_first(&task->threads);
1147 !queue_end(&task->threads, (queue_entry_t)thact);) {
1148 addr = (thuniqueid) ? thact->thread_id : thact->machine.cthread_self;
1149 if (addr == thaddr) {
1150 count = THREAD_BASIC_INFO_COUNT;
1151 if ((kret = thread_info_internal(thact, THREAD_BASIC_INFO, (thread_info_t)&basic_info, &count)) != KERN_SUCCESS) {
1152 err = 1;
1153 goto out;
1154 }
1155 ptinfo->pth_user_time = (((uint64_t)basic_info.user_time.seconds * NSEC_PER_SEC) + ((uint64_t)basic_info.user_time.microseconds * NSEC_PER_USEC));
1156 ptinfo->pth_system_time = (((uint64_t)basic_info.system_time.seconds * NSEC_PER_SEC) + ((uint64_t)basic_info.system_time.microseconds * NSEC_PER_USEC));
1157
1158 ptinfo->pth_cpu_usage = basic_info.cpu_usage;
1159 ptinfo->pth_policy = basic_info.policy;
1160 ptinfo->pth_run_state = basic_info.run_state;
1161 ptinfo->pth_flags = basic_info.flags;
1162 ptinfo->pth_sleep_time = basic_info.sleep_time;
1163 ptinfo->pth_curpri = thact->sched_pri;
1164 ptinfo->pth_priority = thact->base_pri;
1165 ptinfo->pth_maxpriority = thact->max_priority;
1166
1167 if (vpp != NULL) {
1168 bsd_threadcdir(get_bsdthread_info(thact), vpp, vidp);
1169 }
1170 bsd_getthreadname(get_bsdthread_info(thact), ptinfo->pth_name);
1171 err = 0;
1172 goto out;
1173 }
1174 thact = (thread_t)(void *)queue_next(&thact->task_threads);
1175 }
1176 err = 1;
1177
1178 out:
1179 task_unlock(task);
1180 return err;
1181 }
1182
1183 int
fill_taskthreadlist(task_t task,void * buffer,int thcount,bool thuniqueid)1184 fill_taskthreadlist(task_t task, void * buffer, int thcount, bool thuniqueid)
1185 {
1186 int numthr = 0;
1187 thread_t thact;
1188 uint64_t * uptr;
1189 uint64_t thaddr;
1190
1191 uptr = (uint64_t *)buffer;
1192
1193 task_lock(task);
1194
1195 for (thact = (thread_t)(void *)queue_first(&task->threads);
1196 !queue_end(&task->threads, (queue_entry_t)thact);) {
1197 thaddr = (thuniqueid) ? thact->thread_id : thact->machine.cthread_self;
1198 *uptr++ = thaddr;
1199 numthr++;
1200 if (numthr >= thcount) {
1201 goto out;
1202 }
1203 thact = (thread_t)(void *)queue_next(&thact->task_threads);
1204 }
1205
1206 out:
1207 task_unlock(task);
1208 return (int)(numthr * sizeof(uint64_t));
1209 }
1210
1211 int
fill_taskthreadschedinfo(task_t task,uint64_t thread_id,struct proc_threadschedinfo_internal * thread_sched_info)1212 fill_taskthreadschedinfo(task_t task, uint64_t thread_id, struct proc_threadschedinfo_internal *thread_sched_info)
1213 {
1214 int err = 0;
1215
1216 thread_t thread = current_thread();
1217
1218 /*
1219 * Looking up threads is pretty expensive and not realtime-safe
1220 * right now, requiring locking the task and iterating over all
1221 * threads. As long as that is the case, we officially only
1222 * support getting this info for the current thread.
1223 */
1224 if (task != current_task() || thread_id != thread->thread_id) {
1225 return -1;
1226 }
1227
1228 #if SCHED_HYGIENE_DEBUG
1229 absolutetime_to_nanoseconds(thread->machine.int_time_mt, &thread_sched_info->int_time_ns);
1230 #else
1231 (void)thread;
1232 thread_sched_info->int_time_ns = 0;
1233 #endif
1234
1235 return err;
1236 }
1237
1238 int
get_numthreads(task_t task)1239 get_numthreads(task_t task)
1240 {
1241 return task->thread_count;
1242 }
1243
1244 /*
1245 * Gather the various pieces of info about the designated task,
1246 * and collect it all into a single rusage_info.
1247 */
1248 int
fill_task_rusage(task_t task,rusage_info_current * ri)1249 fill_task_rusage(task_t task, rusage_info_current *ri)
1250 {
1251 struct task_power_info powerinfo;
1252
1253 assert(task != TASK_NULL);
1254 task_lock(task);
1255
1256 struct task_power_info_extra extra = { 0 };
1257 task_power_info_locked(task, &powerinfo, NULL, NULL, &extra);
1258 ri->ri_pkg_idle_wkups = powerinfo.task_platform_idle_wakeups;
1259 ri->ri_interrupt_wkups = powerinfo.task_interrupt_wakeups;
1260 ri->ri_user_time = powerinfo.total_user;
1261 ri->ri_system_time = powerinfo.total_system;
1262 ri->ri_runnable_time = extra.runnable_time;
1263 ri->ri_cycles = extra.cycles;
1264 ri->ri_instructions = extra.instructions;
1265 ri->ri_pcycles = extra.pcycles;
1266 ri->ri_pinstructions = extra.pinstructions;
1267 ri->ri_user_ptime = extra.user_ptime;
1268 ri->ri_system_ptime = extra.system_ptime;
1269 ri->ri_energy_nj = extra.energy;
1270 ri->ri_penergy_nj = extra.penergy;
1271 ri->ri_secure_time_in_system = extra.secure_time;
1272 ri->ri_secure_ptime_in_system = extra.secure_ptime;
1273
1274 ri->ri_phys_footprint = get_task_phys_footprint(task);
1275 ledger_get_balance(task->ledger, task_ledgers.phys_mem,
1276 (ledger_amount_t *)&ri->ri_resident_size);
1277 ri->ri_wired_size = get_task_wired_mem(task);
1278
1279 ledger_get_balance(task->ledger, task_ledgers.neural_nofootprint_total,
1280 (ledger_amount_t *)&ri->ri_neural_footprint);
1281 ri->ri_pageins = counter_load(&task->pageins);
1282
1283 task_unlock(task);
1284 return 0;
1285 }
1286
1287 void
fill_task_billed_usage(task_t task __unused,rusage_info_current * ri)1288 fill_task_billed_usage(task_t task __unused, rusage_info_current *ri)
1289 {
1290 bank_billed_balance_safe(task, &ri->ri_billed_system_time, &ri->ri_billed_energy);
1291 bank_serviced_balance_safe(task, &ri->ri_serviced_system_time, &ri->ri_serviced_energy);
1292 }
1293
1294 int
fill_task_io_rusage(task_t task,rusage_info_current * ri)1295 fill_task_io_rusage(task_t task, rusage_info_current *ri)
1296 {
1297 assert(task != TASK_NULL);
1298 task_lock(task);
1299
1300 if (task->task_io_stats) {
1301 ri->ri_diskio_bytesread = task->task_io_stats->disk_reads.size;
1302 ri->ri_diskio_byteswritten = (task->task_io_stats->total_io.size - task->task_io_stats->disk_reads.size);
1303 } else {
1304 /* I/O Stats unavailable */
1305 ri->ri_diskio_bytesread = 0;
1306 ri->ri_diskio_byteswritten = 0;
1307 }
1308 task_unlock(task);
1309 return 0;
1310 }
1311
1312 int
fill_task_qos_rusage(task_t task,rusage_info_current * ri)1313 fill_task_qos_rusage(task_t task, rusage_info_current *ri)
1314 {
1315 thread_t thread;
1316
1317 assert(task != TASK_NULL);
1318 task_lock(task);
1319
1320 /* Rollup QoS time of all the threads to task */
1321 queue_iterate(&task->threads, thread, thread_t, task_threads) {
1322 if (thread->options & TH_OPT_IDLE_THREAD) {
1323 continue;
1324 }
1325
1326 thread_update_qos_cpu_time(thread);
1327 }
1328 ri->ri_cpu_time_qos_default = task->cpu_time_eqos_stats.cpu_time_qos_default;
1329 ri->ri_cpu_time_qos_maintenance = task->cpu_time_eqos_stats.cpu_time_qos_maintenance;
1330 ri->ri_cpu_time_qos_background = task->cpu_time_eqos_stats.cpu_time_qos_background;
1331 ri->ri_cpu_time_qos_utility = task->cpu_time_eqos_stats.cpu_time_qos_utility;
1332 ri->ri_cpu_time_qos_legacy = task->cpu_time_eqos_stats.cpu_time_qos_legacy;
1333 ri->ri_cpu_time_qos_user_initiated = task->cpu_time_eqos_stats.cpu_time_qos_user_initiated;
1334 ri->ri_cpu_time_qos_user_interactive = task->cpu_time_eqos_stats.cpu_time_qos_user_interactive;
1335
1336 task_unlock(task);
1337 return 0;
1338 }
1339
1340 uint64_t
get_task_logical_writes(task_t task,bool external)1341 get_task_logical_writes(task_t task, bool external)
1342 {
1343 assert(task != TASK_NULL);
1344 struct ledger_entry_info lei;
1345 int entry = external ? task_ledgers.logical_writes_to_external :
1346 task_ledgers.logical_writes;
1347
1348 task_lock(task);
1349 ledger_get_entry_info(task->ledger, entry, &lei);
1350 task_unlock(task);
1351
1352 return lei.lei_balance;
1353 }
1354
1355 uint64_t
get_task_dispatchqueue_serialno_offset(task_t task)1356 get_task_dispatchqueue_serialno_offset(task_t task)
1357 {
1358 uint64_t dq_serialno_offset = 0;
1359 void *bsd_info = get_bsdtask_info(task);
1360
1361 if (bsd_info) {
1362 dq_serialno_offset = get_dispatchqueue_serialno_offset_from_proc(bsd_info);
1363 }
1364
1365 return dq_serialno_offset;
1366 }
1367
1368 uint64_t
get_task_dispatchqueue_label_offset(task_t task)1369 get_task_dispatchqueue_label_offset(task_t task)
1370 {
1371 uint64_t dq_label_offset = 0;
1372 void *bsd_info = get_bsdtask_info(task);
1373
1374 if (bsd_info) {
1375 dq_label_offset = get_dispatchqueue_label_offset_from_proc(bsd_info);
1376 }
1377
1378 return dq_label_offset;
1379 }
1380
1381 uint64_t
get_task_uniqueid(task_t task)1382 get_task_uniqueid(task_t task)
1383 {
1384 void *bsd_info = get_bsdtask_info(task);
1385
1386 if (bsd_info) {
1387 return proc_uniqueid_task(bsd_info, task);
1388 } else {
1389 return UINT64_MAX;
1390 }
1391 }
1392
1393 int
get_task_version(task_t task)1394 get_task_version(task_t task)
1395 {
1396 void *bsd_info = get_bsdtask_info(task);
1397
1398 if (bsd_info) {
1399 return proc_pidversion(bsd_info);
1400 } else {
1401 return INT_MAX;
1402 }
1403 }
1404
1405 #if CONFIG_MACF
1406 struct label *
get_task_crash_label(task_t task)1407 get_task_crash_label(task_t task)
1408 {
1409 return task->crash_label;
1410 }
1411
1412 void
set_task_crash_label(task_t task,struct label * label)1413 set_task_crash_label(task_t task, struct label *label)
1414 {
1415 task->crash_label = label;
1416 }
1417 #endif
1418
1419 int
fill_taskipctableinfo(task_t task,uint32_t * table_size,uint32_t * table_free)1420 fill_taskipctableinfo(task_t task, uint32_t *table_size, uint32_t *table_free)
1421 {
1422 ipc_space_t space = task->itk_space;
1423 if (space == NULL) {
1424 return -1;
1425 }
1426
1427 is_read_lock(space);
1428 if (!is_active(space)) {
1429 is_read_unlock(space);
1430 return -1;
1431 }
1432
1433 *table_size = ipc_entry_table_count(is_active_table(space));
1434 *table_free = space->is_table_free;
1435
1436 is_read_unlock(space);
1437
1438 return 0;
1439 }
1440
1441 int
get_task_cdhash(task_t task,char cdhash[static CS_CDHASH_LEN])1442 get_task_cdhash(task_t task, char cdhash[static CS_CDHASH_LEN])
1443 {
1444 int result = 0;
1445 void *bsd_info = NULL;
1446
1447 task_lock(task);
1448 bsd_info = get_bsdtask_info(task);
1449 result = bsd_info ? proc_getcdhash(bsd_info, cdhash) : ESRCH;
1450 task_unlock(task);
1451
1452 return result;
1453 }
1454
1455 bool
current_thread_in_kernel_fault(void)1456 current_thread_in_kernel_fault(void)
1457 {
1458 if (current_thread()->recover) {
1459 return true;
1460 }
1461 return false;
1462 }
1463
1464 /* moved from ubc_subr.c */
1465 int
mach_to_bsd_errno(kern_return_t mach_err)1466 mach_to_bsd_errno(kern_return_t mach_err)
1467 {
1468 switch (mach_err) {
1469 case KERN_SUCCESS:
1470 return 0;
1471
1472 case KERN_INVALID_ADDRESS:
1473 case KERN_INVALID_ARGUMENT:
1474 case KERN_NOT_IN_SET:
1475 case KERN_INVALID_NAME:
1476 case KERN_INVALID_TASK:
1477 case KERN_INVALID_RIGHT:
1478 case KERN_INVALID_VALUE:
1479 case KERN_INVALID_CAPABILITY:
1480 case KERN_INVALID_HOST:
1481 case KERN_MEMORY_PRESENT:
1482 case KERN_INVALID_PROCESSOR_SET:
1483 case KERN_INVALID_POLICY:
1484 case KERN_ALREADY_WAITING:
1485 case KERN_DEFAULT_SET:
1486 case KERN_EXCEPTION_PROTECTED:
1487 case KERN_INVALID_LEDGER:
1488 case KERN_INVALID_MEMORY_CONTROL:
1489 case KERN_INVALID_SECURITY:
1490 case KERN_NOT_DEPRESSED:
1491 case KERN_LOCK_OWNED:
1492 case KERN_LOCK_OWNED_SELF:
1493 return EINVAL;
1494
1495 case KERN_NOT_RECEIVER:
1496 case KERN_NO_ACCESS:
1497 case KERN_POLICY_STATIC:
1498 return EACCES;
1499
1500 case KERN_NO_SPACE:
1501 case KERN_RESOURCE_SHORTAGE:
1502 case KERN_UREFS_OVERFLOW:
1503 case KERN_INVALID_OBJECT:
1504 return ENOMEM;
1505
1506 case KERN_MEMORY_FAILURE:
1507 case KERN_MEMORY_ERROR:
1508 case KERN_PROTECTION_FAILURE:
1509 return EFAULT;
1510
1511 case KERN_POLICY_LIMIT:
1512 case KERN_CODESIGN_ERROR:
1513 case KERN_DENIED:
1514 return EPERM;
1515
1516 case KERN_ALREADY_IN_SET:
1517 case KERN_NAME_EXISTS:
1518 case KERN_RIGHT_EXISTS:
1519 return EEXIST;
1520
1521 case KERN_ABORTED:
1522 return EINTR;
1523
1524 case KERN_TERMINATED:
1525 case KERN_LOCK_SET_DESTROYED:
1526 case KERN_LOCK_UNSTABLE:
1527 case KERN_SEMAPHORE_DESTROYED:
1528 case KERN_NOT_FOUND:
1529 case KERN_NOT_WAITING:
1530 return ENOENT;
1531
1532 case KERN_RPC_SERVER_TERMINATED:
1533 return ECONNRESET;
1534
1535 case KERN_NOT_SUPPORTED:
1536 return ENOTSUP;
1537
1538 case KERN_NODE_DOWN:
1539 return ENETDOWN;
1540
1541 case KERN_OPERATION_TIMED_OUT:
1542 return ETIMEDOUT;
1543
1544 default:
1545 return EIO; /* 5 == KERN_FAILURE */
1546 }
1547 }
1548
1549 /*
1550 * Return the mach return value corresponding to a given BSD errno.
1551 */
1552 kern_return_t
kern_return_for_errno(int bsd_errno)1553 kern_return_for_errno(int bsd_errno)
1554 {
1555 switch (bsd_errno) {
1556 case 0:
1557 return KERN_SUCCESS;
1558 case EIO:
1559 case EACCES:
1560 case ENOMEM:
1561 case EFAULT:
1562 return KERN_MEMORY_ERROR;
1563
1564 case EINVAL:
1565 return KERN_INVALID_ARGUMENT;
1566
1567 case ETIMEDOUT:
1568 case EBUSY:
1569 return KERN_OPERATION_TIMED_OUT;
1570
1571 case ECONNRESET:
1572 return KERN_RPC_SERVER_TERMINATED;
1573
1574 case ENOTSUP:
1575 return KERN_NOT_SUPPORTED;
1576
1577 case ENETDOWN:
1578 return KERN_NODE_DOWN;
1579
1580 case ENOENT:
1581 return KERN_NOT_FOUND;
1582
1583 case EINTR:
1584 return KERN_ABORTED;
1585
1586 case EPERM:
1587 return KERN_DENIED;
1588
1589 case EEXIST:
1590 return KERN_ALREADY_IN_SET;
1591
1592 default:
1593 return KERN_FAILURE;
1594 }
1595 }
1596