| /xnu-11417.101.15/tests/ |
| H A D | ntp_adjtime_29192647.c | 68 old_time_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; in get_abs_to_us_scale_factor() 74 new_time_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; in get_abs_to_us_scale_factor() 139 old_time_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; 166 new_time_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; 249 time1_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; 255 time2_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; 264 app = time_conv / USEC_PER_SEC; //sec elapsed 269 app = time_conv % USEC_PER_SEC; 271 time_delta += (app * (FREQ_PPM)) / USEC_PER_SEC; 324 time1_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; [all …]
|
| H A D | kqueue_timer_tests.c | 49 timeout.tv_sec = (expected / USEC_PER_SEC) + 1; in do_simple_kevent() 50 timeout.tv_nsec = (expected % USEC_PER_SEC) * 1000; in do_simple_kevent() 68 elapsed_usecs = (after.tv_sec - before.tv_sec) * (int64_t)USEC_PER_SEC + in do_simple_kevent() 93 nowus = (uint64_t)tv.tv_sec * USEC_PER_SEC + (uint64_t)tv.tv_usec; in test_absolute_kevent() 103 timescale = USEC_PER_SEC; in test_absolute_kevent() 162 expected = (uint64_t)time * USEC_PER_SEC; in test_oneshot_kevent() 210 uint64_t test_duration_us = USEC_PER_SEC; /* 1 second */ in test_interval_kevent() 251 elapsed_usecs = (uint64_t)((after.tv_sec - before.tv_sec) * (int64_t)USEC_PER_SEC + in test_interval_kevent() 285 uint64_t test_duration_us = USEC_PER_SEC; /* 1 second */ in test_repeating_kevent()
|
| H A D | mach_boottime_usec.c | 21 T_EXPECT_EQ((uint64_t)bt_tv.tv_sec * USEC_PER_SEC + (uint64_t)bt_tv.tv_usec, bt_usec, NULL);
|
| H A D | kevent_continuous_time.c | 164 int64_t nowus = (int64_t)tv.tv_sec * USEC_PER_SEC + (int64_t)tv.tv_usec; 165 int64_t fire_at = (3 * USEC_PER_SEC) + nowus;
|
| H A D | aqm_qdelay_utun.c | 42 } else if (n >= USEC_PER_SEC) { in nsec_to_str() 43 t = n / USEC_PER_SEC; in nsec_to_str()
|
| H A D | disk_mount_conditioner.c | 318 usleep(USEC_PER_SEC / 2); // might still be other I/O inflight
|
| /xnu-11417.101.15/tests/sched/ |
| H A D | cluster_bound_threads.c | 102 …uint64_t bound_usr_usec = (uint64_t)bound_thread_info.user_time.seconds * USEC_PER_SEC + (uint64_t… 120 before_user_us = (uint64_t)basic_thread_info.user_time.seconds * USEC_PER_SEC + in observe_thread_user_time() 127 after_user_us = (uint64_t)basic_thread_info.user_time.seconds * USEC_PER_SEC + in observe_thread_user_time() 170 …T_QUIET; T_EXPECT_GE(recommended_user_us * 1.0, runtime_threshold * observe_seconds * USEC_PER_SEC, 186 T_EXPECT_GE(recommended_user_us * 1.0, runtime_threshold * observe_seconds * USEC_PER_SEC,
|
| H A D | thread_group_fairness.c | 151 …uint64_t thread_usr_usec = (uint64_t) (info.user_time.seconds) * USEC_PER_SEC + (uint64_t) info.us… in snapshot_user_time_usec()
|
| H A D | yield_aggressor.c | 259 T_LOG("Elapsed Runtime: %f seconds", ((double) elapsed_usecs) / USEC_PER_SEC); in run_yielding_test()
|
| /xnu-11417.101.15/bsd/vfs/ |
| H A D | vfs_disk_conditioner.c | 142 …= (uint64_t)(total_size / ((double)(info->read_throughput_mbps * 1024 * 1024 / 8) / USEC_PER_SEC)); in disk_conditioner_delay() 144 … (uint64_t)(total_size / ((double)(info->write_throughput_mbps * 1024 * 1024 / 8) / USEC_PER_SEC)); in disk_conditioner_delay() 153 delay_usec += DISK_SPINUP_SEC * USEC_PER_SEC; in disk_conditioner_delay() 170 if (elapsed.tv_sec * USEC_PER_SEC < delay_usec) { in disk_conditioner_delay() 171 delay_usec -= elapsed.tv_sec * USEC_PER_SEC; in disk_conditioner_delay()
|
| /xnu-11417.101.15/osfmk/kern/ |
| H A D | clock.c | 239 *microsecs = ((uint64_t)USEC_PER_SEC * (uint32_t)(_bt->frac >> 32)) >> 32; in bintime2usclock() 785 TIME_SUB(deltasecs, oldsecs, deltamicrosecs, oldmicrosecs, USEC_PER_SEC); in clock_set_calendar_microtime() 787 TIME_ADD(clock_boottime, deltasecs, clock_boottime_usec, deltamicrosecs, USEC_PER_SEC); in clock_set_calendar_microtime() 795 TIME_SUB(deltasecs, secs, deltamicrosecs, microsecs, USEC_PER_SEC); in clock_set_calendar_microtime() 797 TIME_SUB(clock_boottime, deltasecs, clock_boottime_usec, deltamicrosecs, USEC_PER_SEC); in clock_set_calendar_microtime() 813 commpage_value = clock_boottime * USEC_PER_SEC + clock_boottime_usec; in clock_set_calendar_microtime() 1098 TIME_SUB(utc_offset_secs, sys, utc_offset_microsecs, microsys, USEC_PER_SEC); in clock_initialize_calendar() 1107 commpage_update_boottime(clock_boottime * USEC_PER_SEC + clock_boottime_usec); in clock_initialize_calendar() 1124 OS_ANALYZER_SUPPRESS("82347749") monotonic_sec = monotonic_usec_total / (clock_sec_t)USEC_PER_SEC; in clock_initialize_calendar() 1125 monotonic_usec = monotonic_usec_total % (clock_usec_t)USEC_PER_SEC; in clock_initialize_calendar() [all …]
|
| H A D | machine.c | 626 static const uint64_t TIMEBASE_TICKS_PER_USEC = 24000000ULL / USEC_PER_SEC;
|
| H A D | kern_stackshot.c | 2709 .sc_microsecs = microsecs + (secs * USEC_PER_SEC), in kdp_snapshot_preflight_internal() 4334 .user_usec = user_sec * USEC_PER_SEC + user_usec, in kcdata_record_thread_snapshot() 4335 .system_usec = system_sec * USEC_PER_SEC + system_usec, in kcdata_record_thread_snapshot() 4336 .runnable_usec = (uint64_t)runnable_time.seconds * USEC_PER_SEC + runnable_time.microseconds, in kcdata_record_thread_snapshot()
|
| /xnu-11417.101.15/osfmk/mach/ |
| H A D | clock_types.h | 85 #define USEC_PER_SEC 1000000ull /* microseconds per second */ macro
|
| /xnu-11417.101.15/bsd/kern/ |
| H A D | kern_ntptime.c | 682 ltw = (int64_t)delta->tv_sec * (int64_t)USEC_PER_SEC + delta->tv_usec; in kern_adjtime() 695 atv.tv_sec = (__darwin_time_t)(ltr / (int64_t)USEC_PER_SEC); in kern_adjtime() 696 atv.tv_usec = ltr % (int64_t)USEC_PER_SEC; in kern_adjtime() 698 atv.tv_usec += (suseconds_t)USEC_PER_SEC; in kern_adjtime()
|
| H A D | uipc_domain.c | 970 uint64_t microseconds = ((uint64_t)tvp->tv_sec * USEC_PER_SEC) + (uint64_t)tvp->tv_usec; in net_update_uptime_with_time()
|
| /xnu-11417.101.15/tests/sched/sched_test_harness/shadow_headers/ |
| H A D | sched_prim.c | 304 clock_interval_to_absolutetime_interval(USEC_PER_SEC >> SCHED_TICK_SHIFT, in sched_timeshare_timebase_init()
|
| /xnu-11417.101.15/osfmk/arm64/ |
| H A D | machine_routines.c | 707 if (mtxspin > USEC_PER_SEC >> 4) { in ml_init_lock_timeout() 708 mtxspin = USEC_PER_SEC >> 4; in ml_init_lock_timeout() 2629 } else if (events_per_sec > USEC_PER_SEC) { in wfe_timeout_configure() 2630 events_per_sec = USEC_PER_SEC; in wfe_timeout_configure() 2633 events_per_sec = USEC_PER_SEC; in wfe_timeout_configure()
|
| /xnu-11417.101.15/osfmk/arm/ |
| H A D | rtclock.c | 141 rtclock_usec_divisor = divisor / USEC_PER_SEC; in timebase_callback()
|
| /xnu-11417.101.15/bsd/skywalk/lib/ |
| H A D | cuckoo_hashtable_test.c | 635 struct timespec ts = { 0, 100 * USEC_PER_SEC }; in cht_concurrent_tests() 652 struct timespec ts = { 0, 100 * USEC_PER_SEC }; in cht_concurrent_tests()
|
| /xnu-11417.101.15/osfmk/i386/ |
| H A D | machine_routines.c | 1025 if (mtxspin > USEC_PER_SEC >> 4) { in ml_init_lock_timeout() 1026 mtxspin = USEC_PER_SEC >> 4; in ml_init_lock_timeout()
|
| /xnu-11417.101.15/bsd/net/ |
| H A D | dlil.h | 103 *(nsp) += ((tvp)->tv_sec * USEC_PER_SEC); \
|
| /xnu-11417.101.15/osfmk/arm/commpage/ |
| H A D | commpage.c | 166 commpage_update_boottime(secs * USEC_PER_SEC + microsecs); in commpage_populate()
|
| /xnu-11417.101.15/bsd/skywalk/mem/ |
| H A D | skmem_slab.c | 44 #define SKMEM_SLAB_MAX_BACKOFF (20 * USEC_PER_SEC) /* seconds */
|
| /xnu-11417.101.15/osfmk/i386/commpage/ |
| H A D | commpage.c | 435 commpage_update_boottime(secs * USEC_PER_SEC + microsecs); in commpage_boottime_init()
|