| /xnu-8019.80.24/tests/ |
| H A D | ntp_adjtime_29192647.c | 68 old_time_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; in get_abs_to_us_scale_factor() 74 new_time_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; in get_abs_to_us_scale_factor() 139 old_time_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; 166 new_time_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; 249 time1_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; 255 time2_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; 264 app = time_conv / USEC_PER_SEC; //sec elapsed 269 app = time_conv % USEC_PER_SEC; 271 time_delta += (app * (FREQ_PPM)) / USEC_PER_SEC; 324 time1_usec = (uint64_t)time.tv_sec * USEC_PER_SEC + (uint64_t)time.tv_nsec / 1000; [all …]
|
| H A D | kqueue_timer_tests.c | 49 timeout.tv_sec = (expected / USEC_PER_SEC) + 1; in do_simple_kevent() 50 timeout.tv_nsec = (expected % USEC_PER_SEC) * 1000; in do_simple_kevent() 68 elapsed_usecs = (after.tv_sec - before.tv_sec) * (int64_t)USEC_PER_SEC + in do_simple_kevent() 93 nowus = (uint64_t)tv.tv_sec * USEC_PER_SEC + (uint64_t)tv.tv_usec; in test_absolute_kevent() 103 timescale = USEC_PER_SEC; in test_absolute_kevent() 162 expected = (uint64_t)time * USEC_PER_SEC; in test_oneshot_kevent() 210 uint64_t test_duration_us = USEC_PER_SEC; /* 1 second */ in test_interval_kevent() 251 elapsed_usecs = (uint64_t)((after.tv_sec - before.tv_sec) * (int64_t)USEC_PER_SEC + in test_interval_kevent() 285 uint64_t test_duration_us = USEC_PER_SEC; /* 1 second */ in test_repeating_kevent()
|
| H A D | mach_boottime_usec.c | 21 T_EXPECT_EQ((uint64_t)bt_tv.tv_sec * USEC_PER_SEC + (uint64_t)bt_tv.tv_usec, bt_usec, NULL);
|
| H A D | kevent_continuous_time.c | 164 int64_t nowus = (int64_t)tv.tv_sec * USEC_PER_SEC + (int64_t)tv.tv_usec; 165 int64_t fire_at = (3 * USEC_PER_SEC) + nowus;
|
| H A D | aqm_qdelay_utun.c | 43 } else if (n >= USEC_PER_SEC) { in nsec_to_str() 44 t = n / USEC_PER_SEC; in nsec_to_str()
|
| H A D | sched_cluster_bound_threads.c | 136 …uint64_t bound_usr_usec = bound_thread_info.user_time.seconds * USEC_PER_SEC + bound_thread_info.u…
|
| H A D | disk_mount_conditioner.c | 318 usleep(USEC_PER_SEC / 2); // might still be other I/O inflight
|
| /xnu-8019.80.24/bsd/vfs/ |
| H A D | vfs_disk_conditioner.c | 142 …= (uint64_t)(total_size / ((double)(info->read_throughput_mbps * 1024 * 1024 / 8) / USEC_PER_SEC)); in disk_conditioner_delay() 144 … (uint64_t)(total_size / ((double)(info->write_throughput_mbps * 1024 * 1024 / 8) / USEC_PER_SEC)); in disk_conditioner_delay() 153 delay_usec += DISK_SPINUP_SEC * USEC_PER_SEC; in disk_conditioner_delay() 170 if (elapsed.tv_sec * USEC_PER_SEC < delay_usec) { in disk_conditioner_delay() 171 delay_usec -= elapsed.tv_sec * USEC_PER_SEC; in disk_conditioner_delay()
|
| /xnu-8019.80.24/osfmk/kern/ |
| H A D | clock.c | 238 *microsecs = ((uint64_t)USEC_PER_SEC * (uint32_t)(_bt->frac >> 32)) >> 32; in bintime2usclock() 781 TIME_SUB(deltasecs, oldsecs, deltamicrosecs, oldmicrosecs, USEC_PER_SEC); in clock_set_calendar_microtime() 783 TIME_ADD(clock_boottime, deltasecs, clock_boottime_usec, deltamicrosecs, USEC_PER_SEC); in clock_set_calendar_microtime() 791 TIME_SUB(deltasecs, secs, deltamicrosecs, microsecs, USEC_PER_SEC); in clock_set_calendar_microtime() 793 TIME_SUB(clock_boottime, deltasecs, clock_boottime_usec, deltamicrosecs, USEC_PER_SEC); in clock_set_calendar_microtime() 809 commpage_value = clock_boottime * USEC_PER_SEC + clock_boottime_usec; in clock_set_calendar_microtime() 1094 TIME_SUB(utc_offset_secs, sys, utc_offset_microsecs, microsys, USEC_PER_SEC); in clock_initialize_calendar() 1103 commpage_update_boottime(clock_boottime * USEC_PER_SEC + clock_boottime_usec); in clock_initialize_calendar() 1120 monotonic_sec = monotonic_usec_total / (clock_sec_t)USEC_PER_SEC; in clock_initialize_calendar() 1121 monotonic_usec = monotonic_usec_total % (clock_usec_t)USEC_PER_SEC; in clock_initialize_calendar() [all …]
|
| H A D | cpu_quiesce.c | 88 #define CPU_CHECKIN_MIN_INTERVAL_MAX_US USEC_PER_SEC /* 1s */
|
| H A D | kern_stackshot.c | 890 stackshot_microsecs = microsecs + (secs * USEC_PER_SEC); in kdp_snapshot_preflight() 1937 .user_usec = (uint64_t)user_time.seconds * USEC_PER_SEC + user_time.microseconds, in kcdata_record_thread_snapshot() 1938 .system_usec = (uint64_t)system_time.seconds * USEC_PER_SEC + system_time.microseconds, in kcdata_record_thread_snapshot() 1939 .runnable_usec = (uint64_t)runnable_time.seconds * USEC_PER_SEC + runnable_time.microseconds, in kcdata_record_thread_snapshot()
|
| H A D | thread.c | 616 if (delay_us > USEC_PER_SEC) { in thread_terminate_self() 646 if (delay_us > USEC_PER_SEC) { in thread_terminate_self()
|
| /xnu-8019.80.24/osfmk/mach/ |
| H A D | clock_types.h | 85 #define USEC_PER_SEC 1000000ull /* microseconds per second */ macro
|
| /xnu-8019.80.24/bsd/kern/ |
| H A D | kern_ntptime.c | 679 ltw = (int64_t)delta->tv_sec * (int64_t)USEC_PER_SEC + delta->tv_usec; in kern_adjtime() 692 atv.tv_sec = (__darwin_time_t)(ltr / (int64_t)USEC_PER_SEC); in kern_adjtime() 693 atv.tv_usec = ltr % (int64_t)USEC_PER_SEC; in kern_adjtime() 695 atv.tv_usec += (suseconds_t)USEC_PER_SEC; in kern_adjtime()
|
| /xnu-8019.80.24/osfmk/arm64/ |
| H A D | machine_routines.c | 608 if (mtxspin > USEC_PER_SEC >> 4) { in ml_init_lock_timeout() 609 mtxspin = USEC_PER_SEC >> 4; in ml_init_lock_timeout() 2142 } else if (events_per_sec > USEC_PER_SEC) { in wfe_timeout_configure() 2143 events_per_sec = USEC_PER_SEC; in wfe_timeout_configure() 2146 events_per_sec = USEC_PER_SEC; in wfe_timeout_configure()
|
| /xnu-8019.80.24/osfmk/i386/ |
| H A D | machine_routines.c | 919 if (mtxspin > USEC_PER_SEC >> 4) { in ml_init_lock_timeout() 920 mtxspin = USEC_PER_SEC >> 4; in ml_init_lock_timeout()
|
| /xnu-8019.80.24/osfmk/arm/ |
| H A D | rtclock.c | 143 rtclock_usec_divisor = divisor / USEC_PER_SEC; in timebase_callback()
|
| H A D | machine_routines.c | 169 if (mtxspin > USEC_PER_SEC >> 4) { in ml_init_lock_timeout() 170 mtxspin = USEC_PER_SEC >> 4; in ml_init_lock_timeout()
|
| /xnu-8019.80.24/bsd/skywalk/lib/ |
| H A D | cuckoo_hashtable_test.c | 639 struct timespec ts = { 0, 100 * USEC_PER_SEC }; in cht_concurrent_tests() 656 struct timespec ts = { 0, 100 * USEC_PER_SEC }; in cht_concurrent_tests()
|
| /xnu-8019.80.24/bsd/net/ |
| H A D | dlil.h | 103 *(nsp) += ((tvp)->tv_sec * USEC_PER_SEC); \
|
| /xnu-8019.80.24/bsd/net/pktsched/ |
| H A D | pktsched_netem.c | 814 tb->token += ((sec * USEC_PER_SEC + usec) * tb->rate / USEC_PER_SEC); in token_bucket_update()
|
| /xnu-8019.80.24/pexpert/arm/ |
| H A D | pe_identify_machine.c | 610 rPMGR_EVENT_TMR_PERIOD = gPEClockFrequencyInfo.timebase_frequency_hz / USEC_PER_SEC; in pe_arm_init_timer()
|
| /xnu-8019.80.24/osfmk/arm/commpage/ |
| H A D | commpage.c | 191 commpage_update_boottime(secs * USEC_PER_SEC + microsecs); in commpage_populate()
|
| /xnu-8019.80.24/osfmk/i386/commpage/ |
| H A D | commpage.c | 438 commpage_update_boottime(secs * USEC_PER_SEC + microsecs); in commpage_boottime_init()
|
| /xnu-8019.80.24/bsd/skywalk/mem/ |
| H A D | skmem_test.c | 1753 struct timespec ts = { .tv_sec = 0, .tv_nsec = 100 * USEC_PER_SEC }; in skmem_advanced_tests() 1769 struct timespec ts = { .tv_sec = 0, .tv_nsec = 100 * USEC_PER_SEC }; in skmem_advanced_tests()
|