Home
last modified time | relevance | path

Searched refs:TH_BUCKET_SCHED_MAX (Results 1 – 12 of 12) sorted by relevance

/xnu-12377.41.6/osfmk/kern/
H A Dsched_clutch.h160 bitmap_t scr_unbound_runnable_bitmap[BITMAP_LEN(TH_BUCKET_SCHED_MAX)];
162 bitmap_t scr_unbound_warp_available[BITMAP_LEN(TH_BUCKET_SCHED_MAX)];
164 bitmap_t scr_bound_runnable_bitmap[BITMAP_LEN(TH_BUCKET_SCHED_MAX)];
166 bitmap_t scr_bound_warp_available[BITMAP_LEN(TH_BUCKET_SCHED_MAX)];
174 uint16_t _Atomic scr_cumulative_run_count[TH_BUCKET_SCHED_MAX];
177 struct sched_clutch_root_bucket scr_unbound_buckets[TH_BUCKET_SCHED_MAX];
179 struct sched_clutch_root_bucket scr_bound_buckets[TH_BUCKET_SCHED_MAX];
323 struct sched_clutch_bucket_group sc_clutch_groups[TH_BUCKET_SCHED_MAX];
H A Dprocessor.h215 uint64_t pset_load_average[TH_BUCKET_SCHED_MAX];
220 uint32_t pset_runnable_depth[TH_BUCKET_SCHED_MAX];
307 pset_execution_time_t pset_execution_time[TH_BUCKET_SCHED_MAX];
309 _Atomic sched_clutch_edge sched_edges[MAX_PSETS][TH_BUCKET_SCHED_MAX];
310 sched_pset_search_order_t spill_search_order[TH_BUCKET_SCHED_MAX];
315 uint8_t max_parallel_cores[TH_BUCKET_SCHED_MAX];
316 uint8_t max_parallel_clusters[TH_BUCKET_SCHED_MAX];
H A Dsched_clutch.c84 static_assert(TH_BUCKET_SCHED_MAX == 6, "Ensure layout of sched_clutch_dbg_thread_select_packed can…
199 static uint32_t sched_clutch_root_bucket_wcel_us[TH_BUCKET_SCHED_MAX] = {
207 static uint64_t sched_clutch_root_bucket_wcel[TH_BUCKET_SCHED_MAX] = {0};
223 static uint32_t sched_clutch_root_bucket_warp_us[TH_BUCKET_SCHED_MAX] = {
231 static uint64_t sched_clutch_root_bucket_warp[TH_BUCKET_SCHED_MAX] = {0};
242 static uint32_t sched_clutch_thread_quantum_us[TH_BUCKET_SCHED_MAX] = {
251 static uint32_t sched_clutch_thread_quantum_us[TH_BUCKET_SCHED_MAX] = {
261 static uint64_t sched_clutch_thread_quantum[TH_BUCKET_SCHED_MAX] = {0};
271 for (int i = 0; i < TH_BUCKET_SCHED_MAX; i++) { in sched_clutch_us_to_abstime()
407 static uint32_t _Atomic sched_clutch_global_bucket_load[TH_BUCKET_SCHED_MAX];
[all …]
H A Dthread_group.c1315 #define MAX_EDGE_MATRIX_SIZE (MAX_PSETS * MAX_PSETS * TH_BUCKET_SCHED_MAX)
1325 for (sched_bucket_t bucket = 0; bucket < TH_BUCKET_SCHED_MAX; bucket++) { \
1328 …int sched_ind = (src * (int)num_psets * TH_BUCKET_SCHED_MAX) + (dst * TH_BUCKET_SCHED_MAX) + bucke…
1429 uint32_t expanded_index = (i * TH_BUCKET_SCHED_MAX) + TH_BUCKET_FIXPRI; in sched_perfcontrol_edge_matrix_get()
1438 uint32_t expanded_index = (i * TH_BUCKET_SCHED_MAX) + TH_BUCKET_FIXPRI; in sched_perfcontrol_edge_matrix_get()
1452 for (uint32_t bucket = 0; bucket < TH_BUCKET_SCHED_MAX; bucket++) { in sched_perfcontrol_edge_matrix_set()
1453 uint32_t expanded_index = (i * TH_BUCKET_SCHED_MAX) + bucket; in sched_perfcontrol_edge_matrix_set()
1467 uint32_t tg_bucket_preferred_cluster[TH_BUCKET_SCHED_MAX]; in sched_perfcontrol_thread_group_preferred_clusters_set()
1468 for (sched_bucket_t bucket = 0; bucket < TH_BUCKET_SCHED_MAX; bucket++) { in sched_perfcontrol_thread_group_preferred_clusters_set()
H A Dsched.h232 TH_BUCKET_SCHED_MAX = TH_BUCKET_RUN, /* Maximum schedulable buckets */ enumerator
H A Dprocessor.c538 …e(&processor->processor_set->cpu_running_buckets[processor->cpu_id], TH_BUCKET_SCHED_MAX, relaxed); in processor_state_update_idle()
556 …state & TH_IDLE) || (thread->bound_processor != PROCESSOR_NULL)) ? TH_BUCKET_SCHED_MAX : thread->t… in processor_state_update_from_thread()
H A Dsched_prim.c9089 bzero(running_higher, sizeof(uint32_t) * TH_BUCKET_SCHED_MAX); in sched_edge_pset_running_higher_bucket()
9095 if (cpu_bucket < TH_BUCKET_SCHED_MAX) { in sched_edge_pset_running_higher_bucket()
9101 for (sched_bucket_t bucket = TH_BUCKET_FIXPRI; bucket < TH_BUCKET_SCHED_MAX - 1; bucket++) { in sched_edge_pset_running_higher_bucket()
9155 uint32_t running_higher[TH_BUCKET_SCHED_MAX]; in sched_update_pset_load_average()
9158 …for (sched_bucket_t sched_bucket = TH_BUCKET_FIXPRI; sched_bucket < TH_BUCKET_SCHED_MAX; sched_buc… in sched_update_pset_load_average()
/xnu-12377.41.6/tests/sched/sched_test_harness/
H A Dsched_clutch_harness.h12 extern int root_bucket_to_highest_pri[TH_BUCKET_SCHED_MAX];
15 extern uint64_t clutch_root_bucket_wcel_us[TH_BUCKET_SCHED_MAX];
16 extern uint64_t clutch_root_bucket_warp_us[TH_BUCKET_SCHED_MAX];
H A Dsched_edge_harness.c272 …DLE) || (((thread_t)thread)->bound_processor != PROCESSOR_NULL)) ? TH_BUCKET_SCHED_MAX : ((thread_… in impl_cpu_set_thread_current()
296 …os_atomic_store(&cpus[cpu_id]->processor_set->cpu_running_buckets[cpu_id], TH_BUCKET_SCHED_MAX, re… in impl_cpu_clear_thread_current()
401 assert(sched_bucket < TH_BUCKET_SCHED_MAX); in impl_set_tg_sched_bucket_preferred_pset()
403 bitmap_t modify_bitmap[BITMAP_LEN(TH_BUCKET_SCHED_MAX)] = {0}; in impl_set_tg_sched_bucket_preferred_pset()
405 uint32_t tg_bucket_preferred_cluster[TH_BUCKET_SCHED_MAX] = {0}; in impl_set_tg_sched_bucket_preferred_pset()
413 assert(QoS > 0 && QoS < TH_BUCKET_SCHED_MAX); in impl_set_pset_load_avg()
H A Dsched_clutch_harness_impl.c73 int root_bucket_to_highest_pri[TH_BUCKET_SCHED_MAX] = {
83 uint64_t clutch_root_bucket_wcel_us[TH_BUCKET_SCHED_MAX];
84 uint64_t clutch_root_bucket_warp_us[TH_BUCKET_SCHED_MAX];
196 for (int b = TH_BUCKET_FIXPRI; b < TH_BUCKET_SCHED_MAX; b++) { in clutch_impl_init_params()
218 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) { in clutch_impl_create_tg()
/xnu-12377.41.6/tests/sched/
H A Dclutch_runqueue.c83 int num_threads = TH_BUCKET_SCHED_MAX - 1;
87 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {
110 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {
146 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {
163 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {
179 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {
200 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {
214 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {
252 #define WINDOW_MASK(bucket, cluster_bound) ( 1ULL << (bucket + cluster_bound * TH_BUCKET_SCHED_MAX)…
276 int num_threads = TH_BUCKET_SCHED_MAX - 1;
[all …]
/xnu-12377.41.6/doc/scheduler/
H A Dsched_clutch_edge.md31 static uint32_t sched_clutch_root_bucket_wcel_us[TH_BUCKET_SCHED_MAX] = {
44 static uint32_t sched_clutch_root_bucket_warp_us[TH_BUCKET_SCHED_MAX] = {
102 static uint32_t sched_clutch_thread_quantum_us[TH_BUCKET_SCHED_MAX] = {