Searched refs:TH_BUCKET_SCHED_MAX (Results 1 – 12 of 12) sorted by relevance
133 struct priority_queue_sched_max sess_steal_queues[TH_BUCKET_SCHED_MAX];135 bitmap_t _Atomic sess_populated_steal_queues[BITMAP_LEN(TH_BUCKET_SCHED_MAX)];185 …bitmap_t _Atomic scr_incoming_migration_allowed[TH_BUCKET_SCHED_MAX][BITMAP_LEN(MAX…191 bitmap_t scr_unbound_runnable_bitmap[BITMAP_LEN(TH_BUCKET_SCHED_MAX)];193 bitmap_t scr_unbound_warp_available[BITMAP_LEN(TH_BUCKET_SCHED_MAX)];195 bitmap_t scr_bound_runnable_bitmap[BITMAP_LEN(TH_BUCKET_SCHED_MAX)];197 bitmap_t scr_bound_warp_available[BITMAP_LEN(TH_BUCKET_SCHED_MAX)];205 uint16_t _Atomic scr_cumulative_run_count[TH_BUCKET_SCHED_MAX];208 struct sched_clutch_root_bucket scr_unbound_buckets[TH_BUCKET_SCHED_MAX];210 struct sched_clutch_root_bucket scr_bound_buckets[TH_BUCKET_SCHED_MAX];[all …]
240 uint64_t pset_load_average[TH_BUCKET_SCHED_MAX];245 uint32_t pset_runnable_depth[TH_BUCKET_SCHED_MAX];332 pset_execution_time_t pset_execution_time[TH_BUCKET_SCHED_MAX];334 _Atomic sched_clutch_edge sched_edges[MAX_PSETS][TH_BUCKET_SCHED_MAX];335 sched_pset_search_order_t spill_search_order[TH_BUCKET_SCHED_MAX];340 uint8_t max_parallel_cores[TH_BUCKET_SCHED_MAX];341 uint8_t max_parallel_clusters[TH_BUCKET_SCHED_MAX];
84 static_assert(TH_BUCKET_SCHED_MAX == 6, "Ensure layout of sched_clutch_dbg_thread_select_packed can…191 static uint32_t sched_clutch_root_bucket_wcel_us[TH_BUCKET_SCHED_MAX] = {199 static uint64_t sched_clutch_root_bucket_wcel[TH_BUCKET_SCHED_MAX] = {0};215 static uint32_t sched_clutch_root_bucket_warp_us[TH_BUCKET_SCHED_MAX] = {223 static uint64_t sched_clutch_root_bucket_warp[TH_BUCKET_SCHED_MAX] = {0};234 static uint32_t sched_clutch_thread_quantum_us[TH_BUCKET_SCHED_MAX] = {243 static uint32_t sched_clutch_thread_quantum_us[TH_BUCKET_SCHED_MAX] = {253 static uint64_t sched_clutch_thread_quantum[TH_BUCKET_SCHED_MAX] = {0};263 for (int i = 0; i < TH_BUCKET_SCHED_MAX; i++) { in sched_clutch_us_to_abstime()399 static uint32_t _Atomic sched_clutch_global_bucket_load[TH_BUCKET_SCHED_MAX];[all …]
1315 #define MAX_EDGE_MATRIX_SIZE (MAX_PSETS * MAX_PSETS * TH_BUCKET_SCHED_MAX)1327 …int sched_ind = (src_id * (int)sched_num_psets * TH_BUCKET_SCHED_MAX) + (dst_id * TH_BUCKET_SCHED_…1426 uint32_t expanded_index = (i * TH_BUCKET_SCHED_MAX) + TH_BUCKET_FIXPRI; in sched_perfcontrol_edge_matrix_get()1435 uint32_t expanded_index = (i * TH_BUCKET_SCHED_MAX) + TH_BUCKET_FIXPRI; in sched_perfcontrol_edge_matrix_get()1449 for (uint32_t bucket = 0; bucket < TH_BUCKET_SCHED_MAX; bucket++) { in sched_perfcontrol_edge_matrix_set()1450 uint32_t expanded_index = (i * TH_BUCKET_SCHED_MAX) + bucket; in sched_perfcontrol_edge_matrix_set()1469 uint32_t tg_bucket_preferred_cluster[TH_BUCKET_SCHED_MAX]; in sched_perfcontrol_thread_group_preferred_clusters_set()1470 for (sched_bucket_t bucket = 0; bucket < TH_BUCKET_SCHED_MAX; bucket++) { in sched_perfcontrol_thread_group_preferred_clusters_set()
232 TH_BUCKET_SCHED_MAX = TH_BUCKET_RUN, /* Maximum schedulable buckets */ enumerator
574 …e(&processor->processor_set->cpu_running_buckets[processor->cpu_id], TH_BUCKET_SCHED_MAX, relaxed); in processor_state_update_idle()592 …state & TH_IDLE) || (thread->bound_processor != PROCESSOR_NULL)) ? TH_BUCKET_SCHED_MAX : thread->t… in processor_state_update_from_thread()
83 int num_threads = TH_BUCKET_SCHED_MAX - 1;87 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {110 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {146 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {163 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {179 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {200 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {214 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) {252 #define WINDOW_MASK(bucket, cluster_bound) ( 1ULL << (bucket + cluster_bound * TH_BUCKET_SCHED_MAX)…276 int num_threads = TH_BUCKET_SCHED_MAX - 1;[all …]
430 for (int b = 0; b < TH_BUCKET_SCHED_MAX; b++) { in clear_threads_from_topo()675 test_thread_t per_qos_threads[TH_BUCKET_SCHED_MAX];676 for (int bucket = 0; bucket < TH_BUCKET_SCHED_MAX; bucket++) {683 for (int bucket = 0; bucket < TH_BUCKET_SCHED_MAX; bucket++) {689 for (int qos_with_excess = p_pset_cpus; qos_with_excess < TH_BUCKET_SCHED_MAX; qos_with_excess++) {
12 extern int root_bucket_to_highest_pri[TH_BUCKET_SCHED_MAX];15 extern uint64_t clutch_root_bucket_wcel_us[TH_BUCKET_SCHED_MAX];16 extern uint64_t clutch_root_bucket_warp_us[TH_BUCKET_SCHED_MAX];
236 …os_atomic_store(&cpus[cpu_id]->processor_set->cpu_running_buckets[cpu_id], TH_BUCKET_SCHED_MAX, re… in impl_cpu_clear_thread_current()342 assert(sched_bucket < TH_BUCKET_SCHED_MAX); in impl_set_tg_sched_bucket_preferred_pset()344 bitmap_t modify_bitmap[BITMAP_LEN(TH_BUCKET_SCHED_MAX)] = {0}; in impl_set_tg_sched_bucket_preferred_pset()346 uint32_t tg_bucket_preferred_cluster[TH_BUCKET_SCHED_MAX] = {0}; in impl_set_tg_sched_bucket_preferred_pset()354 assert(QoS >= 0 && QoS < TH_BUCKET_SCHED_MAX); in impl_set_pset_load_avg()
72 int root_bucket_to_highest_pri[TH_BUCKET_SCHED_MAX] = {82 uint64_t clutch_root_bucket_wcel_us[TH_BUCKET_SCHED_MAX];83 uint64_t clutch_root_bucket_warp_us[TH_BUCKET_SCHED_MAX];198 for (int b = TH_BUCKET_FIXPRI; b < TH_BUCKET_SCHED_MAX; b++) { in clutch_impl_init_params()229 for (int bucket = TH_BUCKET_SHARE_FG; bucket < TH_BUCKET_SCHED_MAX; bucket++) { in clutch_impl_create_tg()
31 static uint32_t sched_clutch_root_bucket_wcel_us[TH_BUCKET_SCHED_MAX] = {44 static uint32_t sched_clutch_root_bucket_warp_us[TH_BUCKET_SCHED_MAX] = {102 static uint32_t sched_clutch_thread_quantum_us[TH_BUCKET_SCHED_MAX] = {