/src/SockFuzzer/third_party/xnu/osfmk/kern/sched.h
Line | Count | Source (jump to first uncovered line) |
1 | | /* |
2 | | * Copyright (c) 2000-2019 Apple Inc. All rights reserved. |
3 | | * |
4 | | * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ |
5 | | * |
6 | | * This file contains Original Code and/or Modifications of Original Code |
7 | | * as defined in and that are subject to the Apple Public Source License |
8 | | * Version 2.0 (the 'License'). You may not use this file except in |
9 | | * compliance with the License. The rights granted to you under the License |
10 | | * may not be used to create, or enable the creation or redistribution of, |
11 | | * unlawful or unlicensed copies of an Apple operating system, or to |
12 | | * circumvent, violate, or enable the circumvention or violation of, any |
13 | | * terms of an Apple operating system software license agreement. |
14 | | * |
15 | | * Please obtain a copy of the License at |
16 | | * http://www.opensource.apple.com/apsl/ and read it before using this file. |
17 | | * |
18 | | * The Original Code and all software distributed under the License are |
19 | | * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER |
20 | | * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, |
21 | | * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, |
22 | | * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. |
23 | | * Please see the License for the specific language governing rights and |
24 | | * limitations under the License. |
25 | | * |
26 | | * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ |
27 | | */ |
28 | | /* |
29 | | * @OSF_COPYRIGHT@ |
30 | | */ |
31 | | /* |
32 | | * Mach Operating System |
33 | | * Copyright (c) 1991,1990,1989,1988,1987 Carnegie Mellon University |
34 | | * All Rights Reserved. |
35 | | * |
36 | | * Permission to use, copy, modify and distribute this software and its |
37 | | * documentation is hereby granted, provided that both the copyright |
38 | | * notice and this permission notice appear in all copies of the |
39 | | * software, derivative works or modified versions, and any portions |
40 | | * thereof, and that both notices appear in supporting documentation. |
41 | | * |
42 | | * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" |
43 | | * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR |
44 | | * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. |
45 | | * |
46 | | * Carnegie Mellon requests users of this software to return to |
47 | | * |
48 | | * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU |
49 | | * School of Computer Science |
50 | | * Carnegie Mellon University |
51 | | * Pittsburgh PA 15213-3890 |
52 | | * |
53 | | * any improvements or extensions that they make and grant Carnegie Mellon |
54 | | * the rights to redistribute these changes. |
55 | | */ |
56 | | /* |
57 | | */ |
58 | | /* |
59 | | * File: sched.h |
60 | | * Author: Avadis Tevanian, Jr. |
61 | | * Date: 1985 |
62 | | * |
63 | | * Header file for scheduler. |
64 | | * |
65 | | */ |
66 | | |
67 | | #ifndef _KERN_SCHED_H_ |
68 | | #define _KERN_SCHED_H_ |
69 | | |
70 | | #include <mach/policy.h> |
71 | | #include <kern/kern_types.h> |
72 | | #include <kern/smp.h> |
73 | | #include <kern/circle_queue.h> |
74 | | #include <kern/macro_help.h> |
75 | | #include <kern/timer_call.h> |
76 | | #include <kern/ast.h> |
77 | | #include <kern/bits.h> |
78 | | |
79 | 0 | #define NRQS_MAX (128) /* maximum number of priority levels */ |
80 | | |
81 | 0 | #define MAXPRI (NRQS_MAX-1) |
82 | | #define MINPRI 0 /* lowest legal priority schedulable */ |
83 | | #define IDLEPRI MINPRI /* idle thread priority */ |
84 | | #define NOPRI -1 |
85 | | |
86 | | /* |
87 | | * High-level priority assignments |
88 | | * |
89 | | ************************************************************************* |
90 | | * 127 Reserved (real-time) |
91 | | * A |
92 | | * + |
93 | | * (32 levels) |
94 | | * + |
95 | | * V |
96 | | * 96 Reserved (real-time) |
97 | | * 95 Kernel mode only |
98 | | * A |
99 | | * + |
100 | | * (16 levels) |
101 | | * + |
102 | | * V |
103 | | * 80 Kernel mode only |
104 | | * 79 System high priority |
105 | | * A |
106 | | * + |
107 | | * (16 levels) |
108 | | * + |
109 | | * V |
110 | | * 64 System high priority |
111 | | * 63 Elevated priorities |
112 | | * A |
113 | | * + |
114 | | * (12 levels) |
115 | | * + |
116 | | * V |
117 | | * 52 Elevated priorities |
118 | | * 51 Elevated priorities (incl. BSD +nice) |
119 | | * A |
120 | | * + |
121 | | * (20 levels) |
122 | | * + |
123 | | * V |
124 | | * 32 Elevated priorities (incl. BSD +nice) |
125 | | * 31 Default (default base for threads) |
126 | | * 30 Lowered priorities (incl. BSD -nice) |
127 | | * A |
128 | | * + |
129 | | * (20 levels) |
130 | | * + |
131 | | * V |
132 | | * 11 Lowered priorities (incl. BSD -nice) |
133 | | * 10 Lowered priorities (aged pri's) |
134 | | * A |
135 | | * + |
136 | | * (11 levels) |
137 | | * + |
138 | | * V |
139 | | * 0 Lowered priorities (aged pri's / idle) |
140 | | ************************************************************************* |
141 | | */ |
142 | | |
143 | | #define BASEPRI_RTQUEUES (BASEPRI_REALTIME + 1) /* 97 */ |
144 | 0 | #define BASEPRI_REALTIME (MAXPRI - (NRQS_MAX / 4) + 1) /* 96 */ |
145 | | |
146 | 0 | #define MAXPRI_KERNEL (BASEPRI_REALTIME - 1) /* 95 */ |
147 | | #define BASEPRI_PREEMPT_HIGH (BASEPRI_PREEMPT + 1) /* 93 */ |
148 | | #define BASEPRI_PREEMPT (MAXPRI_KERNEL - 3) /* 92 */ |
149 | | #define BASEPRI_VM (BASEPRI_PREEMPT - 1) /* 91 */ |
150 | | |
151 | | #define BASEPRI_KERNEL (MINPRI_KERNEL + 1) /* 81 */ |
152 | | #define MINPRI_KERNEL (MAXPRI_KERNEL - (NRQS_MAX / 8) + 1) /* 80 */ |
153 | | |
154 | | #define MAXPRI_RESERVED (MINPRI_KERNEL - 1) /* 79 */ |
155 | | #define BASEPRI_GRAPHICS (MAXPRI_RESERVED - 3) /* 76 */ |
156 | | #define MINPRI_RESERVED (MAXPRI_RESERVED - (NRQS_MAX / 8) + 1) /* 64 */ |
157 | | |
158 | | #define MAXPRI_USER (MINPRI_RESERVED - 1) /* 63 */ |
159 | | #define BASEPRI_CONTROL (BASEPRI_DEFAULT + 17) /* 48 */ |
160 | | #define BASEPRI_FOREGROUND (BASEPRI_DEFAULT + 16) /* 47 */ |
161 | | #define BASEPRI_BACKGROUND (BASEPRI_DEFAULT + 15) /* 46 */ |
162 | | #define BASEPRI_USER_INITIATED (BASEPRI_DEFAULT + 6) /* 37 */ |
163 | | #define BASEPRI_DEFAULT (MAXPRI_USER - (NRQS_MAX / 4)) /* 31 */ |
164 | | #define MAXPRI_SUPPRESSED (BASEPRI_DEFAULT - 3) /* 28 */ |
165 | | #define BASEPRI_UTILITY (BASEPRI_DEFAULT - 11) /* 20 */ |
166 | | #define MAXPRI_THROTTLE (MINPRI + 4) /* 4 */ |
167 | | #define MINPRI_USER MINPRI /* 0 */ |
168 | | |
169 | | #define DEPRESSPRI (MINPRI) /* depress priority */ |
170 | | |
171 | | #define MAXPRI_PROMOTE (MAXPRI_KERNEL) /* ceiling for mutex promotion */ |
172 | | #define MINPRI_RWLOCK (BASEPRI_BACKGROUND) /* floor when holding rwlock count */ |
173 | | #define MINPRI_EXEC (BASEPRI_DEFAULT) /* floor when in exec state */ |
174 | | #define MINPRI_WAITQ (BASEPRI_DEFAULT) /* floor when in waitq handover state */ |
175 | | |
176 | | #define NRQS (BASEPRI_REALTIME) /* Non-realtime levels for runqs */ |
177 | | |
178 | | /* Ensure that NRQS is large enough to represent all non-realtime threads; even promoted ones */ |
179 | | _Static_assert((NRQS == (MAXPRI_PROMOTE + 1)), "Runqueues are too small to hold all non-realtime threads"); |
180 | | |
181 | | /* Type used for thread->sched_mode and saved_mode */ |
182 | | typedef enum { |
183 | | TH_MODE_NONE = 0, /* unassigned, usually for saved_mode only */ |
184 | | TH_MODE_REALTIME, /* time constraints supplied */ |
185 | | TH_MODE_FIXED, /* use fixed priorities, no decay */ |
186 | | TH_MODE_TIMESHARE, /* use timesharing algorithm */ |
187 | | } sched_mode_t; |
188 | | |
189 | | /* |
190 | | * Since the clutch scheduler organizes threads based on the thread group |
191 | | * and the scheduling bucket, its important to not mix threads from multiple |
192 | | * priority bands into the same bucket. To achieve that, in the clutch bucket |
193 | | * world, there is a scheduling bucket per QoS effectively. |
194 | | */ |
195 | | |
196 | | /* Buckets used for load calculation */ |
197 | | typedef enum { |
198 | | TH_BUCKET_FIXPRI = 0, /* Fixed-priority */ |
199 | | TH_BUCKET_SHARE_FG, /* Timeshare thread above BASEPRI_DEFAULT */ |
200 | | #if CONFIG_SCHED_CLUTCH |
201 | | TH_BUCKET_SHARE_IN, /* Timeshare thread between BASEPRI_USER_INITIATED and BASEPRI_DEFAULT */ |
202 | | #endif /* CONFIG_SCHED_CLUTCH */ |
203 | | TH_BUCKET_SHARE_DF, /* Timeshare thread between BASEPRI_DEFAULT and BASEPRI_UTILITY */ |
204 | | TH_BUCKET_SHARE_UT, /* Timeshare thread between BASEPRI_UTILITY and MAXPRI_THROTTLE */ |
205 | | TH_BUCKET_SHARE_BG, /* Timeshare thread between MAXPRI_THROTTLE and MINPRI */ |
206 | | TH_BUCKET_RUN, /* All runnable threads */ |
207 | | TH_BUCKET_SCHED_MAX = TH_BUCKET_RUN, /* Maximum schedulable buckets */ |
208 | | TH_BUCKET_MAX, |
209 | | } sched_bucket_t; |
210 | | |
211 | | /* |
212 | | * Macro to check for invalid priorities. |
213 | | */ |
214 | | #define invalid_pri(pri) ((pri) < MINPRI || (pri) > MAXPRI) |
215 | | |
216 | | struct runq_stats { |
217 | | uint64_t count_sum; |
218 | | uint64_t last_change_timestamp; |
219 | | }; |
220 | | |
221 | | #if defined(CONFIG_SCHED_TIMESHARE_CORE) || defined(CONFIG_SCHED_PROTO) |
222 | | |
223 | | struct run_queue { |
224 | | int highq; /* highest runnable queue */ |
225 | | bitmap_t bitmap[BITMAP_LEN(NRQS)]; /* run queue bitmap array */ |
226 | | int count; /* # of threads total */ |
227 | | int urgency; /* level of preemption urgency */ |
228 | | circle_queue_head_t queues[NRQS]; /* one for each priority */ |
229 | | |
230 | | struct runq_stats runq_stats; |
231 | | }; |
232 | | |
233 | | inline static void |
234 | | rq_bitmap_set(bitmap_t *map, u_int n) |
235 | 0 | { |
236 | 0 | assert(n < NRQS); |
237 | 0 | bitmap_set(map, n); |
238 | 0 | } |
239 | | |
240 | | inline static void |
241 | | rq_bitmap_clear(bitmap_t *map, u_int n) |
242 | 0 | { |
243 | 0 | assert(n < NRQS); |
244 | 0 | bitmap_clear(map, n); |
245 | 0 | } |
246 | | |
247 | | #endif /* defined(CONFIG_SCHED_TIMESHARE_CORE) || defined(CONFIG_SCHED_PROTO) */ |
248 | | |
249 | | struct rt_queue { |
250 | | _Atomic int count; /* # of threads total */ |
251 | | queue_head_t queue; /* all runnable RT threads */ |
252 | | struct runq_stats runq_stats; |
253 | | }; |
254 | | typedef struct rt_queue *rt_queue_t; |
255 | | |
256 | | #if defined(CONFIG_SCHED_GRRR_CORE) |
257 | | |
258 | | /* |
259 | | * We map standard Mach priorities to an abstract scale that more properly |
260 | | * indicates how we want processor time allocated under contention. |
261 | | */ |
262 | | typedef uint8_t grrr_proportional_priority_t; |
263 | | typedef uint8_t grrr_group_index_t; |
264 | | |
265 | | #define NUM_GRRR_PROPORTIONAL_PRIORITIES 256 |
266 | | #define MAX_GRRR_PROPORTIONAL_PRIORITY ((grrr_proportional_priority_t)255) |
267 | | |
268 | | #if 0 |
269 | | #define NUM_GRRR_GROUPS 8 /* log(256) */ |
270 | | #endif |
271 | | |
272 | | #define NUM_GRRR_GROUPS 64 /* 256/4 */ |
273 | | |
274 | | struct grrr_group { |
275 | | queue_chain_t priority_order; /* next greatest weight group */ |
276 | | grrr_proportional_priority_t minpriority; |
277 | | grrr_group_index_t index; |
278 | | |
279 | | queue_head_t clients; |
280 | | int count; |
281 | | uint32_t weight; |
282 | | #if 0 |
283 | | uint32_t deferred_removal_weight; |
284 | | #endif |
285 | | uint32_t work; |
286 | | thread_t current_client; |
287 | | }; |
288 | | |
289 | | struct grrr_run_queue { |
290 | | int count; |
291 | | uint32_t last_rescale_tick; |
292 | | struct grrr_group groups[NUM_GRRR_GROUPS]; |
293 | | queue_head_t sorted_group_list; |
294 | | uint32_t weight; |
295 | | grrr_group_t current_group; |
296 | | |
297 | | struct runq_stats runq_stats; |
298 | | }; |
299 | | |
300 | | #endif /* defined(CONFIG_SCHED_GRRR_CORE) */ |
301 | | |
302 | | extern int rt_runq_count(processor_set_t); |
303 | | extern void rt_runq_count_incr(processor_set_t); |
304 | | extern void rt_runq_count_decr(processor_set_t); |
305 | | |
306 | | #if defined(CONFIG_SCHED_MULTIQ) |
307 | | sched_group_t sched_group_create(void); |
308 | | void sched_group_destroy(sched_group_t sched_group); |
309 | | #endif /* defined(CONFIG_SCHED_MULTIQ) */ |
310 | | |
311 | | |
312 | | |
313 | | /* |
314 | | * Scheduler routines. |
315 | | */ |
316 | | |
317 | | /* Handle quantum expiration for an executing thread */ |
318 | | extern void thread_quantum_expire( |
319 | | timer_call_param_t processor, |
320 | | timer_call_param_t thread); |
321 | | |
322 | | /* Context switch check for current processor */ |
323 | | extern ast_t csw_check( |
324 | | thread_t thread, |
325 | | processor_t processor, |
326 | | ast_t check_reason); |
327 | | |
328 | | /* Check for pending ASTs */ |
329 | | extern void ast_check(processor_t processor); |
330 | | |
331 | | extern void sched_update_generation_count(void); |
332 | | |
333 | | #if defined(CONFIG_SCHED_TIMESHARE_CORE) |
334 | | extern uint32_t std_quantum, min_std_quantum; |
335 | | extern uint32_t std_quantum_us; |
336 | | #endif /* CONFIG_SCHED_TIMESHARE_CORE */ |
337 | | |
338 | | extern uint32_t thread_depress_time; |
339 | | extern uint32_t default_timeshare_computation; |
340 | | extern uint32_t default_timeshare_constraint; |
341 | | |
342 | | extern uint32_t max_rt_quantum, min_rt_quantum; |
343 | | |
344 | | extern int default_preemption_rate; |
345 | | |
346 | | #if defined(CONFIG_SCHED_TIMESHARE_CORE) |
347 | | |
348 | | /* |
349 | | * Age usage at approximately (1 << SCHED_TICK_SHIFT) times per second |
350 | | * Aging may be deferred during periods where all processors are idle |
351 | | * and cumulatively applied during periods of activity. |
352 | | */ |
353 | | #define SCHED_TICK_SHIFT 3 |
354 | | #define SCHED_TICK_MAX_DELTA (8) |
355 | | |
356 | | extern unsigned sched_tick; |
357 | | extern uint32_t sched_tick_interval; |
358 | | |
359 | | #endif /* CONFIG_SCHED_TIMESHARE_CORE */ |
360 | | |
361 | | extern uint64_t sched_one_second_interval; |
362 | | |
363 | | /* Periodic computation of various averages */ |
364 | | extern void compute_sched_load(void); |
365 | | |
366 | | extern void compute_averages(uint64_t); |
367 | | |
368 | | extern void compute_averunnable( |
369 | | void *nrun); |
370 | | |
371 | | extern void compute_stack_target( |
372 | | void *arg); |
373 | | |
374 | | extern void compute_pageout_gc_throttle( |
375 | | void *arg); |
376 | | |
377 | | extern void compute_pmap_gc_throttle( |
378 | | void *arg); |
379 | | |
380 | | /* |
381 | | * Conversion factor from usage |
382 | | * to priority. |
383 | | */ |
384 | | #if defined(CONFIG_SCHED_TIMESHARE_CORE) |
385 | | |
386 | | #define MAX_LOAD (NRQS - 1) |
387 | | #define SCHED_PRI_SHIFT_MAX ((8 * sizeof(uint32_t)) - 1) |
388 | | extern uint32_t sched_pri_shifts[TH_BUCKET_MAX]; |
389 | | extern uint32_t sched_fixed_shift; |
390 | | extern int8_t sched_load_shifts[NRQS]; |
391 | | extern uint32_t sched_decay_usage_age_factor; |
392 | | void sched_timeshare_consider_maintenance(uint64_t ctime); |
393 | | #endif /* CONFIG_SCHED_TIMESHARE_CORE */ |
394 | | |
395 | | void sched_consider_recommended_cores(uint64_t ctime, thread_t thread); |
396 | | |
397 | | extern int32_t sched_poll_yield_shift; |
398 | | extern uint64_t sched_safe_duration; |
399 | | |
400 | | extern uint32_t sched_load_average, sched_mach_factor; |
401 | | |
402 | | extern uint32_t avenrun[3], mach_factor[3]; |
403 | | |
404 | | extern uint64_t max_unsafe_computation; |
405 | | extern uint64_t max_poll_computation; |
406 | | |
407 | | extern uint32_t sched_run_buckets[TH_BUCKET_MAX]; |
408 | | |
409 | | extern uint32_t sched_run_incr(thread_t thread); |
410 | | extern uint32_t sched_run_decr(thread_t thread); |
411 | | extern void sched_update_thread_bucket(thread_t thread); |
412 | | |
413 | | extern uint32_t sched_smt_run_incr(thread_t thread); |
414 | | extern uint32_t sched_smt_run_decr(thread_t thread); |
415 | | extern void sched_smt_update_thread_bucket(thread_t thread); |
416 | | |
417 | | #define SCHED_DECAY_TICKS 32 |
418 | | struct shift_data { |
419 | | int shift1; |
420 | | int shift2; |
421 | | }; |
422 | | |
423 | | /* |
424 | | * thread_timer_delta macro takes care of both thread timers. |
425 | | */ |
426 | | #define thread_timer_delta(thread, delta) \ |
427 | | MACRO_BEGIN \ |
428 | | (delta) = (typeof(delta))timer_delta(&(thread)->system_timer, \ |
429 | | &(thread)->system_timer_save); \ |
430 | | (delta) += (typeof(delta))timer_delta(&(thread)->user_timer, \ |
431 | | &(thread)->user_timer_save); \ |
432 | | MACRO_END |
433 | | |
434 | | extern bool system_is_SMT; |
435 | | |
436 | | #endif /* _KERN_SCHED_H_ */ |