Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 | /* * Copyright (c) 2000-2009 Apple Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * * This file contains Original Code and/or Modifications of Original Code * as defined in and that are subject to the Apple Public Source License * Version 2.0 (the 'License'). You may not use this file except in * compliance with the License. The rights granted to you under the License * may not be used to create, or enable the creation or redistribution of, * unlawful or unlicensed copies of an Apple operating system, or to * circumvent, violate, or enable the circumvention or violation of, any * terms of an Apple operating system software license agreement. * * Please obtain a copy of the License at * http://www.opensource.apple.com/apsl/ and read it before using this file. * * The Original Code and all software distributed under the License are * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. * Please see the License for the specific language governing rights and * limitations under the License. * * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* * @OSF_COPYRIGHT@ */ /* * Mach Operating System * Copyright (c) 1991,1990,1989 Carnegie Mellon University * All Rights Reserved. * * Permission to use, copy, modify and distribute this software and its * documentation is hereby granted, provided that both the copyright * notice and this permission notice appear in all copies of the * software, derivative works or modified versions, and any portions * thereof, and that both notices appear in supporting documentation. * * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. * * Carnegie Mellon requests users of this software to return to * * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU * School of Computer Science * Carnegie Mellon University * Pittsburgh PA 15213-3890 * * any improvements or extensions that they make and grant Carnegie Mellon * the rights to redistribute these changes. */ /* */ /* * processor.h: Processor and processor-related definitions. */ #ifndef _KERN_PROCESSOR_H_ #define _KERN_PROCESSOR_H_ #include <mach/boolean.h> #include <mach/kern_return.h> #include <kern/kern_types.h> #include <sys/cdefs.h> #ifdef MACH_KERNEL_PRIVATE #include <mach/mach_types.h> #include <kern/ast.h> #include <kern/cpu_number.h> #include <kern/smp.h> #include <kern/simple_lock.h> #include <kern/locks.h> #include <kern/queue.h> #include <kern/sched.h> #include <mach/sfi_class.h> #include <kern/processor_data.h> struct processor_set { queue_head_t active_queue; /* active processors */ queue_head_t idle_queue; /* idle processors */ queue_head_t idle_secondary_queue; /* idle secondary processors */ int online_processor_count; int cpu_set_low, cpu_set_hi; int cpu_set_count; #if __SMP__ decl_simple_lock_data(,sched_lock) /* lock for above */ #endif #if defined(CONFIG_SCHED_TRADITIONAL) || defined(CONFIG_SCHED_MULTIQ) struct run_queue pset_runq; /* runq for this processor set */ #endif #if defined(CONFIG_SCHED_TRADITIONAL) int pset_runq_bound_count; /* # of threads in runq bound to any processor in pset */ #endif /* CPUs that have been sent an unacknowledged remote AST for scheduling purposes */ uint64_t pending_AST_cpu_mask; #if defined(CONFIG_SCHED_DEFERRED_AST) /* * A seperate mask, for ASTs that we may be able to cancel. This is dependent on * some level of support for requesting an AST on a processor, and then quashing * that request later. * * The purpose of this field (and the associated codepaths) is to infer when we * no longer need a processor that is DISPATCHING to come up, and to prevent it * from coming out of IDLE if possible. This should serve to decrease the number * of spurious ASTs in the system, and let processors spend longer periods in * IDLE. */ uint64_t pending_deferred_AST_cpu_mask; #endif struct ipc_port * pset_self; /* port for operations */ struct ipc_port * pset_name_self; /* port for information */ processor_set_t pset_list; /* chain of associated psets */ pset_node_t node; }; extern struct processor_set pset0; struct pset_node { processor_set_t psets; /* list of associated psets */ pset_node_t nodes; /* list of associated subnodes */ pset_node_t node_list; /* chain of associated nodes */ pset_node_t parent; }; extern struct pset_node pset_node0; extern queue_head_t tasks, terminated_tasks, threads; /* Terminated tasks are ONLY for stackshot */ extern int tasks_count, terminated_tasks_count, threads_count; decl_lck_mtx_data(extern,tasks_threads_lock) struct processor { queue_chain_t processor_queue;/* idle/active queue link, * MUST remain the first element */ int state; /* See below */ boolean_t is_SMT; boolean_t is_recommended; struct thread *active_thread, /* thread running on processor */ *next_thread, /* next thread when dispatched */ *idle_thread; /* this processor's idle thread. */ processor_set_t processor_set; /* assigned set */ int current_pri; /* priority of current thread */ sched_mode_t current_thmode; /* sched mode of current thread */ sfi_class_id_t current_sfi_class; /* SFI class of current thread */ int cpu_id; /* platform numeric id */ timer_call_data_t quantum_timer; /* timer for quantum expiration */ uint64_t quantum_end; /* time when current quantum ends */ uint64_t last_dispatch; /* time of last dispatch */ uint64_t deadline; /* current deadline */ boolean_t first_timeslice; /* has the quantum expired since context switch */ #if defined(CONFIG_SCHED_TRADITIONAL) || defined(CONFIG_SCHED_MULTIQ) struct run_queue runq; /* runq for this processor */ #endif #if defined(CONFIG_SCHED_TRADITIONAL) int runq_bound_count; /* # of threads bound to this processor */ #endif #if defined(CONFIG_SCHED_GRRR) struct grrr_run_queue grrr_runq; /* Group Ratio Round-Robin runq */ #endif processor_t processor_primary; /* pointer to primary processor for * secondary SMT processors, or a pointer * to ourselves for primaries or non-SMT */ processor_t processor_secondary; struct ipc_port * processor_self; /* port for operations */ processor_t processor_list; /* all existing processors */ processor_data_t processor_data; /* per-processor data */ }; extern processor_t processor_list; extern unsigned int processor_count; decl_simple_lock_data(extern,processor_list_lock) extern uint32_t processor_avail_count; extern processor_t master_processor; extern boolean_t sched_stats_active; /* * Processor state is accessed by locking the scheduling lock * for the assigned processor set. * * -------------------- SHUTDOWN * / ^ ^ * _/ | \ * OFF_LINE ---> START ---> RUNNING ---> IDLE ---> DISPATCHING * \_________________^ ^ ^______/ / * \__________________/ * * Most of these state transitions are externally driven as a * a directive (for instance telling an IDLE processor to start * coming out of the idle state to run a thread). However these * are typically paired with a handshake by the processor itself * to indicate that it has completed a transition of indeterminate * length (for example, the DISPATCHING->RUNNING or START->RUNNING * transitions must occur on the processor itself). * * The boot processor has some special cases, and skips the START state, * since it has already bootstrapped and is ready to context switch threads. * * When a processor is in DISPATCHING or RUNNING state, the current_pri, * current_thmode, and deadline fields should be set, so that other * processors can evaluate if it is an appropriate candidate for preemption. */ #if defined(CONFIG_SCHED_DEFERRED_AST) /* * -------------------- SHUTDOWN * / ^ ^ * _/ | \ * OFF_LINE ---> START ---> RUNNING ---> IDLE ---> DISPATCHING * \_________________^ ^ ^______/ ^_____ / / * \__________________/ * * A DISPATCHING processor may be put back into IDLE, if another * processor determines that the target processor will have nothing to do * upon reaching the RUNNING state. This is racy, but if the target * responds and becomes RUNNING, it will not break the processor state * machine. * * This change allows us to cancel an outstanding signal/AST on a processor * (if such an operation is supported through hardware or software), and * push the processor back into the IDLE state as a power optimization. */ #endif #define PROCESSOR_OFF_LINE 0 /* Not available */ #define PROCESSOR_SHUTDOWN 1 /* Going off-line */ #define PROCESSOR_START 2 /* Being started */ /* 3 Formerly Inactive (unavailable) */ #define PROCESSOR_IDLE 4 /* Idle (available) */ #define PROCESSOR_DISPATCHING 5 /* Dispatching (idle -> active) */ #define PROCESSOR_RUNNING 6 /* Normal execution */ extern processor_t current_processor(void); /* Lock macros, always acquired and released with interrupts disabled (splsched()) */ #if __SMP__ #define pset_lock(p) simple_lock(&(p)->sched_lock) #define pset_unlock(p) simple_unlock(&(p)->sched_lock) #define pset_lock_init(p) simple_lock_init(&(p)->sched_lock, 0) #else #define pset_lock(p) do { (void)p; } while(0) #define pset_unlock(p) do { (void)p; } while(0) #define pset_lock_init(p) do { (void)p; } while(0) #endif extern void processor_bootstrap(void); extern void processor_init( processor_t processor, int cpu_id, processor_set_t processor_set); extern void processor_set_primary( processor_t processor, processor_t primary); extern kern_return_t processor_shutdown( processor_t processor); extern void processor_queue_shutdown( processor_t processor); extern processor_set_t processor_pset( processor_t processor); extern pset_node_t pset_node_root(void); extern processor_set_t pset_create( pset_node_t node); extern void pset_init( processor_set_t pset, pset_node_t node); extern kern_return_t processor_info_count( processor_flavor_t flavor, mach_msg_type_number_t *count); #define pset_deallocate(x) #define pset_reference(x) extern void machine_run_count( uint32_t count); extern processor_t machine_choose_processor( processor_set_t pset, processor_t processor); #define next_pset(p) (((p)->pset_list != PROCESSOR_SET_NULL)? (p)->pset_list: (p)->node->psets) #define PSET_THING_TASK 0 #define PSET_THING_THREAD 1 extern kern_return_t processor_set_things( processor_set_t pset, void **thing_list, mach_msg_type_number_t *count, int type); #else /* MACH_KERNEL_PRIVATE */ __BEGIN_DECLS extern void pset_deallocate( processor_set_t pset); extern void pset_reference( processor_set_t pset); __END_DECLS #endif /* MACH_KERNEL_PRIVATE */ #ifdef KERNEL_PRIVATE __BEGIN_DECLS extern processor_t cpu_to_processor(int cpu); __END_DECLS #endif /* KERNEL_PRIVATE */ #endif /* _KERN_PROCESSOR_H_ */ |