1 /* scheduler.h -*-C++-*-
3 *************************************************************************
6 * Copyright (C) 2009-2013, Intel Corporation
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
14 * * Redistributions of source code must retain the above copyright
15 * notice, this list of conditions and the following disclaimer.
16 * * Redistributions in binary form must reproduce the above copyright
17 * notice, this list of conditions and the following disclaimer in
18 * the documentation and/or other materials provided with the
20 * * Neither the name of Intel Corporation nor the names of its
21 * contributors may be used to endorse or promote products derived
22 * from this software without specific prior written permission.
25 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
26 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
27 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
28 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
29 * HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
30 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
31 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
32 * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
33 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
34 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY
35 * WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
36 * POSSIBILITY OF SUCH DAMAGE.
37 **************************************************************************/
42 * @brief scheduler.h declares routines for the Intel Cilk Plus scheduler,
43 * making it the heart of the Intel Cilk Plus implementation.
46 #ifndef INCLUDED_SCHEDULER_DOT_H
47 #define INCLUDED_SCHEDULER_DOT_H
49 #include <cilk/common.h>
50 #include <internal/abi.h>
52 #include "rts-common.h"
53 #include "full_frame.h"
54 #include "reducer_impl.h"
55 #include "global_state.h"
57 #ifdef CILK_RECORD_REPLAY
58 #include "record-replay.h"
61 __CILKRTS_BEGIN_EXTERN_C
65 * @brief Flag to disable parallel reductions.
67 * Set to 0 to allow parallel reductions.
69 #define DISABLE_PARALLEL_REDUCERS 0
72 * @brief Debugging level for parallel reductions.
74 * Print debugging messages and assertions for parallel reducers. 0 is
75 * no debugging. A higher value generates more output.
77 #define REDPAR_DEBUG 0
80 * @brief Lock the worker mutex to allow exclusive access to the
81 * values in the @c __cilkrts_worker and local_state structures.
83 * @pre @c w->l->do_not_steal must not be set. Essentially this
84 * condition asserts that the worker is not locked recursively.
86 * @param w The worker to lock.
89 void __cilkrts_worker_lock(__cilkrts_worker
*w
);
92 * @brief Unlock the worker mutex.
94 * @pre @c w->l->do_not_steal must be set. Essentially this condition
95 * asserts that the worker has been previously locked.
97 * @param w The worker to unlock.
100 void __cilkrts_worker_unlock(__cilkrts_worker
*w
);
103 * @brief Push the next full frame to be made active in this worker
104 * and increment its join counter.
106 * __cilkrts_push_next_frame and pop_next_frame work on a one-element queue.
107 * This queue is used to communicate across the runtime from the code that
108 * wants to activate a frame to the code that can actually begin execution
109 * on that frame. They are asymetrical in that push increments the join
110 * counter but pop does not decrement it. Rather, a single push/pop
111 * combination makes a frame active and increments its join counter once.
113 * @note A system worker may chose to push work onto a user worker if
114 * the work is the continuation from a sync which only the user worker
117 * @param w The worker which the frame is to be pushed onto.
118 * @param ff The full_frame which is to be continued by the worker.
121 void __cilkrts_push_next_frame(__cilkrts_worker
*w
,
125 * @brief Sync on this worker.
127 * If this worker is the last to reach the sync, execution may resume
128 * on this worker after the sync.
130 * If this worker is not the last spawned child to reach the sync,
131 * then execution is suspended and the worker will re-enter the
132 * scheduling loop, looking for work it can steal.
134 * This function will jump into the runtime to switch to the scheduling
135 * stack to implement most of its logic.
137 * @param w The worker which is executing the sync.
138 * @param sf The __cilkrts_stack_frame containing the sync.
141 NORETURN
__cilkrts_c_sync(__cilkrts_worker
*w
,
142 __cilkrts_stack_frame
*sf
);
145 * @brief Worker @c w completely promotes its own deque, simulating the case
146 * where the whole deque is stolen.
148 * We use this mechanism to force the allocation of new storage for
149 * reducers for race-detection purposes.
151 * This method is called from the reducer lookup logic when
152 * @c g->force_reduce is set.
154 * @warning Use of "force_reduce" is known to have bugs when run with
155 * more than 1 worker.
157 * @param w The worker which is to have all entries in its deque
158 * promoted to full frames.
161 void __cilkrts_promote_own_deque(__cilkrts_worker
*w
);
164 * Called when a spawned function attempts to return and
165 * __cilkrts_undo_detach() fails. This can happen for two reasons:
167 * @li If another worker is considering stealing our parent, it bumps the
168 * exception pointer while it did so, which will cause __cilkrts_undo_detach()
169 * to fail. If the other worker didn't complete the steal of our parent, we
170 * still may be able to return to it, either because the steal attempt failed,
171 * or we won the race for the tail pointer.
173 * @li If the function's parent has been stolen then we cannot return. Instead
174 * we'll longjmp into the runtime to switch onto the scheduling stack to
175 * execute do_return_from_spawn() and determine what to do. Either this
176 * worker is the last one to the sync, in which case we need to jump to the
177 * sync, or this worker is not the last one to the sync, in which case we'll
178 * abandon this work and jump to the scheduling loop to search for more work
181 * @param w The worker which attempting to return from a spawn to
183 * @param returning_sf The stack frame which is returning.
186 void __cilkrts_c_THE_exception_check(__cilkrts_worker
*w
,
187 __cilkrts_stack_frame
*returning_sf
);
190 * @brief Return an exception to an stolen parent.
192 * Used by the gcc implementation of exceptions to return an exception
195 * @param w The worker which attempting to return from a spawn with an
196 * exception to a stolen parent.
197 * @param returning_sf The stack frame which is returning.
200 NORETURN
__cilkrts_exception_from_spawn(__cilkrts_worker
*w
,
201 __cilkrts_stack_frame
*returning_sf
);
204 * @brief Used by the Windows implementations of exceptions to migrate an exception
207 * Call this function when an exception has been thrown and has to
208 * traverse across a steal. The exception has already been wrapped
209 * up, so all that remains is to longjmp() into the continuation,
210 * sync, and re-raise it.
212 * @param sf The __cilkrts_stack_frame for the frame that is attempting to
213 * return an exception to a stolen parent.
215 void __cilkrts_migrate_exception (__cilkrts_stack_frame
*sf
);
218 * @brief Return from a call, not a spawn, where this frame has ever
221 * @param w The worker that is returning from a frame which was ever stolen.
224 void __cilkrts_return(__cilkrts_worker
*w
);
227 * @brief Special return from the initial frame.
229 * This method will be called from @c __cilkrts_leave_frame if
230 * @c CILK_FRAME_LAST is set.
232 * This function will do the things necessary to cleanup, and unbind the
233 * thread from the Intel Cilk Plus runtime. If this is the last user
234 * worker unbinding from the runtime, all system worker threads will be
237 * @pre @c w must be the currently executing worker, and must be a user
240 * @param w The worker that's returning from the initial frame.
243 void __cilkrts_c_return_from_initial(__cilkrts_worker
*w
);
246 * @brief Used by exception handling code to pop an entry from the
249 * @param w Worker to pop the entry from
251 * @return __cilkrts_stack_frame of parent call
252 * @return NULL if the deque is empty
255 __cilkrts_stack_frame
*__cilkrts_pop_tail(__cilkrts_worker
*w
);
258 * @brief Modifies the worker's protected_tail to prevent frames from
261 * The Dekker protocol has been extended to only steal if head+1 is also
262 * less than protected_tail.
264 * @param w The worker to be modified.
265 * @param new_protected_tail The new setting for protected_tail, or NULL if the
266 * entire deque is to be protected
268 * @return Previous value of protected tail.
271 __cilkrts_stack_frame
*volatile *__cilkrts_disallow_stealing(
273 __cilkrts_stack_frame
*volatile *new_protected_tail
);
276 * @brief Restores the protected tail to a previous state, possibly
277 * allowing frames to be stolen.
279 * @param w The worker to be modified.
280 * @param saved_protected_tail A previous setting for protected_tail that is
284 void __cilkrts_restore_stealing(
286 __cilkrts_stack_frame
*volatile *saved_protected_tail
);
289 * @brief Initialize a @c __cilkrts_worker.
291 * @note The memory for the worker must have been allocated outside
294 * @param g The global_state_t.
295 * @param self The index into the global_state's array of workers for this
296 * worker, or -1 if this worker was allocated from the heap and cannot be
298 * @param w The worker to be initialized.
300 * @return The initialized __cilkrts_worker.
303 __cilkrts_worker
*make_worker(global_state_t
*g
,
305 __cilkrts_worker
*w
);
308 * @brief Free up any resources allocated for a worker.
310 * @note The memory for the @c __cilkrts_worker itself must be
311 * deallocated outside this call.
313 * @param w The worker to be destroyed.
316 void destroy_worker (__cilkrts_worker
*w
);
319 * @brief Initialize the runtime.
321 * If necessary, allocates and initializes the global state. If
322 * necessary, unsuspends the system workers.
324 * @param start Specifies whether the workers are to be unsuspended if
325 * they are suspended. Allows __cilkrts_init() to start up the runtime without
326 * releasing the system threads.
329 void __cilkrts_init_internal(int start
);
332 * @brief Part of the sequence to shutdown the runtime.
334 * Specifically, this call frees the @c global_state_t for the runtime.
336 * @param g The global_state_t.
339 void __cilkrts_deinit_internal(global_state_t
*g
);
342 * Obsolete. We no longer need to import or export reducer maps.
345 cilkred_map
*__cilkrts_xchg_reducer(
346 __cilkrts_worker
*w
, cilkred_map
*newmap
) cilk_nothrow
;
349 * @brief Called when a user thread is bound to the runtime.
351 * If this action increments the count of bound user threads from 0 to
352 * 1, the system worker threads are unsuspended.
354 * If this action increments the count of bound user threads from 0 to
355 * 1, the system worker threads are unsuspended.
357 * @pre Global lock must be held.
358 * @param g The runtime global state.
361 void __cilkrts_enter_cilk(global_state_t
*g
);
364 * @brief Called when a user thread is unbound from the runtime.
366 * If this action decrements the count of bound user threads to 0, the
367 * system worker threads are suspended.
370 * @pre Global lock must be held.
372 * @param g The runtime global state.
375 void __cilkrts_leave_cilk(global_state_t
*g
);
379 * @brief cilk_fiber_proc that runs the main scheduler loop on a
382 * @pre fiber's owner field should be set to the correct __cilkrts_worker
383 * @pre fiber must be a user worker.
385 * @param fiber The scheduling fiber object.
387 void scheduler_fiber_proc_for_user_worker(cilk_fiber
*fiber
);
391 * @brief Prints out Cilk runtime statistics.
393 * @param g The runtime global state.
395 * This method is useful only for debugging purposes. No guarantees
396 * are made as to the validity of this data. :)
399 void __cilkrts_dump_stats_to_stderr(global_state_t
*g
);
401 #ifdef CILK_RECORD_REPLAY
403 char * walk_pedigree_nodes(char *p
, const __cilkrts_pedigree
*pnode
);
406 * @brief Used by exception handling code to simulate the popping of
407 * an entry from the worker's deque.
409 * @param w Worker whose deque we want to check
411 * @return @c __cilkrts_stack_frame of parent call
412 * @return NULL if the deque is empty
415 __cilkrts_stack_frame
*simulate_pop_tail(__cilkrts_worker
*w
);
419 __CILKRTS_END_EXTERN_C
421 #endif // ! defined(INCLUDED_SCHEDULER_DOT_H)