1 /* Copyright (C) 2005-2023 Free Software Foundation, Inc.
2 Contributed by Richard Henderson <rth@redhat.com>.
4 This file is part of the GNU Offloading and Multi Processing Library
7 Libgomp is free software; you can redistribute it and/or modify it
8 under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3, or (at your option)
12 Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
14 FOR A PARTICULAR PURPOSE. See the GNU General Public License for
17 Under Section 7 of GPL version 3, you are granted additional
18 permissions described in the GCC Runtime Library Exception, version
19 3.1, as published by the Free Software Foundation.
21 You should have received a copy of the GNU General Public License and
22 a copy of the GCC Runtime Library Exception along with this program;
23 see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
24 <http://www.gnu.org/licenses/>. */
26 /* This file handles the (bare) PARALLEL construct. */
32 /* Determine the number of threads to be launched for a PARALLEL construct.
33 This algorithm is explicitly described in OpenMP 3.0 section 2.4.1.
34 SPECIFIED is a combination of the NUM_THREADS clause and the IF clause.
35 If the IF clause is false, SPECIFIED is forced to 1. When NUM_THREADS
36 is not present, SPECIFIED is 0. */
39 gomp_resolve_num_threads (unsigned specified
, unsigned count
)
41 struct gomp_thread
*thr
= gomp_thread ();
42 struct gomp_task_icv
*icv
;
43 unsigned threads_requested
, max_num_threads
, num_threads
;
45 struct gomp_thread_pool
*pool
;
47 icv
= gomp_icv (false);
52 if (thr
->ts
.active_level
>= 1
53 /* Accelerators with fixed thread counts require this to return 1 for
54 nested parallel regions. */
55 #if !defined(__AMDGCN__) && !defined(__nvptx__)
56 && icv
->max_active_levels_var
<= 1
60 else if (thr
->ts
.active_level
>= icv
->max_active_levels_var
)
63 /* If NUM_THREADS not specified, use nthreads_var. */
65 threads_requested
= icv
->nthreads_var
;
67 threads_requested
= specified
;
69 max_num_threads
= threads_requested
;
71 /* If dynamic threads are enabled, bound the number of threads
75 unsigned dyn
= gomp_dynamic_max_threads ();
76 if (dyn
< max_num_threads
)
77 max_num_threads
= dyn
;
79 /* Optimization for parallel sections. */
80 if (count
&& count
< max_num_threads
)
81 max_num_threads
= count
;
84 /* UINT_MAX stands for infinity. */
85 if (__builtin_expect (icv
->thread_limit_var
== UINT_MAX
, 1)
86 || max_num_threads
== 1)
87 return max_num_threads
;
89 /* The threads_busy counter lives in thread_pool, if there
90 isn't a thread_pool yet, there must be just one thread
91 in the contention group. If thr->team is NULL, this isn't
92 nested parallel, so there is just one thread in the
93 contention group as well, no need to handle it atomically. */
94 pool
= thr
->thread_pool
;
95 if (thr
->ts
.team
== NULL
|| pool
== NULL
)
97 num_threads
= max_num_threads
;
98 if (num_threads
> icv
->thread_limit_var
)
99 num_threads
= icv
->thread_limit_var
;
101 pool
->threads_busy
= num_threads
;
105 #ifdef HAVE_SYNC_BUILTINS
108 busy
= pool
->threads_busy
;
109 num_threads
= max_num_threads
;
110 if (icv
->thread_limit_var
- busy
+ 1 < num_threads
)
111 num_threads
= icv
->thread_limit_var
- busy
+ 1;
113 while (__sync_val_compare_and_swap (&pool
->threads_busy
,
114 busy
, busy
+ num_threads
- 1)
117 gomp_mutex_lock (&gomp_managed_threads_lock
);
118 num_threads
= max_num_threads
;
119 busy
= pool
->threads_busy
;
120 if (icv
->thread_limit_var
- busy
+ 1 < num_threads
)
121 num_threads
= icv
->thread_limit_var
- busy
+ 1;
122 pool
->threads_busy
+= num_threads
- 1;
123 gomp_mutex_unlock (&gomp_managed_threads_lock
);
130 GOMP_parallel_start (void (*fn
) (void *), void *data
, unsigned num_threads
)
132 num_threads
= gomp_resolve_num_threads (num_threads
, 0);
133 gomp_team_start (fn
, data
, num_threads
, 0, gomp_new_team (num_threads
),
138 GOMP_parallel_end (void)
140 struct gomp_task_icv
*icv
= gomp_icv (false);
141 if (__builtin_expect (icv
->thread_limit_var
!= UINT_MAX
, 0))
143 struct gomp_thread
*thr
= gomp_thread ();
144 struct gomp_team
*team
= thr
->ts
.team
;
145 unsigned int nthreads
= team
? team
->nthreads
: 1;
149 /* If not nested, there is just one thread in the
150 contention group left, no need for atomicity. */
151 if (thr
->ts
.team
== NULL
)
152 thr
->thread_pool
->threads_busy
= 1;
155 #ifdef HAVE_SYNC_BUILTINS
156 __sync_fetch_and_add (&thr
->thread_pool
->threads_busy
,
159 gomp_mutex_lock (&gomp_managed_threads_lock
);
160 thr
->thread_pool
->threads_busy
-= nthreads
- 1;
161 gomp_mutex_unlock (&gomp_managed_threads_lock
);
169 ialias (GOMP_parallel_end
)
172 GOMP_parallel (void (*fn
) (void *), void *data
, unsigned num_threads
,
175 num_threads
= gomp_resolve_num_threads (num_threads
, 0);
176 gomp_team_start (fn
, data
, num_threads
, flags
, gomp_new_team (num_threads
),
179 ialias_call (GOMP_parallel_end
) ();
183 GOMP_parallel_reductions (void (*fn
) (void *), void *data
,
184 unsigned num_threads
, unsigned int flags
)
186 struct gomp_taskgroup
*taskgroup
;
187 num_threads
= gomp_resolve_num_threads (num_threads
, 0);
188 uintptr_t *rdata
= *(uintptr_t **)data
;
189 taskgroup
= gomp_parallel_reduction_register (rdata
, num_threads
);
190 gomp_team_start (fn
, data
, num_threads
, flags
, gomp_new_team (num_threads
),
193 ialias_call (GOMP_parallel_end
) ();
194 gomp_sem_destroy (&taskgroup
->taskgroup_sem
);
200 GOMP_cancellation_point (int which
)
202 if (!gomp_cancel_var
)
205 struct gomp_thread
*thr
= gomp_thread ();
206 struct gomp_team
*team
= thr
->ts
.team
;
207 if (which
& (GOMP_CANCEL_LOOP
| GOMP_CANCEL_SECTIONS
))
211 return team
->work_share_cancelled
!= 0;
213 else if (which
& GOMP_CANCEL_TASKGROUP
)
215 if (thr
->task
->taskgroup
)
217 if (thr
->task
->taskgroup
->cancelled
)
219 if (thr
->task
->taskgroup
->workshare
220 && thr
->task
->taskgroup
->prev
221 && thr
->task
->taskgroup
->prev
->cancelled
)
224 /* FALLTHRU into the GOMP_CANCEL_PARALLEL case,
225 as #pragma omp cancel parallel also cancels all explicit
229 return gomp_team_barrier_cancelled (&team
->barrier
);
232 ialias (GOMP_cancellation_point
)
235 GOMP_cancel (int which
, bool do_cancel
)
237 if (!gomp_cancel_var
)
241 return ialias_call (GOMP_cancellation_point
) (which
);
243 struct gomp_thread
*thr
= gomp_thread ();
244 struct gomp_team
*team
= thr
->ts
.team
;
245 if (which
& (GOMP_CANCEL_LOOP
| GOMP_CANCEL_SECTIONS
))
247 /* In orphaned worksharing region, all we want to cancel
248 is current thread. */
250 team
->work_share_cancelled
= 1;
253 else if (which
& GOMP_CANCEL_TASKGROUP
)
255 if (thr
->task
->taskgroup
)
257 struct gomp_taskgroup
*taskgroup
= thr
->task
->taskgroup
;
258 if (taskgroup
->workshare
&& taskgroup
->prev
)
259 taskgroup
= taskgroup
->prev
;
260 if (!taskgroup
->cancelled
)
262 gomp_mutex_lock (&team
->task_lock
);
263 taskgroup
->cancelled
= true;
264 gomp_mutex_unlock (&team
->task_lock
);
269 team
->team_cancelled
= 1;
270 gomp_team_barrier_cancel (team
);
274 /* The public OpenMP API for thread and team related inquiries. */
277 omp_get_num_threads (void)
279 struct gomp_team
*team
= gomp_thread ()->ts
.team
;
280 return team
? team
->nthreads
: 1;
284 omp_get_thread_num (void)
286 return gomp_thread ()->ts
.team_id
;
289 /* This wasn't right for OpenMP 2.5. Active region used to be non-zero
290 when the IF clause doesn't evaluate to false, starting with OpenMP 3.0
291 it is non-zero with more than one thread in the team. */
294 omp_in_parallel (void)
296 return gomp_thread ()->ts
.active_level
> 0;
302 return gomp_thread ()->ts
.level
;
306 omp_get_ancestor_thread_num (int level
)
308 struct gomp_team_state
*ts
= &gomp_thread ()->ts
;
309 if (level
< 0 || level
> ts
->level
)
311 for (level
= ts
->level
- level
; level
> 0; --level
)
312 ts
= &ts
->team
->prev_ts
;
317 omp_get_team_size (int level
)
319 struct gomp_team_state
*ts
= &gomp_thread ()->ts
;
320 if (level
< 0 || level
> ts
->level
)
322 for (level
= ts
->level
- level
; level
> 0; --level
)
323 ts
= &ts
->team
->prev_ts
;
324 if (ts
->team
== NULL
)
327 return ts
->team
->nthreads
;
331 omp_get_active_level (void)
333 return gomp_thread ()->ts
.active_level
;
336 ialias (omp_get_num_threads
)
337 ialias (omp_get_thread_num
)
338 ialias (omp_in_parallel
)
339 ialias (omp_get_level
)
340 ialias (omp_get_ancestor_thread_num
)
341 ialias (omp_get_team_size
)
342 ialias (omp_get_active_level
)