Disable tests for strdup/strndup on __hpux__
[official-gcc.git] / libgomp / parallel.c
blob8ffcfc25a2bc3302f974f4ec03655af7c7457f10
1 /* Copyright (C) 2005-2024 Free Software Foundation, Inc.
2 Contributed by Richard Henderson <rth@redhat.com>.
4 This file is part of the GNU Offloading and Multi Processing Library
5 (libgomp).
7 Libgomp is free software; you can redistribute it and/or modify it
8 under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3, or (at your option)
10 any later version.
12 Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
14 FOR A PARTICULAR PURPOSE. See the GNU General Public License for
15 more details.
17 Under Section 7 of GPL version 3, you are granted additional
18 permissions described in the GCC Runtime Library Exception, version
19 3.1, as published by the Free Software Foundation.
21 You should have received a copy of the GNU General Public License and
22 a copy of the GCC Runtime Library Exception along with this program;
23 see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
24 <http://www.gnu.org/licenses/>. */
26 /* This file handles the (bare) PARALLEL construct. */
28 #include "libgomp.h"
29 #include <limits.h>
32 /* Determine the number of threads to be launched for a PARALLEL construct.
33 This algorithm is explicitly described in OpenMP 3.0 section 2.4.1.
34 SPECIFIED is a combination of the NUM_THREADS clause and the IF clause.
35 If the IF clause is false, SPECIFIED is forced to 1. When NUM_THREADS
36 is not present, SPECIFIED is 0. */
38 unsigned
39 gomp_resolve_num_threads (unsigned specified, unsigned count)
41 struct gomp_thread *thr = gomp_thread ();
42 struct gomp_task_icv *icv;
43 unsigned threads_requested, max_num_threads, num_threads;
44 unsigned long busy;
45 struct gomp_thread_pool *pool;
47 icv = gomp_icv (false);
49 if (specified == 1)
50 return 1;
52 if (thr->ts.active_level >= 1
53 /* Accelerators with fixed thread counts require this to return 1 for
54 nested parallel regions. */
55 #if !defined(__AMDGCN__) && !defined(__nvptx__)
56 && icv->max_active_levels_var <= 1
57 #endif
59 return 1;
60 else if (thr->ts.active_level >= icv->max_active_levels_var)
61 return 1;
63 /* If NUM_THREADS not specified, use nthreads_var. */
64 if (specified == 0)
65 threads_requested = icv->nthreads_var;
66 else
67 threads_requested = specified;
69 max_num_threads = threads_requested;
71 /* If dynamic threads are enabled, bound the number of threads
72 that we launch. */
73 if (icv->dyn_var)
75 unsigned dyn = gomp_dynamic_max_threads ();
76 if (dyn < max_num_threads)
77 max_num_threads = dyn;
79 /* Optimization for parallel sections. */
80 if (count && count < max_num_threads)
81 max_num_threads = count;
84 /* UINT_MAX stands for infinity. */
85 if (__builtin_expect (icv->thread_limit_var == UINT_MAX, 1)
86 || max_num_threads == 1)
87 return max_num_threads;
89 /* The threads_busy counter lives in thread_pool, if there
90 isn't a thread_pool yet, there must be just one thread
91 in the contention group. If thr->team is NULL, this isn't
92 nested parallel, so there is just one thread in the
93 contention group as well, no need to handle it atomically. */
94 pool = thr->thread_pool;
95 if (thr->ts.team == NULL || pool == NULL)
97 num_threads = max_num_threads;
98 if (num_threads > icv->thread_limit_var)
99 num_threads = icv->thread_limit_var;
100 if (pool)
101 pool->threads_busy = num_threads;
102 return num_threads;
105 #ifdef HAVE_SYNC_BUILTINS
108 busy = pool->threads_busy;
109 num_threads = max_num_threads;
110 if (icv->thread_limit_var - busy + 1 < num_threads)
111 num_threads = icv->thread_limit_var - busy + 1;
113 while (__sync_val_compare_and_swap (&pool->threads_busy,
114 busy, busy + num_threads - 1)
115 != busy);
116 #else
117 gomp_mutex_lock (&gomp_managed_threads_lock);
118 num_threads = max_num_threads;
119 busy = pool->threads_busy;
120 if (icv->thread_limit_var - busy + 1 < num_threads)
121 num_threads = icv->thread_limit_var - busy + 1;
122 pool->threads_busy += num_threads - 1;
123 gomp_mutex_unlock (&gomp_managed_threads_lock);
124 #endif
126 return num_threads;
129 void
130 GOMP_parallel_start (void (*fn) (void *), void *data, unsigned num_threads)
132 num_threads = gomp_resolve_num_threads (num_threads, 0);
133 gomp_team_start (fn, data, num_threads, 0, gomp_new_team (num_threads),
134 NULL);
137 void
138 GOMP_parallel_end (void)
140 struct gomp_task_icv *icv = gomp_icv (false);
141 if (__builtin_expect (icv->thread_limit_var != UINT_MAX, 0))
143 struct gomp_thread *thr = gomp_thread ();
144 struct gomp_team *team = thr->ts.team;
145 unsigned int nthreads = team ? team->nthreads : 1;
146 gomp_team_end ();
147 if (nthreads > 1)
149 /* If not nested, there is just one thread in the
150 contention group left, no need for atomicity. */
151 if (thr->ts.team == NULL)
152 thr->thread_pool->threads_busy = 1;
153 else
155 #ifdef HAVE_SYNC_BUILTINS
156 __sync_fetch_and_add (&thr->thread_pool->threads_busy,
157 1UL - nthreads);
158 #else
159 gomp_mutex_lock (&gomp_managed_threads_lock);
160 thr->thread_pool->threads_busy -= nthreads - 1;
161 gomp_mutex_unlock (&gomp_managed_threads_lock);
162 #endif
166 else
167 gomp_team_end ();
169 ialias (GOMP_parallel_end)
171 void
172 GOMP_parallel (void (*fn) (void *), void *data, unsigned num_threads,
173 unsigned int flags)
175 num_threads = gomp_resolve_num_threads (num_threads, 0);
176 gomp_team_start (fn, data, num_threads, flags, gomp_new_team (num_threads),
177 NULL);
178 fn (data);
179 ialias_call (GOMP_parallel_end) ();
182 unsigned
183 GOMP_parallel_reductions (void (*fn) (void *), void *data,
184 unsigned num_threads, unsigned int flags)
186 struct gomp_taskgroup *taskgroup;
187 num_threads = gomp_resolve_num_threads (num_threads, 0);
188 uintptr_t *rdata = *(uintptr_t **)data;
189 taskgroup = gomp_parallel_reduction_register (rdata, num_threads);
190 gomp_team_start (fn, data, num_threads, flags, gomp_new_team (num_threads),
191 taskgroup);
192 fn (data);
193 ialias_call (GOMP_parallel_end) ();
194 gomp_sem_destroy (&taskgroup->taskgroup_sem);
195 free (taskgroup);
196 return num_threads;
199 bool
200 GOMP_cancellation_point (int which)
202 if (!gomp_cancel_var)
203 return false;
205 struct gomp_thread *thr = gomp_thread ();
206 struct gomp_team *team = thr->ts.team;
207 if (which & (GOMP_CANCEL_LOOP | GOMP_CANCEL_SECTIONS))
209 if (team == NULL)
210 return false;
211 return team->work_share_cancelled != 0;
213 else if (which & GOMP_CANCEL_TASKGROUP)
215 if (thr->task->taskgroup)
217 if (thr->task->taskgroup->cancelled)
218 return true;
219 if (thr->task->taskgroup->workshare
220 && thr->task->taskgroup->prev
221 && thr->task->taskgroup->prev->cancelled)
222 return true;
224 /* FALLTHRU into the GOMP_CANCEL_PARALLEL case,
225 as #pragma omp cancel parallel also cancels all explicit
226 tasks. */
228 if (team)
229 return gomp_team_barrier_cancelled (&team->barrier);
230 return false;
232 ialias (GOMP_cancellation_point)
234 bool
235 GOMP_cancel (int which, bool do_cancel)
237 if (!gomp_cancel_var)
238 return false;
240 if (!do_cancel)
241 return ialias_call (GOMP_cancellation_point) (which);
243 struct gomp_thread *thr = gomp_thread ();
244 struct gomp_team *team = thr->ts.team;
245 if (which & (GOMP_CANCEL_LOOP | GOMP_CANCEL_SECTIONS))
247 /* In orphaned worksharing region, all we want to cancel
248 is current thread. */
249 if (team != NULL)
250 team->work_share_cancelled = 1;
251 return true;
253 else if (which & GOMP_CANCEL_TASKGROUP)
255 if (thr->task->taskgroup)
257 struct gomp_taskgroup *taskgroup = thr->task->taskgroup;
258 if (taskgroup->workshare && taskgroup->prev)
259 taskgroup = taskgroup->prev;
260 if (!taskgroup->cancelled)
262 gomp_mutex_lock (&team->task_lock);
263 taskgroup->cancelled = true;
264 gomp_mutex_unlock (&team->task_lock);
267 return true;
269 team->team_cancelled = 1;
270 gomp_team_barrier_cancel (team);
271 return true;
274 /* The public OpenMP API for thread and team related inquiries. */
277 omp_get_num_threads (void)
279 struct gomp_team *team = gomp_thread ()->ts.team;
280 return team ? team->nthreads : 1;
284 omp_get_thread_num (void)
286 return gomp_thread ()->ts.team_id;
289 /* This wasn't right for OpenMP 2.5. Active region used to be non-zero
290 when the IF clause doesn't evaluate to false, starting with OpenMP 3.0
291 it is non-zero with more than one thread in the team. */
294 omp_in_parallel (void)
296 return gomp_thread ()->ts.active_level > 0;
300 omp_get_level (void)
302 return gomp_thread ()->ts.level;
306 omp_get_ancestor_thread_num (int level)
308 struct gomp_team_state *ts = &gomp_thread ()->ts;
309 if (level < 0 || level > ts->level)
310 return -1;
311 for (level = ts->level - level; level > 0; --level)
312 ts = &ts->team->prev_ts;
313 return ts->team_id;
317 omp_get_team_size (int level)
319 struct gomp_team_state *ts = &gomp_thread ()->ts;
320 if (level < 0 || level > ts->level)
321 return -1;
322 for (level = ts->level - level; level > 0; --level)
323 ts = &ts->team->prev_ts;
324 if (ts->team == NULL)
325 return 1;
326 else
327 return ts->team->nthreads;
331 omp_get_active_level (void)
333 return gomp_thread ()->ts.active_level;
336 ialias (omp_get_num_threads)
337 ialias (omp_get_thread_num)
338 ialias (omp_in_parallel)
339 ialias (omp_get_level)
340 ialias (omp_get_ancestor_thread_num)
341 ialias (omp_get_team_size)
342 ialias (omp_get_active_level)