1 /* Copyright (C) 2005-2013 Free Software Foundation, Inc.
2 Contributed by Richard Henderson <rth@redhat.com>.
4 This file is part of the GNU OpenMP Library (libgomp).
6 Libgomp is free software; you can redistribute it and/or modify it
7 under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 3, or (at your option)
11 Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
13 FOR A PARTICULAR PURPOSE. See the GNU General Public License for
16 Under Section 7 of GPL version 3, you are granted additional
17 permissions described in the GCC Runtime Library Exception, version
18 3.1, as published by the Free Software Foundation.
20 You should have received a copy of the GNU General Public License and
21 a copy of the GCC Runtime Library Exception along with this program;
22 see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
23 <http://www.gnu.org/licenses/>. */
25 /* This file contains routines to manage the work-share queue for a team
34 /* Allocate a new work share structure, preferably from current team's
35 free gomp_work_share cache. */
37 static struct gomp_work_share
*
38 alloc_work_share (struct gomp_team
*team
)
40 struct gomp_work_share
*ws
;
43 /* This is called in a critical section. */
44 if (team
->work_share_list_alloc
!= NULL
)
46 ws
= team
->work_share_list_alloc
;
47 team
->work_share_list_alloc
= ws
->next_free
;
51 #ifdef HAVE_SYNC_BUILTINS
52 ws
= team
->work_share_list_free
;
53 /* We need atomic read from work_share_list_free,
54 as free_work_share can be called concurrently. */
55 __asm ("" : "+r" (ws
));
57 if (ws
&& ws
->next_free
)
59 struct gomp_work_share
*next
= ws
->next_free
;
61 team
->work_share_list_alloc
= next
->next_free
;
65 gomp_mutex_lock (&team
->work_share_list_free_lock
);
66 ws
= team
->work_share_list_free
;
69 team
->work_share_list_alloc
= ws
->next_free
;
70 team
->work_share_list_free
= NULL
;
71 gomp_mutex_unlock (&team
->work_share_list_free_lock
);
74 gomp_mutex_unlock (&team
->work_share_list_free_lock
);
77 team
->work_share_chunk
*= 2;
78 ws
= gomp_malloc (team
->work_share_chunk
* sizeof (struct gomp_work_share
));
79 ws
->next_alloc
= team
->work_shares
[0].next_alloc
;
80 team
->work_shares
[0].next_alloc
= ws
;
81 team
->work_share_list_alloc
= &ws
[1];
82 for (i
= 1; i
< team
->work_share_chunk
- 1; i
++)
83 ws
[i
].next_free
= &ws
[i
+ 1];
84 ws
[i
].next_free
= NULL
;
88 /* Initialize an already allocated struct gomp_work_share.
89 This shouldn't touch the next_alloc field. */
92 gomp_init_work_share (struct gomp_work_share
*ws
, bool ordered
,
95 gomp_mutex_init (&ws
->lock
);
96 if (__builtin_expect (ordered
, 0))
98 #define INLINE_ORDERED_TEAM_IDS_CNT \
99 ((sizeof (struct gomp_work_share) \
100 - offsetof (struct gomp_work_share, inline_ordered_team_ids)) \
101 / sizeof (((struct gomp_work_share *) 0)->inline_ordered_team_ids[0]))
103 if (nthreads
> INLINE_ORDERED_TEAM_IDS_CNT
)
105 = gomp_malloc (nthreads
* sizeof (*ws
->ordered_team_ids
));
107 ws
->ordered_team_ids
= ws
->inline_ordered_team_ids
;
108 memset (ws
->ordered_team_ids
, '\0',
109 nthreads
* sizeof (*ws
->ordered_team_ids
));
110 ws
->ordered_num_used
= 0;
111 ws
->ordered_owner
= -1;
115 ws
->ordered_team_ids
= NULL
;
116 gomp_ptrlock_init (&ws
->next_ws
, NULL
);
117 ws
->threads_completed
= 0;
120 /* Do any needed destruction of gomp_work_share fields before it
121 is put back into free gomp_work_share cache or freed. */
124 gomp_fini_work_share (struct gomp_work_share
*ws
)
126 gomp_mutex_destroy (&ws
->lock
);
127 if (ws
->ordered_team_ids
!= ws
->inline_ordered_team_ids
)
128 free (ws
->ordered_team_ids
);
129 gomp_ptrlock_destroy (&ws
->next_ws
);
132 /* Free a work share struct, if not orphaned, put it into current
133 team's free gomp_work_share cache. */
136 free_work_share (struct gomp_team
*team
, struct gomp_work_share
*ws
)
138 gomp_fini_work_share (ws
);
139 if (__builtin_expect (team
== NULL
, 0))
143 struct gomp_work_share
*next_ws
;
144 #ifdef HAVE_SYNC_BUILTINS
147 next_ws
= team
->work_share_list_free
;
148 ws
->next_free
= next_ws
;
150 while (!__sync_bool_compare_and_swap (&team
->work_share_list_free
,
153 gomp_mutex_lock (&team
->work_share_list_free_lock
);
154 next_ws
= team
->work_share_list_free
;
155 ws
->next_free
= next_ws
;
156 team
->work_share_list_free
= ws
;
157 gomp_mutex_unlock (&team
->work_share_list_free_lock
);
162 /* The current thread is ready to begin the next work sharing construct.
163 In all cases, thr->ts.work_share is updated to point to the new
164 structure. In all cases the work_share lock is locked. Return true
165 if this was the first thread to reach this point. */
168 gomp_work_share_start (bool ordered
)
170 struct gomp_thread
*thr
= gomp_thread ();
171 struct gomp_team
*team
= thr
->ts
.team
;
172 struct gomp_work_share
*ws
;
174 /* Work sharing constructs can be orphaned. */
177 ws
= gomp_malloc (sizeof (*ws
));
178 gomp_init_work_share (ws
, ordered
, 1);
179 thr
->ts
.work_share
= ws
;
183 ws
= thr
->ts
.work_share
;
184 thr
->ts
.last_work_share
= ws
;
185 ws
= gomp_ptrlock_get (&ws
->next_ws
);
188 /* This thread encountered a new ws first. */
189 struct gomp_work_share
*ws
= alloc_work_share (team
);
190 gomp_init_work_share (ws
, ordered
, team
->nthreads
);
191 thr
->ts
.work_share
= ws
;
196 thr
->ts
.work_share
= ws
;
201 /* The current thread is done with its current work sharing construct.
202 This version does imply a barrier at the end of the work-share. */
205 gomp_work_share_end (void)
207 struct gomp_thread
*thr
= gomp_thread ();
208 struct gomp_team
*team
= thr
->ts
.team
;
209 gomp_barrier_state_t bstate
;
211 /* Work sharing constructs can be orphaned. */
214 free_work_share (NULL
, thr
->ts
.work_share
);
215 thr
->ts
.work_share
= NULL
;
219 bstate
= gomp_barrier_wait_start (&team
->barrier
);
221 if (gomp_barrier_last_thread (bstate
))
223 if (__builtin_expect (thr
->ts
.last_work_share
!= NULL
, 1))
225 team
->work_shares_to_free
= thr
->ts
.work_share
;
226 free_work_share (team
, thr
->ts
.last_work_share
);
230 gomp_team_barrier_wait_end (&team
->barrier
, bstate
);
231 thr
->ts
.last_work_share
= NULL
;
234 /* The current thread is done with its current work sharing construct.
235 This version implies a cancellable barrier at the end of the work-share. */
238 gomp_work_share_end_cancel (void)
240 struct gomp_thread
*thr
= gomp_thread ();
241 struct gomp_team
*team
= thr
->ts
.team
;
242 gomp_barrier_state_t bstate
;
244 /* Cancellable work sharing constructs cannot be orphaned. */
245 bstate
= gomp_barrier_wait_cancel_start (&team
->barrier
);
247 if (gomp_barrier_last_thread (bstate
))
249 if (__builtin_expect (thr
->ts
.last_work_share
!= NULL
, 1))
251 team
->work_shares_to_free
= thr
->ts
.work_share
;
252 free_work_share (team
, thr
->ts
.last_work_share
);
255 thr
->ts
.last_work_share
= NULL
;
257 return gomp_team_barrier_wait_cancel_end (&team
->barrier
, bstate
);
260 /* The current thread is done with its current work sharing construct.
261 This version does NOT imply a barrier at the end of the work-share. */
264 gomp_work_share_end_nowait (void)
266 struct gomp_thread
*thr
= gomp_thread ();
267 struct gomp_team
*team
= thr
->ts
.team
;
268 struct gomp_work_share
*ws
= thr
->ts
.work_share
;
271 /* Work sharing constructs can be orphaned. */
274 free_work_share (NULL
, ws
);
275 thr
->ts
.work_share
= NULL
;
279 if (__builtin_expect (thr
->ts
.last_work_share
== NULL
, 0))
282 #ifdef HAVE_SYNC_BUILTINS
283 completed
= __sync_add_and_fetch (&ws
->threads_completed
, 1);
285 gomp_mutex_lock (&ws
->lock
);
286 completed
= ++ws
->threads_completed
;
287 gomp_mutex_unlock (&ws
->lock
);
290 if (completed
== team
->nthreads
)
292 team
->work_shares_to_free
= thr
->ts
.work_share
;
293 free_work_share (team
, thr
->ts
.last_work_share
);
295 thr
->ts
.last_work_share
= NULL
;