PR c++/61198
[official-gcc.git] / libobjc / objc-sync.c
blobf5169bfcf30fee0f3c64720cb552f465c2eda4ce
1 /* GNU Objective C Runtime @synchronized implementation
2 Copyright (C) 2010-2014 Free Software Foundation, Inc.
3 Contributed by Nicola Pero <nicola.pero@meta-innovation.com>
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it under the
8 terms of the GNU General Public License as published by the Free Software
9 Foundation; either version 3, or (at your option) any later version.
11 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
13 FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
14 details.
16 Under Section 7 of GPL version 3, you are granted additional
17 permissions described in the GCC Runtime Library Exception, version
18 3.1, as published by the Free Software Foundation.
20 You should have received a copy of the GNU General Public License and
21 a copy of the GCC Runtime Library Exception along with this program;
22 see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
23 <http://www.gnu.org/licenses/>. */
25 /* This file implements objc_sync_enter() and objc_sync_exit(), the
26 two functions required to support @synchronized().
28 objc_sync_enter(object) needs to get a recursive lock associated
29 with 'object', and lock it.
31 objc_sync_exit(object) needs to get the recursive lock associated
32 with 'object', and unlock it. */
34 /* To avoid the overhead of continuously allocating and deallocating
35 locks, we implement a pool of locks. When a lock is needed for an
36 object, we get a lock from the pool and associate it with the
37 object.
39 The lock pool need to be protected by its own lock (the
40 "protection" lock), which has to be locked then unlocked each time
41 objc_sync_enter() and objc_sync_exit() are called. To reduce the
42 contention on the protection lock, instead of a single pool with a
43 single (global) protection lock we use a number of smaller pools,
44 each with its own pool protection lock. To decide which lock pool
45 to use for each object, we compute a hash from the object pointer.
47 The implementation of each lock pool uses a linked list of all the
48 locks in the pool (both unlocked, and locked); this works in the
49 assumption that the number of locks concurrently required is very
50 low. In practice, it seems that you rarely see more than a few
51 locks ever concurrently required.
53 A standard case is a thread acquiring a lock recursively, over and
54 over again: for example when most methods of a class are protected
55 by @synchronized(self) but they also call each other. We use
56 thread-local storage to implement a cache and optimize this case.
57 The cache stores locks that the thread successfully acquired,
58 allowing objc_sync_enter() and objc_sync_exit() to locate a lock
59 which is already held by the current thread without having to use
60 any protection lock or synchronization mechanism. It can so detect
61 recursive locks/unlocks, and transform them into no-ops that
62 require no actual locking or synchronization mechanisms at all. */
64 /* You can disable the thread-local cache (most likely to benchmark
65 the code with and without it) by compiling with
66 -DSYNC_CACHE_DISABLE, or commenting out the following line. */
67 /* #define SYNC_CACHE_DISABLE */
69 /* If thread-local storage is not available, automatically disable the
70 cache. */
71 #ifndef HAVE_TLS
72 # define SYNC_CACHE_DISABLE
73 #endif
75 #include "objc-private/common.h"
76 #include "objc/objc-sync.h" /* For objc_sync_enter(), objc_sync_exit() */
77 #include "objc/runtime.h" /* For objc_malloc() */
78 #include "objc/thr.h" /* For objc_mutex_loc() and similar */
79 #include "objc-private/objc-sync.h" /* For __objc_sync_init() */
81 /* We have 32 pools of locks, each of them protected by its own
82 protection lock. It's tempting to increase this number to reduce
83 contention; but in our tests it is high enough. */
84 #define SYNC_NUMBER_OF_POOLS 32
86 /* Given an object, it determines which pool contains the associated
87 lock. */
88 #define SYNC_OBJECT_HASH(OBJECT) ((((size_t)OBJECT >> 8) ^ (size_t)OBJECT) & (SYNC_NUMBER_OF_POOLS - 1))
90 /* The locks protecting each pool. */
91 static objc_mutex_t sync_pool_protection_locks[SYNC_NUMBER_OF_POOLS];
93 /* The data structure (linked list) holding the locks. */
94 typedef struct lock_node
96 /* Pointer to next entry on the list. NULL indicates end of list.
97 You need to hold the appropriate sync_pool_protection_locks[N] to
98 read or write this variable. */
99 struct lock_node *next;
101 /* The (recursive) lock. Allocated when the node is created, and
102 always not-NULL, and unchangeable, after that. */
103 objc_mutex_t lock;
105 /* This is how many times the objc_mutex_lock() has been called on
106 the lock (it is 0 when the lock is unused). Used to track when
107 the lock is no longer associated with an object and can be reused
108 for another object. It records "real" locks, potentially (but
109 not necessarily) by multiple threads. You need to hold the
110 appropriate sync_pool_protection_locks[N] to read or write this
111 variable. */
112 unsigned int usage_count;
114 /* The object that the lock is associated with. This variable can
115 only be written when holding the sync_pool_protection_locks[N]
116 and when node->usage_count == 0, ie, the lock is not being used.
117 You can read this variable either when you hold the
118 sync_pool_protection_locks[N] or when you hold node->lock,
119 because in that case you know that node->usage_count can't get to
120 zero until you release the lock. It is valid to have usage_count
121 == 0 and object != nil; in that case, the lock is not currently
122 being used, but is still currently associated with the
123 object. */
124 id object;
126 /* This is a counter reserved for use by the thread currently
127 holding the lock. So, you need to hold node->lock to read or
128 write this variable. It is normally 0, and if the cache is not
129 being used, it is kept at 0 (even if recursive locks are being
130 done; in that case, no difference is made between recursive and
131 non-recursive locks: they all increase usage_count, and call
132 objc_mutex_lock()). When the cache is being used, a thread may
133 be able to find a lock that it already holds using the cache; in
134 that case, to perform additional locks/unlocks it can
135 increase/decrease the recursive_usage_count (which does not
136 require any synchronization with other threads, since it's
137 protected by the node->lock itself) instead of the usage_count
138 (which requires locking the pool protection lock). And it can
139 skip the call to objc_mutex_lock/unlock too. */
140 unsigned int recursive_usage_count;
141 } *lock_node_ptr;
144 /* The pools of locks. Each of them is a linked list of lock_nodes.
145 In the list we keep both unlocked and locked nodes. */
146 static lock_node_ptr sync_pool_array[SYNC_NUMBER_OF_POOLS];
148 #ifndef SYNC_CACHE_DISABLE
149 /* We store a cache of locks acquired by each thread in thread-local
150 storage. */
151 static __thread lock_node_ptr *lock_cache = NULL;
153 /* This is a conservative implementation that uses a static array of
154 fixed size as cache. Because the cache is an array that we scan
155 linearly, the bigger it is, the slower it gets. This does not
156 matter much at small sizes (eg, the overhead of checking 8 cache
157 slots instead of 4 is very small compared to the other overheads
158 involved such as function calls and lock/unlock operations), but at
159 large sizes it becomes important as obviously there is a size over
160 which using the cache backfires: the lookup is so slow that the
161 cache slows down the software instead of speeding it up. In
162 practice, it seems that most threads use a small number of
163 concurrent locks, so we have a conservative implementation with a
164 fixed-size cache of 8 locks which gives a very predictable
165 behaviour. If a thread locks lots of different locks, only the
166 first 8 get the speed benefits of the cache, but the cache remains
167 always small, fast and predictable.
169 SYNC_CACHE_SIZE is the size of the lock cache for each thread. */
170 #define SYNC_CACHE_SIZE 8
171 #endif /* SYNC_CACHE_DISABLE */
173 /* Called at startup by init.c. */
174 void
175 __objc_sync_init (void)
177 int i;
179 for (i = 0; i < SYNC_NUMBER_OF_POOLS; i++)
181 lock_node_ptr new_node;
183 /* Create a protection lock for each pool. */
184 sync_pool_protection_locks[i] = objc_mutex_allocate ();
186 /* Preallocate a lock per pool. */
187 new_node = objc_malloc (sizeof (struct lock_node));
188 new_node->lock = objc_mutex_allocate ();
189 new_node->object = nil;
190 new_node->usage_count = 0;
191 new_node->recursive_usage_count = 0;
192 new_node->next = NULL;
194 sync_pool_array[i] = new_node;
199 objc_sync_enter (id object)
201 #ifndef SYNC_CACHE_DISABLE
202 int free_cache_slot;
203 #endif
204 int hash;
205 lock_node_ptr node;
206 lock_node_ptr unused_node;
208 if (object == nil)
209 return OBJC_SYNC_SUCCESS;
211 #ifndef SYNC_CACHE_DISABLE
212 if (lock_cache == NULL)
214 /* Note that this calloc only happen only once per thread, the
215 very first time a thread does a objc_sync_enter(). */
216 lock_cache = objc_calloc (SYNC_CACHE_SIZE, sizeof (lock_node_ptr));
219 /* Check the cache to see if we have a record of having already
220 locked the lock corresponding to this object. While doing so,
221 keep track of the first free cache node in case we need it
222 later. */
223 node = NULL;
224 free_cache_slot = -1;
227 int i;
228 for (i = 0; i < SYNC_CACHE_SIZE; i++)
230 lock_node_ptr locked_node = lock_cache[i];
232 if (locked_node == NULL)
234 if (free_cache_slot == -1)
235 free_cache_slot = i;
237 else if (locked_node->object == object)
239 node = locked_node;
240 break;
245 if (node != NULL)
247 /* We found the lock. Increase recursive_usage_count, which is
248 protected by node->lock, which we already hold. */
249 node->recursive_usage_count++;
251 /* There is no need to actually lock anything, since we already
252 hold the lock. Correspondingly, objc_sync_exit() will just
253 decrease recursive_usage_count and do nothing to unlock. */
254 return OBJC_SYNC_SUCCESS;
256 #endif /* SYNC_CACHE_DISABLE */
258 /* The following is the standard lookup for the lock in the standard
259 pool lock. It requires a pool protection lock. */
260 hash = SYNC_OBJECT_HASH(object);
262 /* Search for an existing lock for 'object'. While searching, make
263 note of any unused lock if we find any. */
264 unused_node = NULL;
266 objc_mutex_lock (sync_pool_protection_locks[hash]);
268 node = sync_pool_array[hash];
270 while (node != NULL)
272 if (node->object == object)
274 /* We found the lock. */
275 node->usage_count++;
276 objc_mutex_unlock (sync_pool_protection_locks[hash]);
278 #ifndef SYNC_CACHE_DISABLE
279 /* Put it in the cache. */
280 if (free_cache_slot != -1)
281 lock_cache[free_cache_slot] = node;
282 #endif
284 /* Lock it. */
285 objc_mutex_lock (node->lock);
287 return OBJC_SYNC_SUCCESS;
290 if (unused_node == NULL && node->usage_count == 0)
292 /* We found the first unused node. Record it. */
293 unused_node = node;
296 node = node->next;
299 /* An existing lock for 'object' could not be found. */
300 if (unused_node != NULL)
302 /* But we found a unused lock; use it. */
303 unused_node->object = object;
304 unused_node->usage_count = 1;
305 unused_node->recursive_usage_count = 0;
306 objc_mutex_unlock (sync_pool_protection_locks[hash]);
308 #ifndef SYNC_CACHE_DISABLE
309 if (free_cache_slot != -1)
310 lock_cache[free_cache_slot] = unused_node;
311 #endif
313 objc_mutex_lock (unused_node->lock);
315 return OBJC_SYNC_SUCCESS;
317 else
319 /* There are no unused nodes; allocate a new node. */
320 lock_node_ptr new_node;
322 /* Create the node. */
323 new_node = objc_malloc (sizeof (struct lock_node));
324 new_node->lock = objc_mutex_allocate ();
325 new_node->object = object;
326 new_node->usage_count = 1;
327 new_node->recursive_usage_count = 0;
329 /* Attach it at the beginning of the pool. */
330 new_node->next = sync_pool_array[hash];
331 sync_pool_array[hash] = new_node;
332 objc_mutex_unlock (sync_pool_protection_locks[hash]);
334 #ifndef SYNC_CACHE_DISABLE
335 if (free_cache_slot != -1)
336 lock_cache[free_cache_slot] = new_node;
337 #endif
339 objc_mutex_lock (new_node->lock);
341 return OBJC_SYNC_SUCCESS;
346 objc_sync_exit (id object)
348 int hash;
349 lock_node_ptr node;
351 if (object == nil)
352 return OBJC_SYNC_SUCCESS;
354 #ifndef SYNC_CACHE_DISABLE
355 if (lock_cache != NULL)
357 int i;
359 /* Find the lock in the cache. */
360 node = NULL;
361 for (i = 0; i < SYNC_CACHE_SIZE; i++)
363 lock_node_ptr locked_node = lock_cache[i];
365 if (locked_node != NULL && locked_node->object == object)
367 node = locked_node;
368 break;
371 /* Note that, if a node was found in the cache, the variable i
372 now holds the index where it was found, which will be used to
373 remove it from the cache. */
374 if (node != NULL)
376 if (node->recursive_usage_count > 0)
378 node->recursive_usage_count--;
379 return OBJC_SYNC_SUCCESS;
381 else
383 /* We need to do a real unlock. */
384 hash = SYNC_OBJECT_HASH(object);
386 /* TODO: If we had atomic increase/decrease operations
387 with memory barriers, we could avoid the lock
388 here! */
389 objc_mutex_lock (sync_pool_protection_locks[hash]);
390 node->usage_count--;
391 /* Normally, we do not reset object to nil here. We'll
392 leave the lock associated with that object, at zero
393 usage count. This makes it slightly more efficient to
394 provide a lock for that object if (as likely)
395 requested again. If the object is deallocated, we
396 don't care. It will never match a new lock that is
397 requested, and the node will be reused at some point.
399 But, if garbage collection is enabled, leaving a
400 pointer to the object in memory might prevent the
401 object from being released. In that case, we remove
402 it (TODO: maybe we should avoid using the garbage
403 collector at all ? Nothing is ever deallocated in
404 this file). */
405 #if OBJC_WITH_GC
406 node->object = nil;
407 #endif
408 objc_mutex_unlock (sync_pool_protection_locks[hash]);
410 /* PS: Between objc_mutex_unlock
411 (sync_pool_protection_locks[hash]) and
412 objc_mutex_unlock (node->lock), the pool is unlocked
413 so other threads may allocate this same lock to
414 another object (!). This is not a problem, but it is
415 curious. */
416 objc_mutex_unlock (node->lock);
418 /* Remove the node from the cache. */
419 lock_cache[i] = NULL;
421 return OBJC_SYNC_SUCCESS;
425 #endif
427 /* The cache either wasn't there, or didn't work (eg, we overflowed
428 it at some point and stopped recording new locks in the cache).
429 Proceed with a full search of the lock pool. */
430 hash = SYNC_OBJECT_HASH(object);
432 objc_mutex_lock (sync_pool_protection_locks[hash]);
434 /* Search for an existing lock for 'object'. */
435 node = sync_pool_array[hash];
437 while (node != NULL)
439 if (node->object == object)
441 /* We found the lock. */
442 node->usage_count--;
443 objc_mutex_unlock (sync_pool_protection_locks[hash]);
445 objc_mutex_unlock (node->lock);
447 /* No need to remove the node from the cache, since it
448 wasn't found in the cache when we looked for it! */
449 return OBJC_SYNC_SUCCESS;
452 node = node->next;
455 objc_mutex_unlock (sync_pool_protection_locks[hash]);
457 /* A lock for 'object' to unlock could not be found (!!). */
458 return OBJC_SYNC_NOT_OWNING_THREAD_ERROR;