mirror of
				https://sourceware.org/git/glibc.git
				synced 2025-10-30 10:45:40 +03:00 
			
		
		
		
	This uses atomic operations to access lock elision metadata that is accessed concurrently (ie, adapt_count fields). The size of the data is less than a word but accessed only with atomic loads and stores; therefore, we add support for shorter-size atomic load and stores too. * include/atomic.h (__atomic_check_size_ls): New. (atomic_load_relaxed, atomic_load_acquire, atomic_store_relaxed, atomic_store_release): Use it. * sysdeps/x86/elide.h (ACCESS_ONCE): Remove. (elision_adapt, ELIDE_LOCK): Use atomics. * sysdeps/unix/sysv/linux/x86/elision-lock.c (__lll_lock_elision): Use atomics and improve code comments. * sysdeps/unix/sysv/linux/x86/elision-trylock.c (__lll_trylock_elision): Likewise.
		
			
				
	
	
		
			108 lines
		
	
	
		
			3.4 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			108 lines
		
	
	
		
			3.4 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* elision-lock.c: Elided pthread mutex lock.
 | |
|    Copyright (C) 2011-2016 Free Software Foundation, Inc.
 | |
|    This file is part of the GNU C Library.
 | |
| 
 | |
|    The GNU C Library is free software; you can redistribute it and/or
 | |
|    modify it under the terms of the GNU Lesser General Public
 | |
|    License as published by the Free Software Foundation; either
 | |
|    version 2.1 of the License, or (at your option) any later version.
 | |
| 
 | |
|    The GNU C Library is distributed in the hope that it will be useful,
 | |
|    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
|    Lesser General Public License for more details.
 | |
| 
 | |
|    You should have received a copy of the GNU Lesser General Public
 | |
|    License along with the GNU C Library; if not, see
 | |
|    <http://www.gnu.org/licenses/>.  */
 | |
| 
 | |
| #include <pthread.h>
 | |
| #include "pthreadP.h"
 | |
| #include "lowlevellock.h"
 | |
| #include "hle.h"
 | |
| #include <elision-conf.h>
 | |
| 
 | |
| #if !defined(LLL_LOCK) && !defined(EXTRAARG)
 | |
| /* Make sure the configuration code is always linked in for static
 | |
|    libraries.  */
 | |
| #include "elision-conf.c"
 | |
| #endif
 | |
| 
 | |
| #ifndef EXTRAARG
 | |
| #define EXTRAARG
 | |
| #endif
 | |
| #ifndef LLL_LOCK
 | |
| #define LLL_LOCK(a,b) lll_lock(a,b), 0
 | |
| #endif
 | |
| 
 | |
| #define aconf __elision_aconf
 | |
| 
 | |
| /* Adaptive lock using transactions.
 | |
|    By default the lock region is run as a transaction, and when it
 | |
|    aborts or the lock is busy the lock adapts itself.  */
 | |
| 
 | |
| int
 | |
| __lll_lock_elision (int *futex, short *adapt_count, EXTRAARG int private)
 | |
| {
 | |
|   /* adapt_count can be accessed concurrently; these accesses can be both
 | |
|      inside of transactions (if critical sections are nested and the outer
 | |
|      critical section uses lock elision) and outside of transactions.  Thus,
 | |
|      we need to use atomic accesses to avoid data races.  However, the
 | |
|      value of adapt_count is just a hint, so relaxed MO accesses are
 | |
|      sufficient.  */
 | |
|   if (atomic_load_relaxed (adapt_count) <= 0)
 | |
|     {
 | |
|       unsigned status;
 | |
|       int try_xbegin;
 | |
| 
 | |
|       for (try_xbegin = aconf.retry_try_xbegin;
 | |
| 	   try_xbegin > 0;
 | |
| 	   try_xbegin--)
 | |
| 	{
 | |
| 	  if ((status = _xbegin()) == _XBEGIN_STARTED)
 | |
| 	    {
 | |
| 	      if (*futex == 0)
 | |
| 		return 0;
 | |
| 
 | |
| 	      /* Lock was busy.  Fall back to normal locking.
 | |
| 		 Could also _xend here but xabort with 0xff code
 | |
| 		 is more visible in the profiler.  */
 | |
| 	      _xabort (_ABORT_LOCK_BUSY);
 | |
| 	    }
 | |
| 
 | |
| 	  if (!(status & _XABORT_RETRY))
 | |
| 	    {
 | |
| 	      if ((status & _XABORT_EXPLICIT)
 | |
| 			&& _XABORT_CODE (status) == _ABORT_LOCK_BUSY)
 | |
| 	        {
 | |
| 		  /* Right now we skip here.  Better would be to wait a bit
 | |
| 		     and retry.  This likely needs some spinning.  See
 | |
| 		     above for why relaxed MO is sufficient.  */
 | |
| 		  if (atomic_load_relaxed (adapt_count)
 | |
| 		      != aconf.skip_lock_busy)
 | |
| 		    atomic_store_relaxed (adapt_count, aconf.skip_lock_busy);
 | |
| 		}
 | |
| 	      /* Internal abort.  There is no chance for retry.
 | |
| 		 Use the normal locking and next time use lock.
 | |
| 		 Be careful to avoid writing to the lock.  See above for why
 | |
| 		 relaxed MO is sufficient.  */
 | |
| 	      else if (atomic_load_relaxed (adapt_count)
 | |
| 		  != aconf.skip_lock_internal_abort)
 | |
| 		atomic_store_relaxed (adapt_count,
 | |
| 		    aconf.skip_lock_internal_abort);
 | |
| 	      break;
 | |
| 	    }
 | |
| 	}
 | |
|     }
 | |
|   else
 | |
|     {
 | |
|       /* Use a normal lock until the threshold counter runs out.
 | |
| 	 Lost updates possible.  */
 | |
|       atomic_store_relaxed (adapt_count,
 | |
| 	  atomic_load_relaxed (adapt_count) - 1);
 | |
|     }
 | |
| 
 | |
|   /* Use a normal lock here.  */
 | |
|   return LLL_LOCK ((*futex), private);
 | |
| }
 |