mirror of
				https://sourceware.org/git/glibc.git
				synced 2025-11-03 20:53:13 +03:00 
			
		
		
		
	
		
			
				
	
	
		
			145 lines
		
	
	
		
			5.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			145 lines
		
	
	
		
			5.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/* Copyright (C) 2003-2014 Free Software Foundation, Inc.
 | 
						|
   This file is part of the GNU C Library.
 | 
						|
   Contributed by Jakub Jelinek <jakub@redhat.com>, 2003.
 | 
						|
 | 
						|
   The GNU C Library is free software; you can redistribute it and/or
 | 
						|
   modify it under the terms of the GNU Lesser General Public
 | 
						|
   License as published by the Free Software Foundation; either
 | 
						|
   version 2.1 of the License, or (at your option) any later version.
 | 
						|
 | 
						|
   The GNU C Library is distributed in the hope that it will be useful,
 | 
						|
   but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.	 See the GNU
 | 
						|
   Lesser General Public License for more details.
 | 
						|
 | 
						|
   You should have received a copy of the GNU Lesser General Public
 | 
						|
   License along with the GNU C Library; if not, see
 | 
						|
   <http://www.gnu.org/licenses/>.  */
 | 
						|
 | 
						|
#include "pthreadP.h"
 | 
						|
#include <lowlevellock.h>
 | 
						|
#include <atomic.h>
 | 
						|
 | 
						|
 | 
						|
unsigned long int __fork_generation attribute_hidden;
 | 
						|
 | 
						|
 | 
						|
static void
 | 
						|
clear_once_control (void *arg)
 | 
						|
{
 | 
						|
  pthread_once_t *once_control = (pthread_once_t *) arg;
 | 
						|
 | 
						|
  /* Reset to the uninitialized state here.  We don't need a stronger memory
 | 
						|
     order because we do not need to make any other of our writes visible to
 | 
						|
     other threads that see this value: This function will be called if we
 | 
						|
     get interrupted (see __pthread_once), so all we need to relay to other
 | 
						|
     threads is the state being reset again.  */
 | 
						|
  atomic_store_relaxed (once_control, 0);
 | 
						|
  lll_futex_wake (once_control, INT_MAX, LLL_PRIVATE);
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
/* This is similar to a lock implementation, but we distinguish between three
 | 
						|
   states: not yet initialized (0), initialization in progress
 | 
						|
   (__fork_generation | __PTHREAD_ONCE_INPROGRESS), and initialization
 | 
						|
   finished (__PTHREAD_ONCE_DONE); __fork_generation does not use the bits
 | 
						|
   that are used for __PTHREAD_ONCE_INPROGRESS and __PTHREAD_ONCE_DONE (which
 | 
						|
   is what __PTHREAD_ONCE_FORK_GEN_INCR is used for).  If in the first state,
 | 
						|
   threads will try to run the initialization by moving to the second state;
 | 
						|
   the first thread to do so via a CAS on once_control runs init_routine,
 | 
						|
   other threads block.
 | 
						|
   When forking the process, some threads can be interrupted during the second
 | 
						|
   state; they won't be present in the forked child, so we need to restart
 | 
						|
   initialization in the child.  To distinguish an in-progress initialization
 | 
						|
   from an interrupted initialization (in which case we need to reclaim the
 | 
						|
   lock), we look at the fork generation that's part of the second state: We
 | 
						|
   can reclaim iff it differs from the current fork generation.
 | 
						|
   XXX: This algorithm has an ABA issue on the fork generation: If an
 | 
						|
   initialization is interrupted, we then fork 2^30 times (30 bits of
 | 
						|
   once_control are used for the fork generation), and try to initialize
 | 
						|
   again, we can deadlock because we can't distinguish the in-progress and
 | 
						|
   interrupted cases anymore.
 | 
						|
   XXX: We split out this slow path because current compilers do not generate
 | 
						|
   as efficient code when the fast path in __pthread_once below is not in a
 | 
						|
   separate function.  */
 | 
						|
static int
 | 
						|
__attribute__ ((noinline))
 | 
						|
__pthread_once_slow (pthread_once_t *once_control, void (*init_routine) (void))
 | 
						|
{
 | 
						|
  while (1)
 | 
						|
    {
 | 
						|
      int val, newval;
 | 
						|
 | 
						|
      /* We need acquire memory order for this load because if the value
 | 
						|
         signals that initialization has finished, we need to see any
 | 
						|
         data modifications done during initialization.  */
 | 
						|
      val = atomic_load_acquire (once_control);
 | 
						|
      do
 | 
						|
	{
 | 
						|
	  /* Check if the initialization has already been done.  */
 | 
						|
	  if (__glibc_likely ((val & __PTHREAD_ONCE_DONE) != 0))
 | 
						|
	    return 0;
 | 
						|
 | 
						|
	  /* We try to set the state to in-progress and having the current
 | 
						|
	     fork generation.  We don't need atomic accesses for the fork
 | 
						|
	     generation because it's immutable in a particular process, and
 | 
						|
	     forked child processes start with a single thread that modified
 | 
						|
	     the generation.  */
 | 
						|
	  newval = __fork_generation | __PTHREAD_ONCE_INPROGRESS;
 | 
						|
	  /* We need acquire memory order here for the same reason as for the
 | 
						|
	     load from once_control above.  */
 | 
						|
	}
 | 
						|
      while (__glibc_unlikely (!atomic_compare_exchange_weak_acquire (
 | 
						|
	  once_control, &val, newval)));
 | 
						|
 | 
						|
      /* Check if another thread already runs the initializer.	*/
 | 
						|
      if ((val & __PTHREAD_ONCE_INPROGRESS) != 0)
 | 
						|
	{
 | 
						|
	  /* Check whether the initializer execution was interrupted by a
 | 
						|
	     fork.  We know that for both values, __PTHREAD_ONCE_INPROGRESS
 | 
						|
	     is set and __PTHREAD_ONCE_DONE is not.  */
 | 
						|
	  if (val == newval)
 | 
						|
	    {
 | 
						|
	      /* Same generation, some other thread was faster. Wait.  */
 | 
						|
	      lll_futex_wait (once_control, newval, LLL_PRIVATE);
 | 
						|
	      continue;
 | 
						|
	    }
 | 
						|
	}
 | 
						|
 | 
						|
      /* This thread is the first here.  Do the initialization.
 | 
						|
	 Register a cleanup handler so that in case the thread gets
 | 
						|
	 interrupted the initialization can be restarted.  */
 | 
						|
      pthread_cleanup_push (clear_once_control, once_control);
 | 
						|
 | 
						|
      init_routine ();
 | 
						|
 | 
						|
      pthread_cleanup_pop (0);
 | 
						|
 | 
						|
 | 
						|
      /* Mark *once_control as having finished the initialization.  We need
 | 
						|
         release memory order here because we need to synchronize with other
 | 
						|
         threads that want to use the initialized data.  */
 | 
						|
      atomic_store_release (once_control, __PTHREAD_ONCE_DONE);
 | 
						|
 | 
						|
      /* Wake up all other threads.  */
 | 
						|
      lll_futex_wake (once_control, INT_MAX, LLL_PRIVATE);
 | 
						|
      break;
 | 
						|
    }
 | 
						|
 | 
						|
  return 0;
 | 
						|
}
 | 
						|
 | 
						|
int
 | 
						|
__pthread_once (pthread_once_t *once_control, void (*init_routine) (void))
 | 
						|
{
 | 
						|
  /* Fast path.  See __pthread_once_slow.  */
 | 
						|
  int val;
 | 
						|
  val = atomic_load_acquire (once_control);
 | 
						|
  if (__glibc_likely ((val & __PTHREAD_ONCE_DONE) != 0))
 | 
						|
    return 0;
 | 
						|
  else
 | 
						|
    return __pthread_once_slow (once_control, init_routine);
 | 
						|
}
 | 
						|
weak_alias (__pthread_once, pthread_once)
 | 
						|
hidden_def (__pthread_once)
 |