1
0
mirror of https://sourceware.org/git/glibc.git synced 2025-07-30 22:43:12 +03:00
2004-06-03  Ulrich Drepper  <drepper@redhat.com>

	* sysdeps/i386/i486/bits/atomic.h: Optimize a bit.
This commit is contained in:
Ulrich Drepper
2004-06-03 16:04:11 +00:00
parent 322861e8b6
commit 75fcceded2
30 changed files with 327 additions and 209 deletions

View File

@ -1,3 +1,7 @@
2004-06-03 Ulrich Drepper <drepper@redhat.com>
* sysdeps/i386/i486/bits/atomic.h: Optimize a bit.
2004-05-07 Dmitry V. Levin <ldv@altlinux.org> 2004-05-07 Dmitry V. Levin <ldv@altlinux.org>
* argp/argp-help.c (__argp_error, __argp_failure): Check result * argp/argp-help.c (__argp_error, __argp_failure): Check result

View File

@ -1,3 +1,95 @@
2004-05-21 Jakub Jelinek <jakub@redhat.com>
* Makefile (tests): Add tst-cond16.
* sysdeps/unix/sysv/linux/lowlevelcond.sym (cond_futex): Add.
* pthread_cond_init.c (__pthread_cond_init): Clear __data.__futex.
* sysdeps/unix/sysv/linux/i386/bits/pthreadtypes.h (pthread_cond_t):
Add __data.__futex field, reshuffle __data.__clock.
* sysdeps/unix/sysv/linux/i386/pthread_cond_signal.S
(__pthread_cond_signal): Increment __futex at the same time as
__wakeup_seq or __total_seq. Pass address of __futex instead of
address of low 32-bits of __wakeup_seq to futex syscall.
* sysdeps/unix/sysv/linux/i386/pthread_cond_wait.S
(__pthread_cond_wait): Likewise. Pass __futex value from before
releasing internal lock to FUTEX_WAIT.
* sysdeps/unix/sysv/linux/i386/pthread_cond_timedwait.S
(__pthread_cond_timedwait): Likewise.
* sysdeps/unix/sysv/linux/i386/pthread_cond_broadcast.S
(FUTEX_CMP_REQUEUE): Define.
(__pthread_cond_broadcast): Set __futex to 2 * __total_seq.
Use FUTEX_CMP_REQUEUE operation instead of FUTEX_REQUEUE.
Pass __futex value from before the unlock and __futex address instead
of address of low 32-bits of __wakeup_seq to futex syscall.
Fallback to FUTEX_WAKE all on any errors.
* sysdeps/unix/sysv/linux/alpha/lowlevellock.h (FUTEX_CMP_REQUEUE):
Define.
(lll_futex_requeue): Add val argument, use FUTEX_CMP_REQUEUE
internally. Return non-zero if error, zero if success.
* sysdeps/unix/sysv/linux/alpha/bits/pthreadtypes.h (pthread_cond_t):
Add __data.__futex field, reshuffle __data.__clock.
* sysdeps/unix/sysv/linux/s390/lowlevellock.h (FUTEX_CMP_REQUEUE):
Define.
(lll_futex_requeue): Add val argument, return 1 unconditionally
for the time being.
* sysdeps/unix/sysv/linux/s390/bits/pthreadtypes.h (pthread_cond_t):
Add __data.__futex field, reshuffle __data.__clock.
* sysdeps/unix/sysv/linux/powerpc/lowlevellock.h (FUTEX_CMP_REQUEUE):
Define.
(lll_futex_requeue): Add val argument, use FUTEX_CMP_REQUEUE
internally. Return non-zero if error, zero if success.
* sysdeps/unix/sysv/linux/powerpc/bits/pthreadtypes.h
(pthread_cond_t): Add __data.__futex field, reshuffle __data.__clock.
* sysdeps/unix/sysv/linux/sparc/lowlevellock.h (FUTEX_CMP_REQUEUE):
Define.
(lll_futex_requeue): Add val argument, use FUTEX_CMP_REQUEUE
internally. Return non-zero if error, zero if success.
* sysdeps/unix/sysv/linux/sparc/bits/pthreadtypes.h (pthread_cond_t):
Add __data.__futex field, reshuffle __data.__clock.
* sysdeps/unix/sysv/linux/ia64/lowlevellock.h (FUTEX_CMP_REQUEUE):
Define.
(lll_futex_requeue): Add val argument, use FUTEX_CMP_REQUEUE
internally. Return non-zero if error, zero if success.
* sysdeps/unix/sysv/linux/ia64/bits/pthreadtypes.h (pthread_cond_t):
Add __data.__futex field, reshuffle __data.__clock.
* sysdeps/unix/sysv/linux/x86_64/bits/pthreadtypes.h (pthread_cond_t):
Add __data.__futex field, reshuffle __data.__clock.
* sysdeps/pthread/pthread_cond_signal.c (__pthread_cond_signal):
Increment __futex at the same time as __wakeup_seq or __total_seq.
Pass address of __futex instead of address of low 32-bits of
__wakeup_seq to futex syscall.
* sysdeps/pthread/pthread_cond_wait.c (__pthread_cond_wait): Likewise.
Pass __futex value from before releasing internal lock
to FUTEX_WAIT.
* sysdeps/pthread/pthread_cond_timedwait.c (__pthread_cond_timedwait):
Likewise. Avoid unnecessary shadowing of variables.
* sysdeps/pthread/pthread_cond_broadcast.c (__pthread_cond_broadcast):
Set __futex to 2 * __total_seq. Pass __futex value from before the
unlock and __futex address instead of address of low 32-bits of
__wakeup_seq to futex_requeue macro, adjust for new return value
meaning.
* sysdeps/unix/sysv/linux/x86_64/pthread_cond_signal.S
(__pthread_cond_signal): Increment __futex at the same time as
__wakeup_seq or __total_seq. Pass address of __futex instead of
address of low 32-bits of __wakeup_seq to futex syscall.
* sysdeps/unix/sysv/linux/x86_64/pthread_cond_wait.S
(__pthread_cond_wait): Likewise. Pass __futex value from before
releasing internal lock to FUTEX_WAIT.
* sysdeps/unix/sysv/linux/x86_64/pthread_cond_timedwait.S
(__pthread_cond_timedwait): Likewise.
* sysdeps/unix/sysv/linux/x86_64/pthread_cond_broadcast.S
(FUTEX_CMP_REQUEUE): Define.
(__pthread_cond_broadcast): Set __futex to 2 * __total_seq.
Use FUTEX_CMP_REQUEUE operation instead of FUTEX_REQUEUE.
Pass __futex value from before the unlock and __futex address instead
of address of low 32-bits of __wakeup_seq to futex syscall.
Fallback to FUTEX_WAKE all on any errors.
2004-06-03 Kaz Kojima <kkojima@rr.iij4u.or.jp>
* sysdeps/unix/sysv/linux/sh/lowlevellock.h (lll_mutex_lock):
Add nop to align the end of critical section.
(lll_mutex_cond_lock, lll_mutex_timedlock): Likewise.
2004-06-01 Kaz Kojima <kkojima@rr.iij4u.or.jp> 2004-06-01 Kaz Kojima <kkojima@rr.iij4u.or.jp>
* sysdeps/unix/sysv/linux/sh/bits/pthreadtypes.h (pthread_cond_t): * sysdeps/unix/sysv/linux/sh/bits/pthreadtypes.h (pthread_cond_t):

View File

@ -194,7 +194,7 @@ tests = tst-attr1 tst-attr2 tst-attr3 \
tst-spin1 tst-spin2 tst-spin3 \ tst-spin1 tst-spin2 tst-spin3 \
tst-cond1 tst-cond2 tst-cond3 tst-cond4 tst-cond5 tst-cond6 tst-cond7 \ tst-cond1 tst-cond2 tst-cond3 tst-cond4 tst-cond5 tst-cond6 tst-cond7 \
tst-cond8 tst-cond9 tst-cond10 tst-cond11 tst-cond12 tst-cond13 \ tst-cond8 tst-cond9 tst-cond10 tst-cond11 tst-cond12 tst-cond13 \
tst-cond14 tst-cond15 tst-cond17 tst-cond18 \ tst-cond14 tst-cond15 tst-cond16 tst-cond17 tst-cond18 \
tst-rwlock1 tst-rwlock2 tst-rwlock3 tst-rwlock4 tst-rwlock5 \ tst-rwlock1 tst-rwlock2 tst-rwlock3 tst-rwlock4 tst-rwlock5 \
tst-rwlock6 tst-rwlock7 tst-rwlock8 tst-rwlock9 tst-rwlock10 \ tst-rwlock6 tst-rwlock7 tst-rwlock8 tst-rwlock9 tst-rwlock10 \
tst-rwlock11 tst-rwlock12 tst-rwlock13 \ tst-rwlock11 tst-rwlock12 tst-rwlock13 \

View File

@ -31,6 +31,7 @@ __pthread_cond_init (cond, cond_attr)
struct pthread_condattr *icond_attr = (struct pthread_condattr *) cond_attr; struct pthread_condattr *icond_attr = (struct pthread_condattr *) cond_attr;
cond->__data.__lock = LLL_MUTEX_LOCK_INITIALIZER; cond->__data.__lock = LLL_MUTEX_LOCK_INITIALIZER;
cond->__data.__futex = 0;
cond->__data.__clock = (icond_attr == NULL cond->__data.__clock = (icond_attr == NULL
? CLOCK_REALTIME : (icond_attr->value & 0xfe) >> 1); ? CLOCK_REALTIME : (icond_attr->value & 0xfe) >> 1);
cond->__data.__total_seq = 0; cond->__data.__total_seq = 0;

View File

@ -41,35 +41,29 @@ __pthread_cond_broadcast (cond)
/* Yes. Mark them all as woken. */ /* Yes. Mark them all as woken. */
cond->__data.__wakeup_seq = cond->__data.__total_seq; cond->__data.__wakeup_seq = cond->__data.__total_seq;
cond->__data.__woken_seq = cond->__data.__total_seq; cond->__data.__woken_seq = cond->__data.__total_seq;
cond->__data.__futex = (unsigned int) cond->__data.__total_seq * 2;
int futex_val = cond->__data.__futex;
/* Signal that a broadcast happened. */ /* Signal that a broadcast happened. */
++cond->__data.__broadcast_seq; ++cond->__data.__broadcast_seq;
/* We are done. */ /* We are done. */
lll_mutex_unlock (cond->__data.__lock); lll_mutex_unlock (cond->__data.__lock);
/* The futex syscall operates on a 32-bit word. That is fine,
we just use the low 32 bits of the sequence counter. */
#if BYTE_ORDER == LITTLE_ENDIAN
int *futex = ((int *) (&cond->__data.__wakeup_seq));
#elif BYTE_ORDER == BIG_ENDIAN
int *futex = ((int *) (&cond->__data.__wakeup_seq)) + 1;
#else
# error "No valid byte order"
#endif
/* Do not use requeue for pshared condvars. */ /* Do not use requeue for pshared condvars. */
if (cond->__data.__mutex == (void *) ~0l) if (cond->__data.__mutex == (void *) ~0l)
goto wake_all; goto wake_all;
/* Wake everybody. */ /* Wake everybody. */
pthread_mutex_t *mut = (pthread_mutex_t *) cond->__data.__mutex; pthread_mutex_t *mut = (pthread_mutex_t *) cond->__data.__mutex;
if (__builtin_expect (lll_futex_requeue (futex, 1, INT_MAX, /* lll_futex_requeue returns 0 for success and non-zero
&mut->__data.__lock) == -EINVAL, for errors. */
0)) if (__builtin_expect (lll_futex_requeue (&cond->__data.__futex, 1,
INT_MAX, &mut->__data.__lock,
futex_val), 0))
{ {
/* The requeue functionality is not available. */ /* The requeue functionality is not available. */
wake_all: wake_all:
lll_futex_wake (futex, INT_MAX); lll_futex_wake (&cond->__data.__futex, INT_MAX);
} }
/* That's all. */ /* That's all. */

View File

@ -1,4 +1,4 @@
/* Copyright (C) 2003 Free Software Foundation, Inc. /* Copyright (C) 2003, 2004 Free Software Foundation, Inc.
This file is part of the GNU C Library. This file is part of the GNU C Library.
Contributed by Martin Schwidefsky <schwidefsky@de.ibm.com>, 2003. Contributed by Martin Schwidefsky <schwidefsky@de.ibm.com>, 2003.
@ -40,19 +40,10 @@ __pthread_cond_signal (cond)
{ {
/* Yes. Mark one of them as woken. */ /* Yes. Mark one of them as woken. */
++cond->__data.__wakeup_seq; ++cond->__data.__wakeup_seq;
++cond->__data.__futex;
/* The futex syscall operates on a 32-bit word. That is fine,
we just use the low 32 bits of the sequence counter. */
#if BYTE_ORDER == LITTLE_ENDIAN
int *futex = ((int *) (&cond->__data.__wakeup_seq));
#elif BYTE_ORDER == BIG_ENDIAN
int *futex = ((int *) (&cond->__data.__wakeup_seq)) + 1;
#else
# error "No valid byte order"
#endif
/* Wake one. */ /* Wake one. */
lll_futex_wake (futex, 1); lll_futex_wake (&cond->__data.__futex, 1);
} }
/* We are done. */ /* We are done. */

View File

@ -66,6 +66,7 @@ __pthread_cond_timedwait (cond, mutex, abstime)
/* We have one new user of the condvar. */ /* We have one new user of the condvar. */
++cond->__data.__total_seq; ++cond->__data.__total_seq;
++cond->__data.__futex;
/* Remember the mutex we are using here. If there is already a /* Remember the mutex we are using here. If there is already a
different address store this is a bad user bug. Do not store different address store this is a bad user bug. Do not store
@ -89,27 +90,17 @@ __pthread_cond_timedwait (cond, mutex, abstime)
/* Remember the broadcast counter. */ /* Remember the broadcast counter. */
cbuffer.bc_seq = cond->__data.__broadcast_seq; cbuffer.bc_seq = cond->__data.__broadcast_seq;
/* The futex syscall operates on a 32-bit word. That is fine, we
just use the low 32 bits of the sequence counter. */
#if BYTE_ORDER == LITTLE_ENDIAN
int *futex = ((int *) (&cond->__data.__wakeup_seq));
#elif BYTE_ORDER == BIG_ENDIAN
int *futex = ((int *) (&cond->__data.__wakeup_seq)) + 1;
#else
# error "No valid byte order"
#endif
while (1) while (1)
{ {
struct timespec rt; struct timespec rt;
{ {
#ifdef __NR_clock_gettime #ifdef __NR_clock_gettime
INTERNAL_SYSCALL_DECL (err); INTERNAL_SYSCALL_DECL (err);
int val; int ret;
val = INTERNAL_SYSCALL (clock_gettime, err, 2, ret = INTERNAL_SYSCALL (clock_gettime, err, 2,
cond->__data.__clock, &rt); cond->__data.__clock, &rt);
# ifndef __ASSUME_POSIX_TIMERS # ifndef __ASSUME_POSIX_TIMERS
if (__builtin_expect (INTERNAL_SYSCALL_ERROR_P (val, err), 0)) if (__builtin_expect (INTERNAL_SYSCALL_ERROR_P (ret, err), 0))
{ {
struct timeval tv; struct timeval tv;
(void) gettimeofday (&tv, NULL); (void) gettimeofday (&tv, NULL);
@ -149,15 +140,17 @@ __pthread_cond_timedwait (cond, mutex, abstime)
goto timeout; goto timeout;
} }
unsigned int futex_val = cond->__data.__futex;
/* Prepare to wait. Release the condvar futex. */ /* Prepare to wait. Release the condvar futex. */
lll_mutex_unlock (cond->__data.__lock); lll_mutex_unlock (cond->__data.__lock);
/* Enable asynchronous cancellation. Required by the standard. */ /* Enable asynchronous cancellation. Required by the standard. */
cbuffer.oldtype = __pthread_enable_asynccancel (); cbuffer.oldtype = __pthread_enable_asynccancel ();
/* Wait until woken by signal or broadcast. Note that we /* Wait until woken by signal or broadcast. */
truncate the 'val' value to 32 bits. */ err = lll_futex_timed_wait (&cond->__data.__futex,
err = lll_futex_timed_wait (futex, (unsigned int) val, &rt); futex_val, &rt);
/* Disable asynchronous cancellation. */ /* Disable asynchronous cancellation. */
__pthread_disable_asynccancel (cbuffer.oldtype); __pthread_disable_asynccancel (cbuffer.oldtype);
@ -180,6 +173,7 @@ __pthread_cond_timedwait (cond, mutex, abstime)
timeout: timeout:
/* Yep. Adjust the counters. */ /* Yep. Adjust the counters. */
++cond->__data.__wakeup_seq; ++cond->__data.__wakeup_seq;
++cond->__data.__futex;
/* The error value. */ /* The error value. */
result = ETIMEDOUT; result = ETIMEDOUT;

View File

@ -52,20 +52,14 @@ __condvar_cleanup (void *arg)
appropriately. */ appropriately. */
++cbuffer->cond->__data.__wakeup_seq; ++cbuffer->cond->__data.__wakeup_seq;
++cbuffer->cond->__data.__woken_seq; ++cbuffer->cond->__data.__woken_seq;
++cbuffer->cond->__data.__futex;
} }
/* We are done. */ /* We are done. */
lll_mutex_unlock (cbuffer->cond->__data.__lock); lll_mutex_unlock (cbuffer->cond->__data.__lock);
/* Wake everybody to make sure no condvar signal gets lost. */ /* Wake everybody to make sure no condvar signal gets lost. */
#if BYTE_ORDER == LITTLE_ENDIAN lll_futex_wake (&cbuffer->cond->__data.__futex, INT_MAX);
int *futex = ((int *) (&cbuffer->cond->__data.__wakeup_seq));
#elif BYTE_ORDER == BIG_ENDIAN
int *futex = ((int *) (&cbuffer->cond->__data.__wakeup_seq)) + 1;
#else
# error "No valid byte order"
#endif
lll_futex_wake (futex, INT_MAX);
/* Get the mutex before returning unless asynchronous cancellation /* Get the mutex before returning unless asynchronous cancellation
is in effect. */ is in effect. */
@ -95,6 +89,7 @@ __pthread_cond_wait (cond, mutex)
/* We have one new user of the condvar. */ /* We have one new user of the condvar. */
++cond->__data.__total_seq; ++cond->__data.__total_seq;
++cond->__data.__futex;
/* Remember the mutex we are using here. If there is already a /* Remember the mutex we are using here. If there is already a
different address store this is a bad user bug. Do not store different address store this is a bad user bug. Do not store
@ -118,27 +113,18 @@ __pthread_cond_wait (cond, mutex)
/* Remember the broadcast counter. */ /* Remember the broadcast counter. */
cbuffer.bc_seq = cond->__data.__broadcast_seq; cbuffer.bc_seq = cond->__data.__broadcast_seq;
/* The futex syscall operates on a 32-bit word. That is fine, we
just use the low 32 bits of the sequence counter. */
#if BYTE_ORDER == LITTLE_ENDIAN
int *futex = ((int *) (&cond->__data.__wakeup_seq));
#elif BYTE_ORDER == BIG_ENDIAN
int *futex = ((int *) (&cond->__data.__wakeup_seq)) + 1;
#else
# error "No valid byte order"
#endif
do do
{ {
unsigned int futex_val = cond->__data.__futex;
/* Prepare to wait. Release the condvar futex. */ /* Prepare to wait. Release the condvar futex. */
lll_mutex_unlock (cond->__data.__lock); lll_mutex_unlock (cond->__data.__lock);
/* Enable asynchronous cancellation. Required by the standard. */ /* Enable asynchronous cancellation. Required by the standard. */
cbuffer.oldtype = __pthread_enable_asynccancel (); cbuffer.oldtype = __pthread_enable_asynccancel ();
/* Wait until woken by signal or broadcast. Note that we /* Wait until woken by signal or broadcast. */
truncate the 'val' value to 32 bits. */ lll_futex_wait (&cond->__data.__futex, futex_val);
lll_futex_wait (futex, (unsigned int) val);
/* Disable asynchronous cancellation. */ /* Disable asynchronous cancellation. */
__pthread_disable_asynccancel (cbuffer.oldtype); __pthread_disable_asynccancel (cbuffer.oldtype);

View File

@ -76,11 +76,12 @@ typedef union
struct struct
{ {
int __lock; int __lock;
int __clock; unsigned int __futex;
unsigned long long int __total_seq; unsigned long long int __total_seq;
unsigned long long int __wakeup_seq; unsigned long long int __wakeup_seq;
unsigned long long int __woken_seq; unsigned long long int __woken_seq;
void *__mutex; void *__mutex;
int __clock;
unsigned int __broadcast_seq; unsigned int __broadcast_seq;
} __data; } __data;
char __size[__SIZEOF_PTHREAD_COND_T]; char __size[__SIZEOF_PTHREAD_COND_T];

View File

@ -30,6 +30,7 @@
#define FUTEX_WAIT 0 #define FUTEX_WAIT 0
#define FUTEX_WAKE 1 #define FUTEX_WAKE 1
#define FUTEX_REQUEUE 3 #define FUTEX_REQUEUE 3
#define FUTEX_CMP_REQUEUE 4
/* Initializer for compatibility lock. */ /* Initializer for compatibility lock. */
#define LLL_MUTEX_LOCK_INITIALIZER (0) #define LLL_MUTEX_LOCK_INITIALIZER (0)
@ -61,14 +62,15 @@
INTERNAL_SYSCALL_ERROR_P (__ret, __err)? -__ret : __ret; \ INTERNAL_SYSCALL_ERROR_P (__ret, __err)? -__ret : __ret; \
}) })
#define lll_futex_requeue(futexp, nr_wake, nr_move, mutex) \ /* Returns non-zero if error happened, zero if success. */
#define lll_futex_requeue(futexp, nr_wake, nr_move, mutex, val) \
({ \ ({ \
INTERNAL_SYSCALL_DECL (__err); \ INTERNAL_SYSCALL_DECL (__err); \
long int __ret; \ long int __ret; \
__ret = INTERNAL_SYSCALL (futex, __err, 5, \ __ret = INTERNAL_SYSCALL (futex, __err, 6, \
(futexp), FUTEX_REQUEUE, (nr_wake), (nr_move), \ (futexp), FUTEX_CMP_REQUEUE, (nr_wake), \
(mutex)); \ (nr_move), (mutex), (val)); \
INTERNAL_SYSCALL_ERROR_P (__ret, __err)? -__ret : __ret; \ INTERNAL_SYSCALL_ERROR_P (__ret, __err); \
}) })

View File

@ -76,11 +76,12 @@ typedef union
struct struct
{ {
int __lock; int __lock;
int __clock; unsigned int __futex;
unsigned long long int __total_seq; unsigned long long int __total_seq;
unsigned long long int __wakeup_seq; unsigned long long int __wakeup_seq;
unsigned long long int __woken_seq; unsigned long long int __woken_seq;
void *__mutex; void *__mutex;
int __clock;
unsigned int __broadcast_seq; unsigned int __broadcast_seq;
} __data; } __data;
char __size[__SIZEOF_PTHREAD_COND_T]; char __size[__SIZEOF_PTHREAD_COND_T];

View File

@ -32,6 +32,7 @@
#define FUTEX_WAIT 0 #define FUTEX_WAIT 0
#define FUTEX_WAKE 1 #define FUTEX_WAKE 1
#define FUTEX_REQUEUE 3 #define FUTEX_REQUEUE 3
#define FUTEX_CMP_REQUEUE 4
#define EINVAL 22 #define EINVAL 22
@ -47,8 +48,9 @@ __pthread_cond_broadcast:
pushl %ebx pushl %ebx
pushl %esi pushl %esi
pushl %edi pushl %edi
pushl %ebp
movl 16(%esp), %ebx movl 20(%esp), %ebx
/* Get internal lock. */ /* Get internal lock. */
movl $1, %edx movl $1, %edx
@ -61,29 +63,31 @@ __pthread_cond_broadcast:
#endif #endif
jnz 1f jnz 1f
2: addl $wakeup_seq, %ebx 2: addl $cond_futex, %ebx
movl total_seq+4-wakeup_seq(%ebx), %eax movl total_seq+4-cond_futex(%ebx), %eax
movl total_seq-wakeup_seq(%ebx), %ecx movl total_seq-cond_futex(%ebx), %ebp
cmpl 4(%ebx), %eax cmpl wakeup_seq+4-cond_futex(%ebx), %eax
ja 3f ja 3f
jb 4f jb 4f
cmpl (%ebx), %ecx cmpl wakeup_seq-cond_futex(%ebx), %ebp
jna 4f jna 4f
/* Cause all currently waiting threads to recognize they are /* Cause all currently waiting threads to recognize they are
woken up. */ woken up. */
3: movl %ecx, (%ebx) 3: movl %ebp, wakeup_seq-cond_futex(%ebx)
movl %eax, 4(%ebx) movl %eax, wakeup_seq-cond_futex+4(%ebx)
movl %ecx, woken_seq-wakeup_seq(%ebx) movl %ebp, woken_seq-cond_futex(%ebx)
movl %eax, woken_seq-wakeup_seq+4(%ebx) movl %eax, woken_seq-cond_futex+4(%ebx)
addl $1, broadcast_seq-wakeup_seq(%ebx) addl %ebp, %ebp
addl $1, broadcast_seq-cond_futex(%ebx)
movl %ebp, (%ebx)
/* Get the address of the mutex used. */ /* Get the address of the mutex used. */
movl dep_mutex-wakeup_seq(%ebx), %edi movl dep_mutex-cond_futex(%ebx), %edi
/* Unlock. */ /* Unlock. */
LOCK LOCK
subl $1, cond_lock-wakeup_seq(%ebx) subl $1, cond_lock-cond_futex(%ebx)
jne 7f jne 7f
/* Don't use requeue for pshared condvars. */ /* Don't use requeue for pshared condvars. */
@ -91,7 +95,7 @@ __pthread_cond_broadcast:
je 9f je 9f
/* Wake up all threads. */ /* Wake up all threads. */
movl $FUTEX_REQUEUE, %ecx movl $FUTEX_CMP_REQUEUE, %ecx
movl $SYS_futex, %eax movl $SYS_futex, %eax
movl $0x7fffffff, %esi movl $0x7fffffff, %esi
movl $1, %edx movl $1, %edx
@ -99,14 +103,18 @@ __pthread_cond_broadcast:
# if MUTEX_FUTEX != 0 # if MUTEX_FUTEX != 0
addl $MUTEX_FUTEX, %edi addl $MUTEX_FUTEX, %edi
# endif # endif
ENTER_KERNEL /* FIXME: Until Ingo fixes 4G/4G vDSO, 6 arg syscalls are broken for sysenter.
ENTER_KERNEL */
int $0x80
#ifndef __ASSUME_FUTEX_REQUEUE /* For any kind of error, which mainly is EAGAIN, we try again
cmpl $-EINVAL, %eax with WAKE. The general test also covers running on old
je 9f kernels. */
#endif cmpl $0xfffff001, %eax
jae 9f
10: xorl %eax, %eax 10: xorl %eax, %eax
popl %ebp
popl %edi popl %edi
popl %esi popl %esi
popl %ebx popl %ebx
@ -115,10 +123,11 @@ __pthread_cond_broadcast:
.align 16 .align 16
/* Unlock. */ /* Unlock. */
4: LOCK 4: LOCK
subl $1, cond_lock-wakeup_seq(%ebx) subl $1, cond_lock-cond_futex(%ebx)
jne 5f jne 5f
6: xorl %eax, %eax 6: xorl %eax, %eax
popl %ebp
popl %edi popl %edi
popl %esi popl %esi
popl %ebx popl %ebx
@ -135,12 +144,12 @@ __pthread_cond_broadcast:
jmp 2b jmp 2b
/* Unlock in loop requires waekup. */ /* Unlock in loop requires waekup. */
5: leal cond_lock-wakeup_seq(%ebx), %eax 5: leal cond_lock-cond_futex(%ebx), %eax
call __lll_mutex_unlock_wake call __lll_mutex_unlock_wake
jmp 6b jmp 6b
/* Unlock in loop requires waekup. */ /* Unlock in loop requires waekup. */
7: leal cond_lock-wakeup_seq(%ebx), %eax 7: leal cond_lock-cond_futex(%ebx), %eax
call __lll_mutex_unlock_wake call __lll_mutex_unlock_wake
jmp 8b jmp 8b

View File

@ -60,22 +60,23 @@ __pthread_cond_signal:
#endif #endif
jnz 1f jnz 1f
2: leal wakeup_seq(%edi), %ebx 2: leal cond_futex(%edi), %ebx
movl total_seq+4(%edi), %eax movl total_seq+4(%edi), %eax
movl total_seq(%edi), %ecx movl total_seq(%edi), %ecx
cmpl 4(%ebx), %eax cmpl wakeup_seq+4(%edi), %eax
#if cond_lock != 0 #if cond_lock != 0
/* Must use leal to preserve the flags. */ /* Must use leal to preserve the flags. */
leal cond_lock(%edi), %edi leal cond_lock(%edi), %edi
#endif #endif
ja 3f ja 3f
jb 4f jb 4f
cmpl (%ebx), %ecx cmpl wakeup_seq-cond_futex(%ebx), %ecx
jbe 4f jbe 4f
/* Bump the wakeup number. */ /* Bump the wakeup number. */
3: addl $1, (%ebx) 3: addl $1, wakeup_seq-cond_futex(%ebx)
adcl $0, 4(%ebx) adcl $0, wakeup_seq-cond_futex+4(%ebx)
addl $1, (%ebx)
/* Wake up one thread. */ /* Wake up one thread. */
movl $FUTEX_WAKE, %ecx movl $FUTEX_WAKE, %ecx

View File

@ -79,6 +79,7 @@ __pthread_cond_wait:
addl $1, total_seq(%ebx) addl $1, total_seq(%ebx)
adcl $0, total_seq+4(%ebx) adcl $0, total_seq+4(%ebx)
addl $1, cond_futex(%ebx)
#define FRAME_SIZE 16 #define FRAME_SIZE 16
subl $FRAME_SIZE, %esp subl $FRAME_SIZE, %esp
@ -92,8 +93,10 @@ __pthread_cond_wait:
movl %edx, 8(%esp) movl %edx, 8(%esp)
movl %eax, 12(%esp) movl %eax, 12(%esp)
8: movl cond_futex(%ebx), %edi
/* Unlock. */ /* Unlock. */
8: LOCK LOCK
#if cond_lock == 0 #if cond_lock == 0
subl $1, (%ebx) subl $1, (%ebx)
#else #else
@ -107,12 +110,12 @@ __pthread_cond_wait:
movl %esi, %ecx /* movl $FUTEX_WAIT, %ecx */ movl %esi, %ecx /* movl $FUTEX_WAIT, %ecx */
movl %edi, %edx movl %edi, %edx
addl $wakeup_seq, %ebx addl $cond_futex, %ebx
.Ladd_wakeup: .Ladd_cond_futex:
movl $SYS_futex, %eax movl $SYS_futex, %eax
ENTER_KERNEL ENTER_KERNEL
subl $wakeup_seq, %ebx subl $cond_futex, %ebx
.Lsub_wakeup: .Lsub_cond_futex:
movl (%esp), %eax movl (%esp), %eax
call __pthread_disable_asynccancel call __pthread_disable_asynccancel
@ -246,7 +249,7 @@ versioned_symbol (libpthread, __pthread_cond_wait, pthread_cond_wait,
.type __condvar_w_cleanup2, @function .type __condvar_w_cleanup2, @function
__condvar_w_cleanup2: __condvar_w_cleanup2:
subl $wakeup_seq, %ebx subl $cond_futex, %ebx
.size __condvar_w_cleanup2, .-__condvar_w_cleanup2 .size __condvar_w_cleanup2, .-__condvar_w_cleanup2
.LSbl4: .LSbl4:
.type __condvar_w_cleanup, @function .type __condvar_w_cleanup, @function
@ -278,6 +281,8 @@ __condvar_w_cleanup:
addl $1, wakeup_seq(%ebx) addl $1, wakeup_seq(%ebx)
adcl $0, wakeup_seq+4(%ebx) adcl $0, wakeup_seq+4(%ebx)
addl $1, cond_futex(%ebx)
addl $1, woken_seq(%ebx) addl $1, woken_seq(%ebx)
adcl $0, woken_seq+4(%ebx) adcl $0, woken_seq+4(%ebx)
@ -297,7 +302,7 @@ __condvar_w_cleanup:
call __lll_mutex_unlock_wake call __lll_mutex_unlock_wake
/* Wake up all waiters to make sure no signal gets lost. */ /* Wake up all waiters to make sure no signal gets lost. */
2: addl $wakeup_seq, %ebx 2: addl $cond_futex, %ebx
movl $FUTEX_WAKE, %ecx movl $FUTEX_WAKE, %ecx
movl $SYS_futex, %eax movl $SYS_futex, %eax
movl $0x7fffffff, %edx movl $0x7fffffff, %edx
@ -323,15 +328,15 @@ __condvar_w_cleanup:
.uleb128 .Lcstend-.Lcstbegin .uleb128 .Lcstend-.Lcstbegin
.Lcstbegin: .Lcstbegin:
.long .LcleanupSTART-.LSTARTCODE .long .LcleanupSTART-.LSTARTCODE
.long .Ladd_wakeup-.LcleanupSTART .long .Ladd_cond_futex-.LcleanupSTART
.long __condvar_w_cleanup-.LSTARTCODE .long __condvar_w_cleanup-.LSTARTCODE
.uleb128 0 .uleb128 0
.long .Ladd_wakeup-.LSTARTCODE .long .Ladd_cond_futex-.LSTARTCODE
.long .Lsub_wakeup-.Ladd_wakeup .long .Lsub_cond_futex-.Ladd_cond_futex
.long __condvar_w_cleanup2-.LSTARTCODE .long __condvar_w_cleanup2-.LSTARTCODE
.uleb128 0 .uleb128 0
.long .Lsub_wakeup-.LSTARTCODE .long .Lsub_cond_futex-.LSTARTCODE
.long .LcleanupEND-.Lsub_wakeup .long .LcleanupEND-.Lsub_cond_futex
.long __condvar_w_cleanup-.LSTARTCODE .long __condvar_w_cleanup-.LSTARTCODE
.uleb128 0 .uleb128 0
.long .LcallUR-.LSTARTCODE .long .LcallUR-.LSTARTCODE

View File

@ -76,11 +76,12 @@ typedef union
struct struct
{ {
int __lock; int __lock;
int __clock; unsigned int __futex;
unsigned long long int __total_seq; unsigned long long int __total_seq;
unsigned long long int __wakeup_seq; unsigned long long int __wakeup_seq;
unsigned long long int __woken_seq; unsigned long long int __woken_seq;
void *__mutex; void *__mutex;
int __clock;
unsigned int __broadcast_seq; unsigned int __broadcast_seq;
} __data; } __data;
char __size[__SIZEOF_PTHREAD_COND_T]; char __size[__SIZEOF_PTHREAD_COND_T];

View File

@ -30,6 +30,7 @@
#define FUTEX_WAIT 0 #define FUTEX_WAIT 0
#define FUTEX_WAKE 1 #define FUTEX_WAKE 1
#define FUTEX_REQUEUE 3 #define FUTEX_REQUEUE 3
#define FUTEX_CMP_REQUEUE 4
/* Delay in spinlock loop. */ /* Delay in spinlock loop. */
#define BUSY_WAIT_NOP asm ("hint @pause") #define BUSY_WAIT_NOP asm ("hint @pause")
@ -52,11 +53,13 @@
_r10 == -1 ? -_retval : _retval; \ _r10 == -1 ? -_retval : _retval; \
}) })
#define lll_futex_requeue(ftx, nr_wake, nr_move, mutex) \ /* Returns non-zero if error happened, zero if success. */
#define lll_futex_requeue(ftx, nr_wake, nr_move, mutex, val) \
({ \ ({ \
DO_INLINE_SYSCALL(futex, 5, (long) (ftx), FUTEX_REQUEUE, (int) (nr_wake), \ DO_INLINE_SYSCALL(futex, 6, (long) (ftx), FUTEX_CMP_REQUEUE, \
(int) (nr_move), (long) (mutex)); \ (int) (nr_wake), (int) (nr_move), (long) (mutex), \
_r10 == -1 ? -_retval : _retval; \ (int) val); \
_r10 == -1; \
}) })

View File

@ -4,6 +4,7 @@
-- --
cond_lock offsetof (pthread_cond_t, __data.__lock) cond_lock offsetof (pthread_cond_t, __data.__lock)
cond_futex offsetof (pthread_cond_t, __data.__futex)
cond_clock offsetof (pthread_cond_t, __data.__clock) cond_clock offsetof (pthread_cond_t, __data.__clock)
total_seq offsetof (pthread_cond_t, __data.__total_seq) total_seq offsetof (pthread_cond_t, __data.__total_seq)
wakeup_seq offsetof (pthread_cond_t, __data.__wakeup_seq) wakeup_seq offsetof (pthread_cond_t, __data.__wakeup_seq)

View File

@ -96,11 +96,12 @@ typedef union
struct struct
{ {
int __lock; int __lock;
int __clock; unsigned int __futex;
unsigned long long int __total_seq; unsigned long long int __total_seq;
unsigned long long int __wakeup_seq; unsigned long long int __wakeup_seq;
unsigned long long int __woken_seq; unsigned long long int __woken_seq;
void *__mutex; void *__mutex;
int __clock;
unsigned int __broadcast_seq; unsigned int __broadcast_seq;
} __data; } __data;
char __size[__SIZEOF_PTHREAD_COND_T]; char __size[__SIZEOF_PTHREAD_COND_T];

View File

@ -30,6 +30,7 @@
#define FUTEX_WAIT 0 #define FUTEX_WAIT 0
#define FUTEX_WAKE 1 #define FUTEX_WAKE 1
#define FUTEX_REQUEUE 3 #define FUTEX_REQUEUE 3
#define FUTEX_CMP_REQUEUE 4
/* Initializer for compatibility lock. */ /* Initializer for compatibility lock. */
#define LLL_MUTEX_LOCK_INITIALIZER (0) #define LLL_MUTEX_LOCK_INITIALIZER (0)
@ -64,15 +65,16 @@
INTERNAL_SYSCALL_ERROR_P (__ret, __err) ? -__ret : __ret; \ INTERNAL_SYSCALL_ERROR_P (__ret, __err) ? -__ret : __ret; \
}) })
#define lll_futex_requeue(futexp, nr_wake, nr_move, mutex) \ /* Returns non-zero if error happened, zero if success. */
#define lll_futex_requeue(futexp, nr_wake, nr_move, mutex, val) \
({ \ ({ \
INTERNAL_SYSCALL_DECL (__err); \ INTERNAL_SYSCALL_DECL (__err); \
long int __ret; \ long int __ret; \
\ \
__ret = INTERNAL_SYSCALL (futex, __err, 5, \ __ret = INTERNAL_SYSCALL (futex, __err, 6, \
(futexp), FUTEX_REQUEUE, (nr_wake), (nr_move), \ (futexp), FUTEX_CMP_REQUEUE, (nr_wake), \
(mutex)); \ (nr_move), (mutex), (val)); \
INTERNAL_SYSCALL_ERROR_P (__ret, __err) ? -__ret : __ret; \ INTERNAL_SYSCALL_ERROR_P (__ret, __err); \
}) })
#ifdef UP #ifdef UP

View File

@ -95,11 +95,12 @@ typedef union
struct struct
{ {
int __lock; int __lock;
int __clock; unsigned int __futex;
unsigned long long int __total_seq; unsigned long long int __total_seq;
unsigned long long int __wakeup_seq; unsigned long long int __wakeup_seq;
unsigned long long int __woken_seq; unsigned long long int __woken_seq;
void *__mutex; void *__mutex;
int __clock;
unsigned int __broadcast_seq; unsigned int __broadcast_seq;
} __data; } __data;
char __size[__SIZEOF_PTHREAD_COND_T]; char __size[__SIZEOF_PTHREAD_COND_T];

View File

@ -29,6 +29,7 @@
#define FUTEX_WAIT 0 #define FUTEX_WAIT 0
#define FUTEX_WAKE 1 #define FUTEX_WAKE 1
#define FUTEX_REQUEUE 3 #define FUTEX_REQUEUE 3
#define FUTEX_CMP_REQUEUE 4
/* Initializer for compatibility lock. */ /* Initializer for compatibility lock. */
#define LLL_MUTEX_LOCK_INITIALIZER (0) #define LLL_MUTEX_LOCK_INITIALIZER (0)
@ -82,22 +83,29 @@
}) })
#define lll_futex_requeue(futex, nr_wake, nr_move, mutex) \ /* Returns non-zero if error happened, zero if success. */
#if 0
/* FIXME: s390 only supports up to 5 argument syscalls. Once FUTEX_CMP_REQUEUE
kernel interface for s390 is finalized, adjust this. */
#define lll_futex_requeue(futex, nr_wake, nr_move, mutex, val) \
({ \ ({ \
register unsigned long int __r2 asm ("2") = (unsigned long int) (futex); \ register unsigned long int __r2 asm ("2") = (unsigned long int) (futex); \
register unsigned long int __r3 asm ("3") = FUTEX_REQUEUE; \ register unsigned long int __r3 asm ("3") = FUTEX_CMP_REQUEUE; \
register unsigned long int __r4 asm ("4") = (long int) (nr_wake); \ register unsigned long int __r4 asm ("4") = (long int) (nr_wake); \
register unsigned long int __r5 asm ("5") = (long int) (nr_move); \ register unsigned long int __r5 asm ("5") = (long int) (nr_move); \
register unsigned long int __r6 asm ("6") = (unsigned long int) (mutex); \ register unsigned long int __r6 asm ("6") = (unsigned long int) (mutex); \
register unsigned long __result asm ("2"); \ register unsigned long __result asm ("2"); \
\ \
__asm __volatile ("svc %b1" \ __asm __volatile ("svc %b1" \
: "=d" (__result) \ : "=d" (__result) \
: "i" (SYS_futex), "0" (__r2), "d" (__r3), \ : "i" (SYS_futex), "0" (__r2), "d" (__r3), \
"d" (__r4), "d" (__r5), "d" (__r6) \ "d" (__r4), "d" (__r5), "d" (__r6) \
: "cc", "memory" ); \ : "cc", "memory" ); \
__result; \ __result > -4096UL; \
}) })
#else
#define lll_futex_requeue(futex, nr_wake, nr_move, mutex, val) 1
#endif
#define lll_compare_and_swap(futex, oldval, newval, operation) \ #define lll_compare_and_swap(futex, oldval, newval, operation) \

View File

@ -89,6 +89,7 @@ extern int __lll_mutex_unlock_wake (int *__futex) attribute_hidden;
__asm __volatile ("\ __asm __volatile ("\
.align 2\n\ .align 2\n\
mova 1f,r0\n\ mova 1f,r0\n\
nop\n\
mov r15,r1\n\ mov r15,r1\n\
mov #-8,r15\n\ mov #-8,r15\n\
0: mov.l @%2,%0\n\ 0: mov.l @%2,%0\n\
@ -108,6 +109,7 @@ extern int __lll_mutex_unlock_wake (int *__futex) attribute_hidden;
__asm __volatile ("\ __asm __volatile ("\
.align 2\n\ .align 2\n\
mova 1f,r0\n\ mova 1f,r0\n\
nop\n\
mov r15,r1\n\ mov r15,r1\n\
mov #-8,r15\n\ mov #-8,r15\n\
0: mov.l @%2,%0\n\ 0: mov.l @%2,%0\n\
@ -125,6 +127,7 @@ extern int __lll_mutex_unlock_wake (int *__futex) attribute_hidden;
__asm __volatile ("\ __asm __volatile ("\
.align 2\n\ .align 2\n\
mova 1f,r0\n\ mova 1f,r0\n\
nop\n\
mov r15,r1\n\ mov r15,r1\n\
mov #-8,r15\n\ mov #-8,r15\n\
0: mov.l @%2,%0\n\ 0: mov.l @%2,%0\n\

View File

@ -96,11 +96,12 @@ typedef union
struct struct
{ {
int __lock; int __lock;
int __clock; unsigned int __futex;
unsigned long long int __total_seq; unsigned long long int __total_seq;
unsigned long long int __wakeup_seq; unsigned long long int __wakeup_seq;
unsigned long long int __woken_seq; unsigned long long int __woken_seq;
void *__mutex; void *__mutex;
int __clock;
unsigned int __broadcast_seq; unsigned int __broadcast_seq;
} __data; } __data;
char __size[__SIZEOF_PTHREAD_COND_T]; char __size[__SIZEOF_PTHREAD_COND_T];

View File

@ -29,6 +29,7 @@
#define FUTEX_WAIT 0 #define FUTEX_WAIT 0
#define FUTEX_WAKE 1 #define FUTEX_WAKE 1
#define FUTEX_REQUEUE 3 #define FUTEX_REQUEUE 3
#define FUTEX_CMP_REQUEUE 4
/* Initializer for compatibility lock. */ /* Initializer for compatibility lock. */
#define LLL_MUTEX_LOCK_INITIALIZER (0) #define LLL_MUTEX_LOCK_INITIALIZER (0)
@ -63,15 +64,16 @@
__ret; \ __ret; \
}) })
#define lll_futex_requeue(futexp, nr_wake, nr_move, mutex) \ /* Returns non-zero if error happened, zero if success. */
#define lll_futex_requeue(futexp, nr_wake, nr_move, mutex, val) \
({ \ ({ \
INTERNAL_SYSCALL_DECL (__err); \ INTERNAL_SYSCALL_DECL (__err); \
long int __ret; \ long int __ret; \
\ \
__ret = INTERNAL_SYSCALL (futex, __err, 5, \ __ret = INTERNAL_SYSCALL (futex, __err, 6, \
(futexp), FUTEX_REQUEUE, (nr_wake), (nr_move), \ (futexp), FUTEX_CMP_REQUEUE, (nr_wake), \
(mutex)); \ (nr_move), (mutex), (val)); \
__ret; \ INTERNAL_SYSCALL_ERROR_P (__ret, __err); \
}) })
#ifdef __sparc32_atomic_do_lock #ifdef __sparc32_atomic_do_lock

View File

@ -76,11 +76,12 @@ typedef union
struct struct
{ {
int __lock; int __lock;
int __clock; unsigned int __futex;
unsigned long long int __total_seq; unsigned long long int __total_seq;
unsigned long long int __wakeup_seq; unsigned long long int __wakeup_seq;
unsigned long long int __woken_seq; unsigned long long int __woken_seq;
void *__mutex; void *__mutex;
int __clock;
unsigned int __broadcast_seq; unsigned int __broadcast_seq;
} __data; } __data;
char __size[__SIZEOF_PTHREAD_COND_T]; char __size[__SIZEOF_PTHREAD_COND_T];

View File

@ -32,6 +32,7 @@
#define FUTEX_WAIT 0 #define FUTEX_WAIT 0
#define FUTEX_WAKE 1 #define FUTEX_WAKE 1
#define FUTEX_REQUEUE 3 #define FUTEX_REQUEUE 3
#define FUTEX_CMP_REQUEUE 4
#define EINVAL 22 #define EINVAL 22
@ -55,39 +56,42 @@ __pthread_cond_broadcast:
#endif #endif
jnz 1f jnz 1f
2: addq $wakeup_seq, %rdi 2: addq $cond_futex, %rdi
movq total_seq-wakeup_seq(%rdi), %rcx movq total_seq-cond_futex(%rdi), %r9
cmpq (%rdi), %rcx cmpq wakeup_seq-cond_futex(%rdi), %r9
jna 4f jna 4f
/* Cause all currently waiting threads to recognize they are /* Cause all currently waiting threads to recognize they are
woken up. */ woken up. */
movq %rcx, (%rdi) movq %r9, wakeup_seq-cond_futex(%rdi)
movq %rcx, woken_seq-wakeup_seq(%rdi) movq %r9, woken_seq-cond_futex(%rdi)
incl broadcast_seq-wakeup_seq(%rdi) addq %r9, %r9
movl %r9d, (%rdi)
incl broadcast_seq-cond_futex(%rdi)
/* Get the address of the mutex used. */ /* Get the address of the mutex used. */
movq dep_mutex-wakeup_seq(%rdi), %r8 movq dep_mutex-cond_futex(%rdi), %r8
/* Unlock. */ /* Unlock. */
LOCK LOCK
decl cond_lock-wakeup_seq(%rdi) decl cond_lock-cond_futex(%rdi)
jne 7f jne 7f
8: cmpq $-1, %r8 8: cmpq $-1, %r8
je 9f je 9f
/* Wake up all threads. */ /* Wake up all threads. */
movq $FUTEX_REQUEUE, %rsi movq $FUTEX_CMP_REQUEUE, %rsi
movq $SYS_futex, %rax movq $SYS_futex, %rax
movl $1, %edx movl $1, %edx
movq $0x7fffffff, %r10 movq $0x7fffffff, %r10
syscall syscall
#ifndef __ASSUME_FUTEX_REQUEUE /* For any kind of error, which mainly is EAGAIN, we try again
cmpq $-EINVAL, %rax with WAKE. The general test also covers running on old
je 9f kernels. */
#endif cmpq $-4095, %rax
jae 9f
10: xorl %eax, %eax 10: xorl %eax, %eax
retq retq
@ -95,7 +99,7 @@ __pthread_cond_broadcast:
.align 16 .align 16
/* Unlock. */ /* Unlock. */
4: LOCK 4: LOCK
decl cond_lock-wakeup_seq(%rdi) decl cond_lock-cond_futex(%rdi)
jne 5f jne 5f
6: xorl %eax, %eax 6: xorl %eax, %eax
@ -113,14 +117,14 @@ __pthread_cond_broadcast:
jmp 2b jmp 2b
/* Unlock in loop requires wakeup. */ /* Unlock in loop requires wakeup. */
5: addq $cond_lock-wakeup_seq, %rdi 5: addq $cond_lock-cond_futex, %rdi
callq __lll_mutex_unlock_wake callq __lll_mutex_unlock_wake
jmp 6b jmp 6b
/* Unlock in loop requires wakeup. */ /* Unlock in loop requires wakeup. */
7: addq $cond_lock-wakeup_seq, %rdi 7: addq $cond_lock-cond_futex, %rdi
callq __lll_mutex_unlock_wake callq __lll_mutex_unlock_wake
subq $cond_lock-wakeup_seq, %rdi subq $cond_lock-cond_futex, %rdi
jmp 8b jmp 8b
9: /* The futex requeue functionality is not available. */ 9: /* The futex requeue functionality is not available. */

View File

@ -1,4 +1,4 @@
/* Copyright (C) 2002, 2003 Free Software Foundation, Inc. /* Copyright (C) 2002, 2003, 2004 Free Software Foundation, Inc.
This file is part of the GNU C Library. This file is part of the GNU C Library.
Contributed by Ulrich Drepper <drepper@redhat.com>, 2002. Contributed by Ulrich Drepper <drepper@redhat.com>, 2002.
@ -56,13 +56,14 @@ __pthread_cond_signal:
#endif #endif
jnz 1f jnz 1f
2: addq $wakeup_seq, %rdi 2: addq $cond_futex, %rdi
movq total_seq(%r8), %rcx movq total_seq(%r8), %rcx
cmpq (%rdi), %rcx cmpq wakeup_seq(%r8), %rcx
jbe 4f jbe 4f
/* Bump the wakeup number. */ /* Bump the wakeup number. */
addq $1, (%rdi) addq $1, wakeup_seq(%r8)
addl $1, (%rdi)
/* Wake up one thread. */ /* Wake up one thread. */
movq $FUTEX_WAKE, %rsi movq $FUTEX_WAKE, %rsi
@ -95,11 +96,7 @@ __pthread_cond_signal:
/* Unlock in loop requires wakeup. */ /* Unlock in loop requires wakeup. */
5: 5:
#if cond_lock != 0
addq $cond_lock-wakeup_seq, %rdi
#else
movq %r8, %rdi movq %r8, %rdi
#endif
callq __lll_mutex_unlock_wake callq __lll_mutex_unlock_wake
jmp 6b jmp 6b
.size __pthread_cond_signal, .-__pthread_cond_signal .size __pthread_cond_signal, .-__pthread_cond_signal

View File

@ -105,6 +105,7 @@ __pthread_cond_timedwait:
movq 8(%rsp), %rdi movq 8(%rsp), %rdi
incq total_seq(%rdi) incq total_seq(%rdi)
incl cond_futex(%rdi)
/* Install cancellation handler. */ /* Install cancellation handler. */
#ifdef PIC #ifdef PIC
@ -118,9 +119,9 @@ __pthread_cond_timedwait:
/* Get and store current wakeup_seq value. */ /* Get and store current wakeup_seq value. */
movq 8(%rsp), %rdi movq 8(%rsp), %rdi
movq wakeup_seq(%rdi), %r12 movq wakeup_seq(%rdi), %r9
movl broadcast_seq(%rdi), %edx movl broadcast_seq(%rdi), %edx
movq %r12, 40(%rsp) movq %r9, 40(%rsp)
movl %edx, 4(%rsp) movl %edx, 4(%rsp)
/* Get the current time. */ /* Get the current time. */
@ -172,6 +173,8 @@ __pthread_cond_timedwait:
21: movq %rcx, 24(%rsp) 21: movq %rcx, 24(%rsp)
movq %rdx, 32(%rsp) movq %rdx, 32(%rsp)
movl cond_futex(%rdi), %r12d
/* Unlock. */ /* Unlock. */
LOCK LOCK
#if cond_lock == 0 #if cond_lock == 0
@ -187,7 +190,7 @@ __pthread_cond_timedwait:
leaq 24(%rsp), %r10 leaq 24(%rsp), %r10
xorq %rsi, %rsi /* movq $FUTEX_WAIT, %rsi */ xorq %rsi, %rsi /* movq $FUTEX_WAIT, %rsi */
movq %r12, %rdx movq %r12, %rdx
addq $wakeup_seq-cond_lock, %rdi addq $cond_futex, %rdi
movq $SYS_futex, %rax movq $SYS_futex, %rax
syscall syscall
movq %rax, %r14 movq %rax, %r14
@ -211,21 +214,22 @@ __pthread_cond_timedwait:
movq woken_seq(%rdi), %rax movq woken_seq(%rdi), %rax
movq wakeup_seq(%rdi), %r12 movq wakeup_seq(%rdi), %r9
cmpl 4(%rsp), %edx cmpl 4(%rsp), %edx
jne 23f jne 23f
cmpq 40(%rsp), %r12 cmpq 40(%rsp), %r9
jbe 15f jbe 15f
cmpq %rax, %r12 cmpq %rax, %r9
ja 9f ja 9f
15: cmpq $-ETIMEDOUT, %r14 15: cmpq $-ETIMEDOUT, %r14
jne 8b jne 8b
13: incq wakeup_seq(%rdi) 13: incq wakeup_seq(%rdi)
incl cond_futex(%rdi)
movq $ETIMEDOUT, %r14 movq $ETIMEDOUT, %r14
jmp 14f jmp 14f

View File

@ -1,4 +1,4 @@
/* Copyright (C) 2002, 2003 Free Software Foundation, Inc. /* Copyright (C) 2002, 2003, 2004 Free Software Foundation, Inc.
This file is part of the GNU C Library. This file is part of the GNU C Library.
Contributed by Ulrich Drepper <drepper@redhat.com>, 2002. Contributed by Ulrich Drepper <drepper@redhat.com>, 2002.
@ -69,6 +69,8 @@ __condvar_cleanup:
incq woken_seq(%rdi) incq woken_seq(%rdi)
incl cond_futex(%rdi)
3: LOCK 3: LOCK
#if cond_lock == 0 #if cond_lock == 0
decl (%rdi) decl (%rdi)
@ -82,7 +84,7 @@ __condvar_cleanup:
callq __lll_mutex_unlock_wake callq __lll_mutex_unlock_wake
/* Wake up all waiters to make sure no signal gets lost. */ /* Wake up all waiters to make sure no signal gets lost. */
2: addq $wakeup_seq, %rdi 2: addq $cond_futex, %rdi
movq $FUTEX_WAKE, %rsi movq $FUTEX_WAKE, %rsi
movl $0x7fffffff, %edx movl $0x7fffffff, %edx
movq $SYS_futex, %rax movq $SYS_futex, %rax
@ -154,6 +156,7 @@ __pthread_cond_wait:
movq 8(%rsp), %rdi movq 8(%rsp), %rdi
incq total_seq(%rdi) incq total_seq(%rdi)
incl cond_futex(%rdi)
/* Install cancellation handler. */ /* Install cancellation handler. */
#ifdef PIC #ifdef PIC
@ -167,13 +170,14 @@ __pthread_cond_wait:
/* Get and store current wakeup_seq value. */ /* Get and store current wakeup_seq value. */
movq 8(%rsp), %rdi movq 8(%rsp), %rdi
movq wakeup_seq(%rdi), %r12 movq wakeup_seq(%rdi), %r9
movl broadcast_seq(%rdi), %edx movl broadcast_seq(%rdi), %edx
movq %r12, 24(%rsp) movq %r9, 24(%rsp)
movl %edx, 4(%rsp) movl %edx, 4(%rsp)
/* Unlock. */ /* Unlock. */
8: LOCK 8: movl cond_futex(%rdi), %r12d
LOCK
#if cond_lock == 0 #if cond_lock == 0
decl (%rdi) decl (%rdi)
#else #else
@ -187,7 +191,7 @@ __pthread_cond_wait:
movq 8(%rsp), %rdi movq 8(%rsp), %rdi
xorq %r10, %r10 xorq %r10, %r10
movq %r12, %rdx movq %r12, %rdx
addq $wakeup_seq-cond_lock, %rdi addq $cond_futex-cond_lock, %rdi
movq $SYS_futex, %rax movq $SYS_futex, %rax
movq %r10, %rsi /* movq $FUTEX_WAIT, %rsi */ movq %r10, %rsi /* movq $FUTEX_WAIT, %rsi */
syscall syscall
@ -211,15 +215,15 @@ __pthread_cond_wait:
movq woken_seq(%rdi), %rax movq woken_seq(%rdi), %rax
movq wakeup_seq(%rdi), %r12 movq wakeup_seq(%rdi), %r9
cmpl 4(%rsp), %edx cmpl 4(%rsp), %edx
jne 16f jne 16f
cmpq 24(%rsp), %r12 cmpq 24(%rsp), %r9
jbe 8b jbe 8b
cmpq %rax, %r12 cmpq %rax, %r9
jna 8b jna 8b
incq woken_seq(%rdi) incq woken_seq(%rdi)
@ -359,8 +363,8 @@ versioned_symbol (libpthread, __pthread_cond_wait, pthread_cond_wait,
.byte 0x40+.Lsubq-.Lpush_r12 # DW_CFA_advance_loc+N .byte 0x40+.Lsubq-.Lpush_r12 # DW_CFA_advance_loc+N
.byte 14 # DW_CFA_def_cfa_offset .byte 14 # DW_CFA_def_cfa_offset
.uleb128 16+FRAME_SIZE .uleb128 16+FRAME_SIZE
.byte 2 # DW_CFA_advance_loc1 .byte 3 # DW_CFA_advance_loc2
.byte .Laddq-.Lsubq .2byte .Laddq-.Lsubq
.byte 14 # DW_CFA_def_cfa_offset .byte 14 # DW_CFA_def_cfa_offset
.uleb128 16 .uleb128 16
.byte 0x40+.Lpop_r12-.Laddq # DW_CFA_advance_loc+N .byte 0x40+.Lpop_r12-.Laddq # DW_CFA_advance_loc+N

View File

@ -140,30 +140,31 @@ typedef uintmax_t uatomic_max_t;
#define atomic_exchange_and_add(mem, value) \ #define atomic_exchange_and_add(mem, value) \
({ __typeof (*mem) result; \ ({ __typeof (*mem) __result; \
__typeof (value) addval = (value); \ __typeof (value) __addval = (value); \
if (sizeof (*mem) == 1) \ if (sizeof (*mem) == 1) \
__asm __volatile (LOCK_PREFIX "xaddb %b0, %1" \ __asm __volatile (LOCK_PREFIX "xaddb %b0, %1" \
: "=r" (result), "=m" (*mem) \ : "=r" (__result), "=m" (*mem) \
: "0" (addval), "m" (*mem)); \ : "0" (__addval), "m" (*mem)); \
else if (sizeof (*mem) == 2) \ else if (sizeof (*mem) == 2) \
__asm __volatile (LOCK_PREFIX "xaddw %w0, %1" \ __asm __volatile (LOCK_PREFIX "xaddw %w0, %1" \
: "=r" (result), "=m" (*mem) \ : "=r" (__result), "=m" (*mem) \
: "0" (addval), "m" (*mem)); \ : "0" (__addval), "m" (*mem)); \
else if (sizeof (*mem) == 4) \ else if (sizeof (*mem) == 4) \
__asm __volatile (LOCK_PREFIX "xaddl %0, %1" \ __asm __volatile (LOCK_PREFIX "xaddl %0, %1" \
: "=r" (result), "=m" (*mem) \ : "=r" (__result), "=m" (*mem) \
: "0" (addval), "m" (*mem)); \ : "0" (__addval), "m" (*mem)); \
else \ else \
{ \ { \
__typeof (mem) memp = (mem); \ __typeof (mem) __memp = (mem); \
__typeof (*mem) __tmpval; \
__result = *__memp; \
do \ do \
result = *memp; \ __tmpval = __result; \
while (__arch_compare_and_exchange_val_64_acq (memp, \ while ((__result = __arch_compare_and_exchange_val_64_acq \
result + addval, \ (__memp, __result + __addval, __result)) == __tmpval); \
result) == result); \
} \ } \
result; }) __result; })
#define atomic_add(mem, value) \ #define atomic_add(mem, value) \
@ -185,13 +186,14 @@ typedef uintmax_t uatomic_max_t;
: "ir" (value), "m" (*mem)); \ : "ir" (value), "m" (*mem)); \
else \ else \
{ \ { \
__typeof (value) addval = (value); \ __typeof (value) __addval = (value); \
__typeof (*mem) oldval; \ __typeof (mem) __memp = (mem); \
__typeof (mem) memp = (mem); \ __typeof (*mem) __oldval = *__memp; \
__typeof (*mem) __tmpval; \
do \ do \
oldval = *memp; \ __tmpval = __oldval; \
while (__arch_compare_and_exchange_val_64_acq \ while ((__oldval = __arch_compare_and_exchange_val_64_acq \
(memp, oldval + addval, oldval) == oldval); \ (__memp, __oldval + __addval, __oldval)) == __tmpval); \
} \ } \
}) })
@ -249,12 +251,13 @@ typedef uintmax_t uatomic_max_t;
: "m" (*mem)); \ : "m" (*mem)); \
else \ else \
{ \ { \
__typeof (*mem) oldval; \ __typeof (mem) __memp = (mem); \
__typeof (mem) memp = (mem); \ __typeof (*mem) __oldval = *__memp; \
__typeof (*mem) __tmpval; \
do \ do \
oldval = *memp; \ __tmpval = __oldval; \
while (__arch_compare_and_exchange_val_64_acq \ while ((__oldval = __arch_compare_and_exchange_val_64_acq \
(memp, oldval + 1, oldval) == oldval); \ (__memp, __oldval + 1, __oldval)) == __tmpval); \
} \ } \
}) })
@ -293,12 +296,13 @@ typedef uintmax_t uatomic_max_t;
: "m" (*mem)); \ : "m" (*mem)); \
else \ else \
{ \ { \
__typeof (*mem) oldval; \ __typeof (mem) __memp = (mem); \
__typeof (mem) memp = (mem); \ __typeof (*mem) __oldval = *__memp; \
__typeof (*mem) __tmpval; \
do \ do \
oldval = *memp; \ __tmpval = __oldval; \
while (__arch_compare_and_exchange_val_64_acq \ while ((__oldval = __arch_compare_and_exchange_val_64_acq \
(memp, oldval - 1, oldval) == oldval); \ (__memp, __oldval - 1, __oldval)) == __tmpval); \
} \ } \
}) })