e51deae7f6
Use __asm __volatile (__lll_acq_instr ::: "memory") instead of atomic_full_barrier. 2007-07-31 Jakub Jelinek <jakub@redhat.com> * allocatestack.c (stack_cache_lock): Change type to int. (get_cached_stack, allocate_stack, __deallocate_stack, __make_stacks_executable, __find_thread_by_id, __nptl_setxid, __pthread_init_static_tls, __wait_lookup_done): Add LLL_PRIVATE as second argument to lll_lock and lll_unlock macros on stack_cache_lock. * pthread_create.c (__find_in_stack_list): Likewise. (start_thread): Similarly with pd->lock. Use lll_robust_dead macro instead of lll_robust_mutex_dead, pass LLL_SHARED to it as second argument. * descr.h (struct pthread): Change lock and setxid_futex field type to int. * old_pthread_cond_broadcast.c (__pthread_cond_broadcast_2_0): Use LLL_LOCK_INITIALIZER instead of LLL_MUTEX_LOCK_INITIALIZER. * old_pthread_cond_signal.c (__pthread_cond_signal_2_0): Likewise. * old_pthread_cond_timedwait.c (__pthread_cond_timedwait_2_0): Likewise. * old_pthread_cond_wait.c (__pthread_cond_wait_2_0): Likewise. * pthread_cond_init.c (__pthread_cond_init): Likewise. * pthreadP.h (__attr_list_lock): Change type to int. * pthread_attr_init.c (__attr_list_lock): Likewise. * pthread_barrier_destroy.c (pthread_barrier_destroy): Pass ibarrier->private ^ FUTEX_PRIVATE_FLAG as second argument to lll_{,un}lock. * pthread_barrier_wait.c (pthread_barrier_wait): Likewise and also for lll_futex_{wake,wait}. * pthread_barrier_init.c (pthread_barrier_init): Make iattr a pointer to const. * pthread_cond_broadcast.c (__pthread_cond_broadcast): Pass LLL_SHARED as second argument to lll_{,un}lock. * pthread_cond_destroy.c (__pthread_cond_destroy): Likewise. * pthread_cond_signal.c (__pthread_cond_singal): Likewise. * pthread_cond_timedwait.c (__pthread_cond_timedwait): Likewise. * pthread_cond_wait.c (__condvar_cleanup, __pthread_cond_wait): Likewise. * pthread_getattr_np.c (pthread_getattr_np): Add LLL_PRIVATE as second argument to lll_{,un}lock macros on pd->lock. * pthread_getschedparam.c (__pthread_getschedparam): Likewise. * pthread_setschedparam.c (__pthread_setschedparam): Likewise. * pthread_setschedprio.c (pthread_setschedprio): Likewise. * tpp.c (__pthread_tpp_change_priority, __pthread_current_priority): Likewise. * sysdeps/pthread/createthread.c (do_clone, create_thread): Likewise. * pthread_once.c (once_lock): Change type to int. (__pthread_once): Pass LLL_PRIVATE as second argument to lll_{,un}lock macros on once_lock. * pthread_rwlock_rdlock.c (__pthread_rwlock_rdlock): Use lll_{,un}lock macros instead of lll_mutex_{,un}lock, pass rwlock->__data.__shared as second argument to them and similarly for lll_futex_w*. * pthread_rwlock_timedrdlock.c (pthread_rwlock_timedrdlock): Likewise. * pthread_rwlock_timedwrlock.c (pthread_rwlock_timedwrlock): Likewise. * pthread_rwlock_tryrdlock.c (__pthread_rwlock_tryrdlock): Likewise. * pthread_rwlock_trywrlock.c (__pthread_rwlock_trywrlock): Likewise. * pthread_rwlock_unlock.c (__pthread_rwlock_unlock): Likewise. * pthread_rwlock_wrlock.c (__pthread_rwlock_wrlock): Likewise. * sem_close.c (sem_close): Pass LLL_PRIVATE as second argument to lll_{,un}lock macros on __sem_mappings_lock. * sem_open.c (check_add_mapping): Likewise. (__sem_mappings_lock): Change type to int. * semaphoreP.h (__sem_mappings_lock): Likewise. * pthread_mutex_lock.c (LLL_MUTEX_LOCK, LLL_MUTEX_TRYLOCK, LLL_ROBUST_MUTEX_LOCK): Use lll_{,try,robust_}lock macros instead of lll_*mutex_*, pass LLL_SHARED as last argument. (__pthread_mutex_lock): Use lll_unlock instead of lll_mutex_unlock, pass LLL_SHARED as last argument. * sysdeps/unix/sysv/linux/pthread_mutex_cond_lock.c (LLL_MUTEX_LOCK, LLL_MUTEX_TRYLOCK, LLL_ROBUST_MUTEX_LOCK): Use lll_{cond_,cond_try,robust_cond}lock macros instead of lll_*mutex_*, pass LLL_SHARED as last argument. * pthread_mutex_timedlock.c (pthread_mutex_timedlock): Use lll_{timed,try,robust_timed,un}lock instead of lll_*mutex*, pass LLL_SHARED as last argument. * pthread_mutex_trylock.c (__pthread_mutex_trylock): Similarly. * pthread_mutex_unlock.c (__pthread_mutex_unlock_usercnt): Similarly. * sysdeps/pthread/bits/libc-lock.h (__libc_lock_lock, __libc_lock_lock_recursive, __libc_lock_unlock, __libc_lock_unlock_recursive): Pass LLL_PRIVATE as second argument to lll_{,un}lock. * sysdeps/pthread/bits/stdio-lock.h (_IO_lock_lock, _IO_lock_unlock): Likewise. * sysdeps/unix/sysv/linux/fork.c (__libc_fork): Don't use compound literal. * sysdeps/unix/sysv/linux/unregister-atfork.c (__unregister_atfork): Pass LLL_PRIVATE as second argument to lll_{,un}lock macros on __fork_lock. * sysdeps/unix/sysv/linux/register-atfork.c (__register_atfork, free_mem): Likewise. (__fork_lock): Change type to int. * sysdeps/unix/sysv/linux/fork.h (__fork_lock): Likewise. * sysdeps/unix/sysv/linux/sem_post.c (__new_sem_post): Pass isem->private ^ FUTEX_PRIVATE_FLAG as second argument to lll_futex_wake. * sysdeps/unix/sysv/linux/sem_timedwait.c (sem_timedwait): Likewise. * sysdeps/unix/sysv/linux/sem_wait.c (__new_sem_wait): Likewise. * sysdeps/unix/sysv/linux/lowlevellock.c (__lll_lock_wait_private): New function. (__lll_lock_wait, __lll_timedlock_wait): Add private argument and pass it through to lll_futex_*wait, only compile in when IS_IN_libpthread. * sysdeps/unix/sysv/linux/lowlevelrobustlock.c (__lll_robust_lock_wait, __lll_robust_timedlock_wait): Add private argument and pass it through to lll_futex_*wait. * sysdeps/unix/sysv/linux/alpha/lowlevellock.h: Renamed all lll_mutex_* resp. lll_robust_mutex_* macros to lll_* resp. lll_robust_*. Renamed all __lll_mutex_* resp. __lll_robust_mutex_* inline functions to __lll_* resp. __lll_robust_*. (LLL_MUTEX_LOCK_INITIALIZER): Remove. (lll_mutex_dead): Add private argument. (__lll_lock_wait_private): New prototype. (__lll_lock_wait, __lll_robust_lock_wait, __lll_lock_timedwait, __lll_robust_lock_timedwait): Add private argument to prototypes. (__lll_lock): Add private argument, if it is constant LLL_PRIVATE, call __lll_lock_wait_private, otherwise pass private to __lll_lock_wait. (__lll_robust_lock, __lll_cond_lock, __lll_timedlock, __lll_robust_timedlock): Add private argument, pass it to __lll_*wait functions. (__lll_unlock): Add private argument, if it is constant LLL_PRIVATE, call __lll_unlock_wake_private, otherwise pass private to __lll_unlock_wake. (__lll_robust_unlock): Add private argument, pass it to __lll_robust_unlock_wake. (lll_lock, lll_robust_lock, lll_cond_lock, lll_timedlock, lll_robust_timedlock, lll_unlock, lll_robust_unlock): Add private argument, pass it through to __lll_* inline function. (__lll_mutex_unlock_force, lll_mutex_unlock_force): Remove. (lll_lock_t): Remove. (__lll_cond_wait, __lll_cond_timedwait, __lll_cond_wake, __lll_cond_broadcast, lll_cond_wait, lll_cond_timedwait, lll_cond_wake, lll_cond_broadcast): Remove. * sysdeps/unix/sysv/linux/ia64/lowlevellock.h: Likewise. * sysdeps/unix/sysv/linux/powerpc/lowlevellock.h: Likewise. * sysdeps/unix/sysv/linux/s390/lowlevellock.h: Likewise. * sysdeps/unix/sysv/linux/sparc/lowlevellock.h: Likewise. * sysdeps/unix/sysv/linux/i386/lowlevellock.h: Allow including the header from assembler. Renamed all lll_mutex_* resp. lll_robust_mutex_* macros to lll_* resp. lll_robust_*. (LOCK, FUTEX_CMP_REQUEUE, FUTEX_WAKE_OP, FUTEX_OP_CLEAR_WAKE_IF_GT_ONE): Define. (LLL_MUTEX_LOCK_INITIALIZER, LLL_MUTEX_LOCK_INITIALIZER_LOCKED, LLL_MUTEX_LOCK_INITIALIZER_WAITERS): Remove. (__lll_mutex_lock_wait, __lll_mutex_timedlock_wait, __lll_mutex_unlock_wake, __lll_lock_wait, __lll_unlock_wake): Remove prototype. (__lll_trylock_asm, __lll_lock_asm_start, __lll_unlock_asm): Define. (lll_robust_trylock, lll_cond_trylock): Use LLL_LOCK_INITIALIZER* rather than LLL_MUTEX_LOCK_INITIALIZER* macros. (lll_trylock): Likewise, use __lll_trylock_asm, pass MULTIPLE_THREADS_OFFSET as another asm operand. (lll_lock): Add private argument, use __lll_lock_asm_start, pass MULTIPLE_THREADS_OFFSET as last asm operand, call __lll_lock_wait_private if private is constant LLL_PRIVATE, otherwise pass private as another argument to __lll_lock_wait. (lll_robust_lock, lll_cond_lock, lll_robust_cond_lock, lll_timedlock, lll_robust_timedlock): Add private argument, pass private as another argument to __lll_*lock_wait call. (lll_unlock): Add private argument, use __lll_unlock_asm, pass MULTIPLE_THREADS_OFFSET as another asm operand, call __lll_unlock_wake_private if private is constant LLL_PRIVATE, otherwise pass private as another argument to __lll_unlock_wake. (lll_robust_unlock): Add private argument, pass private as another argument to __lll_unlock_wake. (lll_robust_dead): Add private argument, use __lll_private_flag macro. (lll_islocked): Use LLL_LOCK_INITIALIZER instead of LLL_MUTEX_LOCK_INITIALIZER. (lll_lock_t): Remove. (LLL_LOCK_INITIALIZER_WAITERS): Define. (__lll_cond_wait, __lll_cond_timedwait, __lll_cond_wake, __lll_cond_broadcast, lll_cond_wait, lll_cond_timedwait, lll_cond_wake, lll_cond_broadcast): Remove. * sysdeps/unix/sysv/linux/x86_64/lowlevellock.h: Likewise. * sysdeps/unix/sysv/linux/i386/i486/libc-lowlevellock.S: Revert 2007-05-2{3,9} changes. * sysdeps/unix/sysv/linux/i386/i486/lowlevellock.S: Include kernel-features.h and lowlevellock.h. (LOAD_PRIVATE_FUTEX_WAIT): Define. (LOAD_FUTEX_WAIT): Rewritten. (LOCK, SYS_gettimeofday, SYS_futex, FUTEX_WAIT, FUTEX_WAKE): Don't define. (__lll_lock_wait_private, __lll_unlock_wake_private): New functions. (__lll_mutex_lock_wait): Rename to ... (__lll_lock_wait): ... this. Take futex addr from %edx instead of %ecx, %ecx is now private argument. Don't compile in for libc.so. (__lll_mutex_timedlock_wait): Rename to ... (__lll_timedlock_wait): ... this. Use __NR_gettimeofday. %esi contains private argument. Don't compile in for libc.so. (__lll_mutex_unlock_wake): Rename to ... (__lll_unlock_wake): ... this. %ecx contains private argument. Don't compile in for libc.so. (__lll_timedwait_tid): Use __NR_gettimeofday. * sysdeps/unix/sysv/linux/i386/i486/lowlevelrobustlock.S: Include kernel-features.h and lowlevellock.h. (LOAD_FUTEX_WAIT): Define. (LOCK, SYS_gettimeofday, SYS_futex, FUTEX_WAIT, FUTEX_WAKE): Don't define. (__lll_robust_mutex_lock_wait): Rename to ... (__lll_robust_lock_wait): ... this. Futex addr is now in %edx argument, %ecx argument contains private. Use LOAD_FUTEX_WAIT macro. (__lll_robust_mutex_timedlock_wait): Rename to ... (__lll_robust_timedlock_wait): ... this. Use __NR_gettimeofday. %esi argument contains private, use LOAD_FUTEX_WAIT macro. * sysdeps/unix/sysv/linux/i386/i486/pthread_barrier_wait.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (pthread_barrier_wait): Rename __lll_mutex_* to __lll_*, pass PRIVATE(%ebx) ^ LLL_SHARED as private argument in %ecx to __lll_lock_wait and __lll_unlock_wake, pass MUTEX(%ebx) address to __lll_lock_wait in %edx. * sysdeps/unix/sysv/linux/i386/i486/pthread_cond_broadcast.S: Include lowlevellock.h and pthread-errnos.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_REQUEUE, FUTEX_CMP_REQUEUE, EINVAL, LOCK): Don't define. (__pthread_cond_broadcast): Rename __lll_mutex_* to __lll_*, pass cond_lock address in %edx rather than %ecx to __lll_lock_wait, pass LLL_SHARED in %ecx to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/i386/i486/pthread_cond_signal.S: Include lowlevellock.h and pthread-errnos.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_WAKE_OP, FUTEX_OP_CLEAR_WAKE_IF_GT_ONE, EINVAL, LOCK): Don't define. (__pthread_cond_signal): Rename __lll_mutex_* to __lll_*, pass cond_lock address in %edx rather than %ecx to __lll_lock_wait, pass LLL_SHARED in %ecx to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/i386/i486/pthread_cond_timedwait.S: Include lowlevellock.h. (SYS_futex, SYS_gettimeofday, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (__pthread_cond_timedwait): Rename __lll_mutex_* to __lll_*, pass cond_lock address in %edx rather than %ecx to __lll_lock_wait, pass LLL_SHARED in %ecx to both __lll_lock_wait and __lll_unlock_wake. Use __NR_gettimeofday. * sysdeps/unix/sysv/linux/i386/i486/pthread_cond_wait.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (__pthread_cond_wait, __condvar_w_cleanup): Rename __lll_mutex_* to __lll_*, pass cond_lock address in %edx rather than %ecx to __lll_lock_wait, pass LLL_SHARED in %ecx to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/i386/i486/pthread_rwlock_rdlock.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (__pthread_rwlock_rdlock): Rename __lll_mutex_* to __lll_*, pass MUTEX(%ebx) address in %edx rather than %ecx to __lll_lock_wait, pass PSHARED(%ebx) in %ecx to both __lll_lock_wait and __lll_unlock_wake. Move return value from %ecx to %edx register. * sysdeps/unix/sysv/linux/i386/i486/pthread_rwlock_timedrdlock.S: Include lowlevellock.h. (SYS_futex, SYS_gettimeofday, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (__pthread_rwlock_wrlock): Rename __lll_mutex_* to __lll_*, pass MUTEX(%ebp) address in %edx rather than %ecx to __lll_lock_wait, pass PSHARED(%ebp) in %ecx to both __lll_lock_wait and __lll_unlock_wake. Move return value from %ecx to %edx register. Use __NR_gettimeofday. * sysdeps/unix/sysv/linux/i386/i486/pthread_rwlock_timedwrlock.S: Include lowlevellock.h. (SYS_futex, SYS_gettimeofday, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (__pthread_rwlock_wrlock): Rename __lll_mutex_* to __lll_*, pass MUTEX(%ebp) address in %edx rather than %ecx to __lll_lock_wait, pass PSHARED(%ebp) in %ecx to both __lll_lock_wait and __lll_unlock_wake. Move return value from %ecx to %edx register. Use __NR_gettimeofday. * sysdeps/unix/sysv/linux/i386/i486/pthread_rwlock_unlock.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (__pthread_rwlock_unlock): Rename __lll_mutex_* to __lll_*, pass MUTEX(%edi) address in %edx rather than %ecx to __lll_lock_wait, pass PSHARED(%edi) in %ecx to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/i386/i486/pthread_rwlock_wrlock.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (__pthread_rwlock_wrlock): Rename __lll_mutex_* to __lll_*, pass MUTEX(%ebx) address in %edx rather than %ecx to __lll_lock_wait, pass PSHARED(%ebx) in %ecx to both __lll_lock_wait and __lll_unlock_wake. Move return value from %ecx to %edx register. * sysdeps/unix/sysv/linux/i386/pthread_once.S: Include lowlevellock.h. (LOCK, SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_PRIVATE_FLAG): Don't define. * sysdeps/unix/sysv/linux/i386/i486/sem_post.S: Include lowlevellock.h. (LOCK, SYS_futex, FUTEX_WAKE): Don't define. * sysdeps/unix/sysv/linux/i386/i486/sem_timedwait.S: Include lowlevellock.h. (LOCK, SYS_futex, SYS_gettimeofday, FUTEX_WAIT): Don't define. (sem_timedwait): Use __NR_gettimeofday. * sysdeps/unix/sysv/linux/i386/i486/sem_trywait.S: Include lowlevellock.h. (LOCK): Don't define. * sysdeps/unix/sysv/linux/i386/i486/sem_wait.S: Include lowlevellock.h. (LOCK, SYS_futex, FUTEX_WAIT): Don't define. * sysdeps/unix/sysv/linux/powerpc/sem_post.c: Wake only when there are waiters. * sysdeps/unix/sysv/linux/x86_64/libc-lowlevellock.S: Revert 2007-05-2{3,9} changes. * sysdeps/unix/sysv/linux/x86_64/lowlevellock.S: Include kernel-features.h and lowlevellock.h. (LOAD_PRIVATE_FUTEX_WAIT): Define. (LOAD_FUTEX_WAIT): Rewritten. (LOCK, SYS_futex, FUTEX_WAIT, FUTEX_WAKE): Don't define. (__lll_lock_wait_private, __lll_unlock_wake_private): New functions. (__lll_mutex_lock_wait): Rename to ... (__lll_lock_wait): ... this. %esi is now private argument. Don't compile in for libc.so. (__lll_mutex_timedlock_wait): Rename to ... (__lll_timedlock_wait): ... this. %esi contains private argument. Don't compile in for libc.so. (__lll_mutex_unlock_wake): Rename to ... (__lll_unlock_wake): ... this. %esi contains private argument. Don't compile in for libc.so. * sysdeps/unix/sysv/linux/x86_64/lowlevelrobustlock.S: Include kernel-features.h and lowlevellock.h. (LOAD_FUTEX_WAIT): Define. (LOCK, SYS_futex, FUTEX_WAIT, FUTEX_WAKE): Don't define. (__lll_robust_mutex_lock_wait): Rename to ... (__lll_robust_lock_wait): ... this. %esi argument contains private. Use LOAD_FUTEX_WAIT macro. (__lll_robust_mutex_timedlock_wait): Rename to ... (__lll_robust_timedlock_wait): ... this. %esi argument contains private, use LOAD_FUTEX_WAIT macro. * sysdeps/unix/sysv/linux/x86_64/pthread_barrier_wait.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (pthread_barrier_wait): Rename __lll_mutex_* to __lll_*, pass PRIVATE(%rdi) ^ LLL_SHARED as private argument in %esi to __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_cond_broadcast.S: Include lowlevellock.h and pthread-errnos.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_REQUEUE, FUTEX_CMP_REQUEUE, EINVAL, LOCK): Don't define. (__pthread_cond_broadcast): Rename __lll_mutex_* to __lll_*, pass LLL_SHARED in %esi to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_cond_signal.S: Include lowlevellock.h and pthread-errnos.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_WAKE_OP, FUTEX_OP_CLEAR_WAKE_IF_GT_ONE, EINVAL, LOCK): Don't define. (__pthread_cond_signal): Rename __lll_mutex_* to __lll_*, pass LLL_SHARED in %esi to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_cond_timedwait.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (__pthread_cond_timedwait): Rename __lll_mutex_* to __lll_*, pass LLL_SHARED in %esi to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_cond_wait.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, LOCK): Don't define. (__pthread_cond_wait, __condvar_cleanup): Rename __lll_mutex_* to __lll_*, pass LLL_SHARED in %esi to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_rwlock_rdlock.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_PRIVATE_FLAG, LOCK): Don't define. (__pthread_rwlock_rdlock): Rename __lll_mutex_* to __lll_*, pass PSHARED(%rdi) in %esi to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_rwlock_timedrdlock.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_PRIVATE_FLAG, LOCK): Don't define. (__pthread_rwlock_wrlock): Rename __lll_mutex_* to __lll_*, pass PSHARED(%rdi) in %esi to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_rwlock_timedwrlock.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_PRIVATE_FLAG, LOCK): Don't define. (__pthread_rwlock_wrlock): Rename __lll_mutex_* to __lll_*, pass PSHARED(%rdi) in %esi to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_rwlock_unlock.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_PRIVATE_FLAG, LOCK): Don't define. (__pthread_rwlock_unlock): Rename __lll_mutex_* to __lll_*, pass PSHARED(%rdi) in %esi to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_rwlock_wrlock.S: Include lowlevellock.h. (SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_PRIVATE_FLAG, LOCK): Don't define. (__pthread_rwlock_wrlock): Rename __lll_mutex_* to __lll_*, pass PSHARED(%rdi) in %ecx to both __lll_lock_wait and __lll_unlock_wake. * sysdeps/unix/sysv/linux/x86_64/pthread_once.S: Include lowlevellock.h. (LOCK, SYS_futex, FUTEX_WAIT, FUTEX_WAKE, FUTEX_PRIVATE_FLAG): Don't define. * sysdeps/unix/sysv/linux/x86_64/sem_post.S: Include lowlevellock.h. (LOCK, SYS_futex, FUTEX_WAKE): Don't define. * sysdeps/unix/sysv/linux/x86_64/sem_timedwait.S: Include lowlevellock.h. (LOCK, SYS_futex, FUTEX_WAIT): Don't define. * sysdeps/unix/sysv/linux/x86_64/sem_trywait.S: Include lowlevellock.h. (LOCK): Don't define. * sysdeps/unix/sysv/linux/x86_64/sem_wait.S: Include lowlevellock.h. (LOCK, SYS_futex, FUTEX_WAIT): Don't define. * sysdeps/unix/sysv/linux/sparc/internaltypes.h: New file. * sysdeps/unix/sysv/linux/sparc/pthread_barrier_destroy.c: New file. * sysdeps/unix/sysv/linux/sparc/pthread_barrier_init.c: New file. * sysdeps/unix/sysv/linux/sparc/pthread_barrier_wait.c: New file. * sysdeps/unix/sysv/linux/sparc/sparc32/lowlevellock.c (__lll_lock_wait_private): New function. (__lll_lock_wait, __lll_timedlock_wait): Add private argument, pass it to lll_futex_*wait. Don't compile in for libc.so. * sysdeps/unix/sysv/linux/sparc/sparc32/pthread_barrier_init.c: Remove. * sysdeps/unix/sysv/linux/sparc/sparc32/pthread_barrier_wait.c (struct sparc_pthread_barrier): Remove. (pthread_barrier_wait): Use union sparc_pthread_barrier instead of struct sparc_pthread_barrier. Pass ibarrier->s.pshared ? LLL_SHARED : LLL_PRIVATE to lll_{,un}lock and lll_futex_wait macros. * sysdeps/unix/sysv/linux/sparc/sparc32/sparcv9/pthread_barrier_init.c: Remove. * sysdeps/unix/sysv/linux/sparc/sparc32/sparcv9/pthread_barrier_wait.c: Include sparc pthread_barrier_wait.c instead of generic one.
1068 lines
30 KiB
C
1068 lines
30 KiB
C
/* Copyright (C) 2002,2003,2004,2005,2006,2007 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
Contributed by Ulrich Drepper <drepper@redhat.com>, 2002.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, write to the Free
|
|
Software Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA
|
|
02111-1307 USA. */
|
|
|
|
#include <assert.h>
|
|
#include <errno.h>
|
|
#include <signal.h>
|
|
#include <stdint.h>
|
|
#include <string.h>
|
|
#include <unistd.h>
|
|
#include <sys/mman.h>
|
|
#include <sys/param.h>
|
|
#include <dl-sysdep.h>
|
|
#include <tls.h>
|
|
#include <lowlevellock.h>
|
|
|
|
|
|
#ifndef NEED_SEPARATE_REGISTER_STACK
|
|
|
|
/* Most architectures have exactly one stack pointer. Some have more. */
|
|
# define STACK_VARIABLES void *stackaddr = NULL
|
|
|
|
/* How to pass the values to the 'create_thread' function. */
|
|
# define STACK_VARIABLES_ARGS stackaddr
|
|
|
|
/* How to declare function which gets there parameters. */
|
|
# define STACK_VARIABLES_PARMS void *stackaddr
|
|
|
|
/* How to declare allocate_stack. */
|
|
# define ALLOCATE_STACK_PARMS void **stack
|
|
|
|
/* This is how the function is called. We do it this way to allow
|
|
other variants of the function to have more parameters. */
|
|
# define ALLOCATE_STACK(attr, pd) allocate_stack (attr, pd, &stackaddr)
|
|
|
|
#else
|
|
|
|
/* We need two stacks. The kernel will place them but we have to tell
|
|
the kernel about the size of the reserved address space. */
|
|
# define STACK_VARIABLES void *stackaddr = NULL; size_t stacksize = 0
|
|
|
|
/* How to pass the values to the 'create_thread' function. */
|
|
# define STACK_VARIABLES_ARGS stackaddr, stacksize
|
|
|
|
/* How to declare function which gets there parameters. */
|
|
# define STACK_VARIABLES_PARMS void *stackaddr, size_t stacksize
|
|
|
|
/* How to declare allocate_stack. */
|
|
# define ALLOCATE_STACK_PARMS void **stack, size_t *stacksize
|
|
|
|
/* This is how the function is called. We do it this way to allow
|
|
other variants of the function to have more parameters. */
|
|
# define ALLOCATE_STACK(attr, pd) \
|
|
allocate_stack (attr, pd, &stackaddr, &stacksize)
|
|
|
|
#endif
|
|
|
|
|
|
/* Default alignment of stack. */
|
|
#ifndef STACK_ALIGN
|
|
# define STACK_ALIGN __alignof__ (long double)
|
|
#endif
|
|
|
|
/* Default value for minimal stack size after allocating thread
|
|
descriptor and guard. */
|
|
#ifndef MINIMAL_REST_STACK
|
|
# define MINIMAL_REST_STACK 4096
|
|
#endif
|
|
|
|
|
|
/* Let the architecture add some flags to the mmap() call used to
|
|
allocate stacks. */
|
|
#ifndef ARCH_MAP_FLAGS
|
|
# define ARCH_MAP_FLAGS 0
|
|
#endif
|
|
|
|
/* This yields the pointer that TLS support code calls the thread pointer. */
|
|
#if TLS_TCB_AT_TP
|
|
# define TLS_TPADJ(pd) (pd)
|
|
#elif TLS_DTV_AT_TP
|
|
# define TLS_TPADJ(pd) ((struct pthread *)((char *) (pd) + TLS_PRE_TCB_SIZE))
|
|
#endif
|
|
|
|
/* Cache handling for not-yet free stacks. */
|
|
|
|
/* Maximum size in kB of cache. */
|
|
static size_t stack_cache_maxsize = 40 * 1024 * 1024; /* 40MiBi by default. */
|
|
static size_t stack_cache_actsize;
|
|
|
|
/* Mutex protecting this variable. */
|
|
static int stack_cache_lock = LLL_LOCK_INITIALIZER;
|
|
|
|
/* List of queued stack frames. */
|
|
static LIST_HEAD (stack_cache);
|
|
|
|
/* List of the stacks in use. */
|
|
static LIST_HEAD (stack_used);
|
|
|
|
/* List of the threads with user provided stacks in use. No need to
|
|
initialize this, since it's done in __pthread_initialize_minimal. */
|
|
list_t __stack_user __attribute__ ((nocommon));
|
|
hidden_data_def (__stack_user)
|
|
|
|
#if COLORING_INCREMENT != 0
|
|
/* Number of threads created. */
|
|
static unsigned int nptl_ncreated;
|
|
#endif
|
|
|
|
|
|
/* Check whether the stack is still used or not. */
|
|
#define FREE_P(descr) ((descr)->tid <= 0)
|
|
|
|
|
|
/* We create a double linked list of all cache entries. Double linked
|
|
because this allows removing entries from the end. */
|
|
|
|
|
|
/* Get a stack frame from the cache. We have to match by size since
|
|
some blocks might be too small or far too large. */
|
|
static struct pthread *
|
|
get_cached_stack (size_t *sizep, void **memp)
|
|
{
|
|
size_t size = *sizep;
|
|
struct pthread *result = NULL;
|
|
list_t *entry;
|
|
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
/* Search the cache for a matching entry. We search for the
|
|
smallest stack which has at least the required size. Note that
|
|
in normal situations the size of all allocated stacks is the
|
|
same. As the very least there are only a few different sizes.
|
|
Therefore this loop will exit early most of the time with an
|
|
exact match. */
|
|
list_for_each (entry, &stack_cache)
|
|
{
|
|
struct pthread *curr;
|
|
|
|
curr = list_entry (entry, struct pthread, list);
|
|
if (FREE_P (curr) && curr->stackblock_size >= size)
|
|
{
|
|
if (curr->stackblock_size == size)
|
|
{
|
|
result = curr;
|
|
break;
|
|
}
|
|
|
|
if (result == NULL
|
|
|| result->stackblock_size > curr->stackblock_size)
|
|
result = curr;
|
|
}
|
|
}
|
|
|
|
if (__builtin_expect (result == NULL, 0)
|
|
/* Make sure the size difference is not too excessive. In that
|
|
case we do not use the block. */
|
|
|| __builtin_expect (result->stackblock_size > 4 * size, 0))
|
|
{
|
|
/* Release the lock. */
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* Dequeue the entry. */
|
|
list_del (&result->list);
|
|
|
|
/* And add to the list of stacks in use. */
|
|
list_add (&result->list, &stack_used);
|
|
|
|
/* And decrease the cache size. */
|
|
stack_cache_actsize -= result->stackblock_size;
|
|
|
|
/* Release the lock early. */
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
/* Report size and location of the stack to the caller. */
|
|
*sizep = result->stackblock_size;
|
|
*memp = result->stackblock;
|
|
|
|
/* Cancellation handling is back to the default. */
|
|
result->cancelhandling = 0;
|
|
result->cleanup = NULL;
|
|
|
|
/* No pending event. */
|
|
result->nextevent = NULL;
|
|
|
|
/* Clear the DTV. */
|
|
dtv_t *dtv = GET_DTV (TLS_TPADJ (result));
|
|
memset (dtv, '\0', (dtv[-1].counter + 1) * sizeof (dtv_t));
|
|
|
|
/* Re-initialize the TLS. */
|
|
_dl_allocate_tls_init (TLS_TPADJ (result));
|
|
|
|
return result;
|
|
}
|
|
|
|
|
|
/* Free stacks until cache size is lower than LIMIT. */
|
|
static void
|
|
free_stacks (size_t limit)
|
|
{
|
|
/* We reduce the size of the cache. Remove the last entries until
|
|
the size is below the limit. */
|
|
list_t *entry;
|
|
list_t *prev;
|
|
|
|
/* Search from the end of the list. */
|
|
list_for_each_prev_safe (entry, prev, &stack_cache)
|
|
{
|
|
struct pthread *curr;
|
|
|
|
curr = list_entry (entry, struct pthread, list);
|
|
if (FREE_P (curr))
|
|
{
|
|
/* Unlink the block. */
|
|
list_del (entry);
|
|
|
|
/* Account for the freed memory. */
|
|
stack_cache_actsize -= curr->stackblock_size;
|
|
|
|
/* Free the memory associated with the ELF TLS. */
|
|
_dl_deallocate_tls (TLS_TPADJ (curr), false);
|
|
|
|
/* Remove this block. This should never fail. If it does
|
|
something is really wrong. */
|
|
if (munmap (curr->stackblock, curr->stackblock_size) != 0)
|
|
abort ();
|
|
|
|
/* Maybe we have freed enough. */
|
|
if (stack_cache_actsize <= limit)
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/* Add a stack frame which is not used anymore to the stack. Must be
|
|
called with the cache lock held. */
|
|
static inline void
|
|
__attribute ((always_inline))
|
|
queue_stack (struct pthread *stack)
|
|
{
|
|
/* We unconditionally add the stack to the list. The memory may
|
|
still be in use but it will not be reused until the kernel marks
|
|
the stack as not used anymore. */
|
|
list_add (&stack->list, &stack_cache);
|
|
|
|
stack_cache_actsize += stack->stackblock_size;
|
|
if (__builtin_expect (stack_cache_actsize > stack_cache_maxsize, 0))
|
|
free_stacks (stack_cache_maxsize);
|
|
}
|
|
|
|
|
|
/* This function is called indirectly from the freeres code in libc. */
|
|
void
|
|
__free_stack_cache (void)
|
|
{
|
|
free_stacks (0);
|
|
}
|
|
|
|
|
|
static int
|
|
internal_function
|
|
change_stack_perm (struct pthread *pd
|
|
#ifdef NEED_SEPARATE_REGISTER_STACK
|
|
, size_t pagemask
|
|
#endif
|
|
)
|
|
{
|
|
#ifdef NEED_SEPARATE_REGISTER_STACK
|
|
void *stack = (pd->stackblock
|
|
+ (((((pd->stackblock_size - pd->guardsize) / 2)
|
|
& pagemask) + pd->guardsize) & pagemask));
|
|
size_t len = pd->stackblock + pd->stackblock_size - stack;
|
|
#elif _STACK_GROWS_DOWN
|
|
void *stack = pd->stackblock + pd->guardsize;
|
|
size_t len = pd->stackblock_size - pd->guardsize;
|
|
#elif _STACK_GROWS_UP
|
|
void *stack = pd->stackblock;
|
|
size_t len = (uintptr_t) pd - pd->guardsize - (uintptr_t) pd->stackblock;
|
|
#else
|
|
# error "Define either _STACK_GROWS_DOWN or _STACK_GROWS_UP"
|
|
#endif
|
|
if (mprotect (stack, len, PROT_READ | PROT_WRITE | PROT_EXEC) != 0)
|
|
return errno;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int
|
|
allocate_stack (const struct pthread_attr *attr, struct pthread **pdp,
|
|
ALLOCATE_STACK_PARMS)
|
|
{
|
|
struct pthread *pd;
|
|
size_t size;
|
|
size_t pagesize_m1 = __getpagesize () - 1;
|
|
void *stacktop;
|
|
|
|
assert (attr != NULL);
|
|
assert (powerof2 (pagesize_m1 + 1));
|
|
assert (TCB_ALIGNMENT >= STACK_ALIGN);
|
|
|
|
/* Get the stack size from the attribute if it is set. Otherwise we
|
|
use the default we determined at start time. */
|
|
size = attr->stacksize ?: __default_stacksize;
|
|
|
|
/* Get memory for the stack. */
|
|
if (__builtin_expect (attr->flags & ATTR_FLAG_STACKADDR, 0))
|
|
{
|
|
uintptr_t adj;
|
|
|
|
/* If the user also specified the size of the stack make sure it
|
|
is large enough. */
|
|
if (attr->stacksize != 0
|
|
&& attr->stacksize < (__static_tls_size + MINIMAL_REST_STACK))
|
|
return EINVAL;
|
|
|
|
/* Adjust stack size for alignment of the TLS block. */
|
|
#if TLS_TCB_AT_TP
|
|
adj = ((uintptr_t) attr->stackaddr - TLS_TCB_SIZE)
|
|
& __static_tls_align_m1;
|
|
assert (size > adj + TLS_TCB_SIZE);
|
|
#elif TLS_DTV_AT_TP
|
|
adj = ((uintptr_t) attr->stackaddr - __static_tls_size)
|
|
& __static_tls_align_m1;
|
|
assert (size > adj);
|
|
#endif
|
|
|
|
/* The user provided some memory. Let's hope it matches the
|
|
size... We do not allocate guard pages if the user provided
|
|
the stack. It is the user's responsibility to do this if it
|
|
is wanted. */
|
|
#if TLS_TCB_AT_TP
|
|
pd = (struct pthread *) ((uintptr_t) attr->stackaddr
|
|
- TLS_TCB_SIZE - adj);
|
|
#elif TLS_DTV_AT_TP
|
|
pd = (struct pthread *) (((uintptr_t) attr->stackaddr
|
|
- __static_tls_size - adj)
|
|
- TLS_PRE_TCB_SIZE);
|
|
#endif
|
|
|
|
/* The user provided stack memory needs to be cleared. */
|
|
memset (pd, '\0', sizeof (struct pthread));
|
|
|
|
/* The first TSD block is included in the TCB. */
|
|
pd->specific[0] = pd->specific_1stblock;
|
|
|
|
/* Remember the stack-related values. */
|
|
pd->stackblock = (char *) attr->stackaddr - size;
|
|
pd->stackblock_size = size;
|
|
|
|
/* This is a user-provided stack. It will not be queued in the
|
|
stack cache nor will the memory (except the TLS memory) be freed. */
|
|
pd->user_stack = true;
|
|
|
|
/* This is at least the second thread. */
|
|
pd->header.multiple_threads = 1;
|
|
#ifndef TLS_MULTIPLE_THREADS_IN_TCB
|
|
__pthread_multiple_threads = *__libc_multiple_threads_ptr = 1;
|
|
#endif
|
|
|
|
#ifndef __ASSUME_PRIVATE_FUTEX
|
|
/* The thread must know when private futexes are supported. */
|
|
pd->header.private_futex = THREAD_GETMEM (THREAD_SELF,
|
|
header.private_futex);
|
|
#endif
|
|
|
|
#ifdef NEED_DL_SYSINFO
|
|
/* Copy the sysinfo value from the parent. */
|
|
THREAD_SYSINFO(pd) = THREAD_SELF_SYSINFO;
|
|
#endif
|
|
|
|
/* The process ID is also the same as that of the caller. */
|
|
pd->pid = THREAD_GETMEM (THREAD_SELF, pid);
|
|
|
|
/* Allocate the DTV for this thread. */
|
|
if (_dl_allocate_tls (TLS_TPADJ (pd)) == NULL)
|
|
{
|
|
/* Something went wrong. */
|
|
assert (errno == ENOMEM);
|
|
return EAGAIN;
|
|
}
|
|
|
|
|
|
/* Prepare to modify global data. */
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
/* And add to the list of stacks in use. */
|
|
list_add (&pd->list, &__stack_user);
|
|
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
}
|
|
else
|
|
{
|
|
/* Allocate some anonymous memory. If possible use the cache. */
|
|
size_t guardsize;
|
|
size_t reqsize;
|
|
void *mem;
|
|
const int prot = (PROT_READ | PROT_WRITE
|
|
| ((GL(dl_stack_flags) & PF_X) ? PROT_EXEC : 0));
|
|
|
|
#if COLORING_INCREMENT != 0
|
|
/* Add one more page for stack coloring. Don't do it for stacks
|
|
with 16 times pagesize or larger. This might just cause
|
|
unnecessary misalignment. */
|
|
if (size <= 16 * pagesize_m1)
|
|
size += pagesize_m1 + 1;
|
|
#endif
|
|
|
|
/* Adjust the stack size for alignment. */
|
|
size &= ~__static_tls_align_m1;
|
|
assert (size != 0);
|
|
|
|
/* Make sure the size of the stack is enough for the guard and
|
|
eventually the thread descriptor. */
|
|
guardsize = (attr->guardsize + pagesize_m1) & ~pagesize_m1;
|
|
if (__builtin_expect (size < ((guardsize + __static_tls_size
|
|
+ MINIMAL_REST_STACK + pagesize_m1)
|
|
& ~pagesize_m1),
|
|
0))
|
|
/* The stack is too small (or the guard too large). */
|
|
return EINVAL;
|
|
|
|
/* Try to get a stack from the cache. */
|
|
reqsize = size;
|
|
pd = get_cached_stack (&size, &mem);
|
|
if (pd == NULL)
|
|
{
|
|
/* To avoid aliasing effects on a larger scale than pages we
|
|
adjust the allocated stack size if necessary. This way
|
|
allocations directly following each other will not have
|
|
aliasing problems. */
|
|
#if MULTI_PAGE_ALIASING != 0
|
|
if ((size % MULTI_PAGE_ALIASING) == 0)
|
|
size += pagesize_m1 + 1;
|
|
#endif
|
|
|
|
mem = mmap (NULL, size, prot,
|
|
MAP_PRIVATE | MAP_ANONYMOUS | ARCH_MAP_FLAGS, -1, 0);
|
|
|
|
if (__builtin_expect (mem == MAP_FAILED, 0))
|
|
{
|
|
#ifdef ARCH_RETRY_MMAP
|
|
mem = ARCH_RETRY_MMAP (size);
|
|
if (__builtin_expect (mem == MAP_FAILED, 0))
|
|
#endif
|
|
return errno;
|
|
}
|
|
|
|
/* SIZE is guaranteed to be greater than zero.
|
|
So we can never get a null pointer back from mmap. */
|
|
assert (mem != NULL);
|
|
|
|
#if COLORING_INCREMENT != 0
|
|
/* Atomically increment NCREATED. */
|
|
unsigned int ncreated = atomic_increment_val (&nptl_ncreated);
|
|
|
|
/* We chose the offset for coloring by incrementing it for
|
|
every new thread by a fixed amount. The offset used
|
|
module the page size. Even if coloring would be better
|
|
relative to higher alignment values it makes no sense to
|
|
do it since the mmap() interface does not allow us to
|
|
specify any alignment for the returned memory block. */
|
|
size_t coloring = (ncreated * COLORING_INCREMENT) & pagesize_m1;
|
|
|
|
/* Make sure the coloring offsets does not disturb the alignment
|
|
of the TCB and static TLS block. */
|
|
if (__builtin_expect ((coloring & __static_tls_align_m1) != 0, 0))
|
|
coloring = (((coloring + __static_tls_align_m1)
|
|
& ~(__static_tls_align_m1))
|
|
& ~pagesize_m1);
|
|
#else
|
|
/* Unless specified we do not make any adjustments. */
|
|
# define coloring 0
|
|
#endif
|
|
|
|
/* Place the thread descriptor at the end of the stack. */
|
|
#if TLS_TCB_AT_TP
|
|
pd = (struct pthread *) ((char *) mem + size - coloring) - 1;
|
|
#elif TLS_DTV_AT_TP
|
|
pd = (struct pthread *) ((((uintptr_t) mem + size - coloring
|
|
- __static_tls_size)
|
|
& ~__static_tls_align_m1)
|
|
- TLS_PRE_TCB_SIZE);
|
|
#endif
|
|
|
|
/* Remember the stack-related values. */
|
|
pd->stackblock = mem;
|
|
pd->stackblock_size = size;
|
|
|
|
/* We allocated the first block thread-specific data array.
|
|
This address will not change for the lifetime of this
|
|
descriptor. */
|
|
pd->specific[0] = pd->specific_1stblock;
|
|
|
|
/* This is at least the second thread. */
|
|
pd->header.multiple_threads = 1;
|
|
#ifndef TLS_MULTIPLE_THREADS_IN_TCB
|
|
__pthread_multiple_threads = *__libc_multiple_threads_ptr = 1;
|
|
#endif
|
|
|
|
#ifndef __ASSUME_PRIVATE_FUTEX
|
|
/* The thread must know when private futexes are supported. */
|
|
pd->header.private_futex = THREAD_GETMEM (THREAD_SELF,
|
|
header.private_futex);
|
|
#endif
|
|
|
|
#ifdef NEED_DL_SYSINFO
|
|
/* Copy the sysinfo value from the parent. */
|
|
THREAD_SYSINFO(pd) = THREAD_SELF_SYSINFO;
|
|
#endif
|
|
|
|
/* The process ID is also the same as that of the caller. */
|
|
pd->pid = THREAD_GETMEM (THREAD_SELF, pid);
|
|
|
|
/* Allocate the DTV for this thread. */
|
|
if (_dl_allocate_tls (TLS_TPADJ (pd)) == NULL)
|
|
{
|
|
/* Something went wrong. */
|
|
assert (errno == ENOMEM);
|
|
|
|
/* Free the stack memory we just allocated. */
|
|
(void) munmap (mem, size);
|
|
|
|
return EAGAIN;
|
|
}
|
|
|
|
|
|
/* Prepare to modify global data. */
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
/* And add to the list of stacks in use. */
|
|
list_add (&pd->list, &stack_used);
|
|
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
|
|
/* There might have been a race. Another thread might have
|
|
caused the stacks to get exec permission while this new
|
|
stack was prepared. Detect if this was possible and
|
|
change the permission if necessary. */
|
|
if (__builtin_expect ((GL(dl_stack_flags) & PF_X) != 0
|
|
&& (prot & PROT_EXEC) == 0, 0))
|
|
{
|
|
int err = change_stack_perm (pd
|
|
#ifdef NEED_SEPARATE_REGISTER_STACK
|
|
, ~pagesize_m1
|
|
#endif
|
|
);
|
|
if (err != 0)
|
|
{
|
|
/* Free the stack memory we just allocated. */
|
|
(void) munmap (mem, size);
|
|
|
|
return err;
|
|
}
|
|
}
|
|
|
|
|
|
/* Note that all of the stack and the thread descriptor is
|
|
zeroed. This means we do not have to initialize fields
|
|
with initial value zero. This is specifically true for
|
|
the 'tid' field which is always set back to zero once the
|
|
stack is not used anymore and for the 'guardsize' field
|
|
which will be read next. */
|
|
}
|
|
|
|
/* Create or resize the guard area if necessary. */
|
|
if (__builtin_expect (guardsize > pd->guardsize, 0))
|
|
{
|
|
#ifdef NEED_SEPARATE_REGISTER_STACK
|
|
char *guard = mem + (((size - guardsize) / 2) & ~pagesize_m1);
|
|
#elif _STACK_GROWS_DOWN
|
|
char *guard = mem;
|
|
# elif _STACK_GROWS_UP
|
|
char *guard = (char *) (((uintptr_t) pd - guardsize) & ~pagesize_m1);
|
|
#endif
|
|
if (mprotect (guard, guardsize, PROT_NONE) != 0)
|
|
{
|
|
int err;
|
|
mprot_error:
|
|
err = errno;
|
|
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
/* Remove the thread from the list. */
|
|
list_del (&pd->list);
|
|
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
/* Get rid of the TLS block we allocated. */
|
|
_dl_deallocate_tls (TLS_TPADJ (pd), false);
|
|
|
|
/* Free the stack memory regardless of whether the size
|
|
of the cache is over the limit or not. If this piece
|
|
of memory caused problems we better do not use it
|
|
anymore. Uh, and we ignore possible errors. There
|
|
is nothing we could do. */
|
|
(void) munmap (mem, size);
|
|
|
|
return err;
|
|
}
|
|
|
|
pd->guardsize = guardsize;
|
|
}
|
|
else if (__builtin_expect (pd->guardsize - guardsize > size - reqsize,
|
|
0))
|
|
{
|
|
/* The old guard area is too large. */
|
|
|
|
#ifdef NEED_SEPARATE_REGISTER_STACK
|
|
char *guard = mem + (((size - guardsize) / 2) & ~pagesize_m1);
|
|
char *oldguard = mem + (((size - pd->guardsize) / 2) & ~pagesize_m1);
|
|
|
|
if (oldguard < guard
|
|
&& mprotect (oldguard, guard - oldguard, prot) != 0)
|
|
goto mprot_error;
|
|
|
|
if (mprotect (guard + guardsize,
|
|
oldguard + pd->guardsize - guard - guardsize,
|
|
prot) != 0)
|
|
goto mprot_error;
|
|
#elif _STACK_GROWS_DOWN
|
|
if (mprotect ((char *) mem + guardsize, pd->guardsize - guardsize,
|
|
prot) != 0)
|
|
goto mprot_error;
|
|
#elif _STACK_GROWS_UP
|
|
if (mprotect ((char *) pd - pd->guardsize,
|
|
pd->guardsize - guardsize, prot) != 0)
|
|
goto mprot_error;
|
|
#endif
|
|
|
|
pd->guardsize = guardsize;
|
|
}
|
|
/* The pthread_getattr_np() calls need to get passed the size
|
|
requested in the attribute, regardless of how large the
|
|
actually used guardsize is. */
|
|
pd->reported_guardsize = guardsize;
|
|
}
|
|
|
|
/* Initialize the lock. We have to do this unconditionally since the
|
|
stillborn thread could be canceled while the lock is taken. */
|
|
pd->lock = LLL_LOCK_INITIALIZER;
|
|
|
|
/* The robust mutex lists also need to be initialized
|
|
unconditionally because the cleanup for the previous stack owner
|
|
might have happened in the kernel. */
|
|
pd->robust_head.futex_offset = (offsetof (pthread_mutex_t, __data.__lock)
|
|
- offsetof (pthread_mutex_t,
|
|
__data.__list.__next));
|
|
pd->robust_head.list_op_pending = NULL;
|
|
#ifdef __PTHREAD_MUTEX_HAVE_PREV
|
|
pd->robust_prev = &pd->robust_head;
|
|
#endif
|
|
pd->robust_head.list = &pd->robust_head;
|
|
|
|
/* We place the thread descriptor at the end of the stack. */
|
|
*pdp = pd;
|
|
|
|
#if TLS_TCB_AT_TP
|
|
/* The stack begins before the TCB and the static TLS block. */
|
|
stacktop = ((char *) (pd + 1) - __static_tls_size);
|
|
#elif TLS_DTV_AT_TP
|
|
stacktop = (char *) (pd - 1);
|
|
#endif
|
|
|
|
#ifdef NEED_SEPARATE_REGISTER_STACK
|
|
*stack = pd->stackblock;
|
|
*stacksize = stacktop - *stack;
|
|
#elif _STACK_GROWS_DOWN
|
|
*stack = stacktop;
|
|
#elif _STACK_GROWS_UP
|
|
*stack = pd->stackblock;
|
|
assert (*stack > 0);
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
void
|
|
internal_function
|
|
__deallocate_stack (struct pthread *pd)
|
|
{
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
/* Remove the thread from the list of threads with user defined
|
|
stacks. */
|
|
list_del (&pd->list);
|
|
|
|
/* Not much to do. Just free the mmap()ed memory. Note that we do
|
|
not reset the 'used' flag in the 'tid' field. This is done by
|
|
the kernel. If no thread has been created yet this field is
|
|
still zero. */
|
|
if (__builtin_expect (! pd->user_stack, 1))
|
|
(void) queue_stack (pd);
|
|
else
|
|
/* Free the memory associated with the ELF TLS. */
|
|
_dl_deallocate_tls (TLS_TPADJ (pd), false);
|
|
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
}
|
|
|
|
|
|
int
|
|
internal_function
|
|
__make_stacks_executable (void **stack_endp)
|
|
{
|
|
/* First the main thread's stack. */
|
|
int err = _dl_make_stack_executable (stack_endp);
|
|
if (err != 0)
|
|
return err;
|
|
|
|
#ifdef NEED_SEPARATE_REGISTER_STACK
|
|
const size_t pagemask = ~(__getpagesize () - 1);
|
|
#endif
|
|
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
list_t *runp;
|
|
list_for_each (runp, &stack_used)
|
|
{
|
|
err = change_stack_perm (list_entry (runp, struct pthread, list)
|
|
#ifdef NEED_SEPARATE_REGISTER_STACK
|
|
, pagemask
|
|
#endif
|
|
);
|
|
if (err != 0)
|
|
break;
|
|
}
|
|
|
|
/* Also change the permission for the currently unused stacks. This
|
|
might be wasted time but better spend it here than adding a check
|
|
in the fast path. */
|
|
if (err == 0)
|
|
list_for_each (runp, &stack_cache)
|
|
{
|
|
err = change_stack_perm (list_entry (runp, struct pthread, list)
|
|
#ifdef NEED_SEPARATE_REGISTER_STACK
|
|
, pagemask
|
|
#endif
|
|
);
|
|
if (err != 0)
|
|
break;
|
|
}
|
|
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
return err;
|
|
}
|
|
|
|
|
|
/* In case of a fork() call the memory allocation in the child will be
|
|
the same but only one thread is running. All stacks except that of
|
|
the one running thread are not used anymore. We have to recycle
|
|
them. */
|
|
void
|
|
__reclaim_stacks (void)
|
|
{
|
|
struct pthread *self = (struct pthread *) THREAD_SELF;
|
|
|
|
/* No locking necessary. The caller is the only stack in use. */
|
|
|
|
/* Mark all stacks except the still running one as free. */
|
|
list_t *runp;
|
|
list_for_each (runp, &stack_used)
|
|
{
|
|
struct pthread *curp = list_entry (runp, struct pthread, list);
|
|
if (curp != self)
|
|
{
|
|
/* This marks the stack as free. */
|
|
curp->tid = 0;
|
|
|
|
/* The PID field must be initialized for the new process. */
|
|
curp->pid = self->pid;
|
|
|
|
/* Account for the size of the stack. */
|
|
stack_cache_actsize += curp->stackblock_size;
|
|
}
|
|
}
|
|
|
|
/* Reset the PIDs in any cached stacks. */
|
|
list_for_each (runp, &stack_cache)
|
|
{
|
|
struct pthread *curp = list_entry (runp, struct pthread, list);
|
|
curp->pid = self->pid;
|
|
}
|
|
|
|
/* Add the stack of all running threads to the cache. */
|
|
list_splice (&stack_used, &stack_cache);
|
|
|
|
/* Remove the entry for the current thread to from the cache list
|
|
and add it to the list of running threads. Which of the two
|
|
lists is decided by the user_stack flag. */
|
|
list_del (&self->list);
|
|
|
|
/* Re-initialize the lists for all the threads. */
|
|
INIT_LIST_HEAD (&stack_used);
|
|
INIT_LIST_HEAD (&__stack_user);
|
|
|
|
if (__builtin_expect (THREAD_GETMEM (self, user_stack), 0))
|
|
list_add (&self->list, &__stack_user);
|
|
else
|
|
list_add (&self->list, &stack_used);
|
|
|
|
/* There is one thread running. */
|
|
__nptl_nthreads = 1;
|
|
|
|
/* Initialize the lock. */
|
|
stack_cache_lock = LLL_LOCK_INITIALIZER;
|
|
}
|
|
|
|
|
|
#if HP_TIMING_AVAIL
|
|
# undef __find_thread_by_id
|
|
/* Find a thread given the thread ID. */
|
|
attribute_hidden
|
|
struct pthread *
|
|
__find_thread_by_id (pid_t tid)
|
|
{
|
|
struct pthread *result = NULL;
|
|
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
/* Iterate over the list with system-allocated threads first. */
|
|
list_t *runp;
|
|
list_for_each (runp, &stack_used)
|
|
{
|
|
struct pthread *curp;
|
|
|
|
curp = list_entry (runp, struct pthread, list);
|
|
|
|
if (curp->tid == tid)
|
|
{
|
|
result = curp;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
/* Now the list with threads using user-allocated stacks. */
|
|
list_for_each (runp, &__stack_user)
|
|
{
|
|
struct pthread *curp;
|
|
|
|
curp = list_entry (runp, struct pthread, list);
|
|
|
|
if (curp->tid == tid)
|
|
{
|
|
result = curp;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
out:
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
return result;
|
|
}
|
|
#endif
|
|
|
|
|
|
static void
|
|
internal_function
|
|
setxid_signal_thread (struct xid_command *cmdp, struct pthread *t)
|
|
{
|
|
if (! IS_DETACHED (t))
|
|
{
|
|
int ch;
|
|
do
|
|
{
|
|
ch = t->cancelhandling;
|
|
|
|
/* If the thread is exiting right now, ignore it. */
|
|
if ((ch & EXITING_BITMASK) != 0)
|
|
return;
|
|
}
|
|
while (atomic_compare_and_exchange_bool_acq (&t->cancelhandling,
|
|
ch | SETXID_BITMASK, ch));
|
|
}
|
|
|
|
int val;
|
|
INTERNAL_SYSCALL_DECL (err);
|
|
#if __ASSUME_TGKILL
|
|
val = INTERNAL_SYSCALL (tgkill, err, 3, THREAD_GETMEM (THREAD_SELF, pid),
|
|
t->tid, SIGSETXID);
|
|
#else
|
|
# ifdef __NR_tgkill
|
|
val = INTERNAL_SYSCALL (tgkill, err, 3, THREAD_GETMEM (THREAD_SELF, pid),
|
|
t->tid, SIGSETXID);
|
|
if (INTERNAL_SYSCALL_ERROR_P (val, err)
|
|
&& INTERNAL_SYSCALL_ERRNO (val, err) == ENOSYS)
|
|
# endif
|
|
val = INTERNAL_SYSCALL (tkill, err, 2, t->tid, SIGSETXID);
|
|
#endif
|
|
|
|
if (!INTERNAL_SYSCALL_ERROR_P (val, err))
|
|
atomic_increment (&cmdp->cntr);
|
|
}
|
|
|
|
|
|
int
|
|
attribute_hidden
|
|
__nptl_setxid (struct xid_command *cmdp)
|
|
{
|
|
int result;
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
__xidcmd = cmdp;
|
|
cmdp->cntr = 0;
|
|
|
|
struct pthread *self = THREAD_SELF;
|
|
|
|
/* Iterate over the list with system-allocated threads first. */
|
|
list_t *runp;
|
|
list_for_each (runp, &stack_used)
|
|
{
|
|
struct pthread *t = list_entry (runp, struct pthread, list);
|
|
if (t == self)
|
|
continue;
|
|
|
|
setxid_signal_thread (cmdp, t);
|
|
}
|
|
|
|
/* Now the list with threads using user-allocated stacks. */
|
|
list_for_each (runp, &__stack_user)
|
|
{
|
|
struct pthread *t = list_entry (runp, struct pthread, list);
|
|
if (t == self)
|
|
continue;
|
|
|
|
setxid_signal_thread (cmdp, t);
|
|
}
|
|
|
|
int cur = cmdp->cntr;
|
|
while (cur != 0)
|
|
{
|
|
lll_futex_wait (&cmdp->cntr, cur, LLL_PRIVATE);
|
|
cur = cmdp->cntr;
|
|
}
|
|
|
|
/* This must be last, otherwise the current thread might not have
|
|
permissions to send SIGSETXID syscall to the other threads. */
|
|
INTERNAL_SYSCALL_DECL (err);
|
|
result = INTERNAL_SYSCALL_NCS (cmdp->syscall_no, err, 3,
|
|
cmdp->id[0], cmdp->id[1], cmdp->id[2]);
|
|
if (INTERNAL_SYSCALL_ERROR_P (result, err))
|
|
{
|
|
__set_errno (INTERNAL_SYSCALL_ERRNO (result, err));
|
|
result = -1;
|
|
}
|
|
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
return result;
|
|
}
|
|
|
|
static inline void __attribute__((always_inline))
|
|
init_one_static_tls (struct pthread *curp, struct link_map *map)
|
|
{
|
|
dtv_t *dtv = GET_DTV (TLS_TPADJ (curp));
|
|
# if TLS_TCB_AT_TP
|
|
void *dest = (char *) curp - map->l_tls_offset;
|
|
# elif TLS_DTV_AT_TP
|
|
void *dest = (char *) curp + map->l_tls_offset + TLS_PRE_TCB_SIZE;
|
|
# else
|
|
# error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined"
|
|
# endif
|
|
|
|
/* Fill in the DTV slot so that a later LD/GD access will find it. */
|
|
dtv[map->l_tls_modid].pointer.val = dest;
|
|
dtv[map->l_tls_modid].pointer.is_static = true;
|
|
|
|
/* Initialize the memory. */
|
|
memset (__mempcpy (dest, map->l_tls_initimage, map->l_tls_initimage_size),
|
|
'\0', map->l_tls_blocksize - map->l_tls_initimage_size);
|
|
}
|
|
|
|
void
|
|
attribute_hidden
|
|
__pthread_init_static_tls (struct link_map *map)
|
|
{
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
/* Iterate over the list with system-allocated threads first. */
|
|
list_t *runp;
|
|
list_for_each (runp, &stack_used)
|
|
init_one_static_tls (list_entry (runp, struct pthread, list), map);
|
|
|
|
/* Now the list with threads using user-allocated stacks. */
|
|
list_for_each (runp, &__stack_user)
|
|
init_one_static_tls (list_entry (runp, struct pthread, list), map);
|
|
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
}
|
|
|
|
|
|
void
|
|
attribute_hidden
|
|
__wait_lookup_done (void)
|
|
{
|
|
lll_lock (stack_cache_lock, LLL_PRIVATE);
|
|
|
|
struct pthread *self = THREAD_SELF;
|
|
|
|
/* Iterate over the list with system-allocated threads first. */
|
|
list_t *runp;
|
|
list_for_each (runp, &stack_used)
|
|
{
|
|
struct pthread *t = list_entry (runp, struct pthread, list);
|
|
if (t == self || t->header.gscope_flag == THREAD_GSCOPE_FLAG_UNUSED)
|
|
continue;
|
|
|
|
int *const gscope_flagp = &t->header.gscope_flag;
|
|
|
|
/* We have to wait until this thread is done with the global
|
|
scope. First tell the thread that we are waiting and
|
|
possibly have to be woken. */
|
|
if (atomic_compare_and_exchange_bool_acq (gscope_flagp,
|
|
THREAD_GSCOPE_FLAG_WAIT,
|
|
THREAD_GSCOPE_FLAG_USED))
|
|
continue;
|
|
|
|
do
|
|
lll_futex_wait (gscope_flagp, THREAD_GSCOPE_FLAG_WAIT, LLL_PRIVATE);
|
|
while (*gscope_flagp == THREAD_GSCOPE_FLAG_WAIT);
|
|
}
|
|
|
|
/* Now the list with threads using user-allocated stacks. */
|
|
list_for_each (runp, &__stack_user)
|
|
{
|
|
struct pthread *t = list_entry (runp, struct pthread, list);
|
|
if (t == self || t->header.gscope_flag == THREAD_GSCOPE_FLAG_UNUSED)
|
|
continue;
|
|
|
|
int *const gscope_flagp = &t->header.gscope_flag;
|
|
|
|
/* We have to wait until this thread is done with the global
|
|
scope. First tell the thread that we are waiting and
|
|
possibly have to be woken. */
|
|
if (atomic_compare_and_exchange_bool_acq (gscope_flagp,
|
|
THREAD_GSCOPE_FLAG_WAIT,
|
|
THREAD_GSCOPE_FLAG_USED))
|
|
continue;
|
|
|
|
do
|
|
lll_futex_wait (gscope_flagp, THREAD_GSCOPE_FLAG_WAIT, LLL_PRIVATE);
|
|
while (*gscope_flagp == THREAD_GSCOPE_FLAG_WAIT);
|
|
}
|
|
|
|
lll_unlock (stack_cache_lock, LLL_PRIVATE);
|
|
}
|