about summary refs log tree commit diff
path: root/src/thread
Commit message (Collapse)AuthorAgeFilesLines
* pthread_create need not set errnoRich Felker2011-04-031-1/+1
|
* block all signals during rsyscallRich Felker2011-04-031-4/+9
| | | | | otherwise a signal handler could see an inconsistent and nonconformant program state where different threads have different uids/gids.
* fix race condition in rsyscall handlerRich Felker2011-04-031-1/+1
| | | | | | | | | | | | | | | | | | the problem: there is a (single-instruction) race condition window between a thread flagging itself dead and decrementing itself from the thread count. if it receives the rsyscall signal at this exact moment, the rsyscall caller will never succeed in signalling enough flags to succeed, and will deadlock forever. in previous versions of musl, the about-to-terminate thread masked all signals prior to decrementing the thread count, but this cost a whole syscall just to account for extremely rare races. the solution is a huge hack: rather than blocking in the signal handler if the thread is dead, modify the signal mask of the saved context and return in order to prevent further signal handling by the dead thread. this allows the dead thread to continue decrementing the thread count (if it had not yet done so) and exiting, even while the live part of the program blocks for rsyscall.
* don't trust siginfo in rsyscall handlerRich Felker2011-04-031-3/+2
| | | | | | | | for some inexplicable reason, linux allows the sender of realtime signals to spoof its identity. permission checks for sending signals should limit the impact to same-user processes, but just to be safe, we avoid trusting the siginfo structure and instead simply examine the program state to see if we're in the middle of a legitimate rsyscall.
* simplify calling of timer signal handlerRich Felker2011-04-031-7/+4
|
* simplify pthread tsd key handlingRich Felker2011-04-032-8/+6
|
* omit pthread tsd dtor code if tsd is not usedRich Felker2011-04-032-14/+24
|
* simplify setting result on thread cancellationRich Felker2011-04-011-1/+1
|
* use bss instead of mmap for main thread's pthread thread-specific dataRich Felker2011-04-012-9/+4
| | | | this simplifies code and removes a failure case
* fix misspelled PTHREAD_CANCELED constantRich Felker2011-04-011-1/+1
|
* use a_store to set cancel flag in pthread_cancel, to ensure a barrierRich Felker2011-04-011-1/+1
|
* simplify pthread_key_deleteRich Felker2011-03-311-1/+1
| | | | | calling this function on an uninitialized key value is UB, so there is no need to check that the table pointer was initialized.
* greatly simplify pthread_key_create (~20% size reduction)Rich Felker2011-03-311-10/+9
|
* avoid crash on stupid but allowable usage of pthread_mutex_unlockRich Felker2011-03-301-1/+3
| | | | | | | unlocking an unlocked mutex is not UB for robust or error-checking mutexes, so we must avoid calling __pthread_self (which might crash due to lack of thread-register initialization) until after checking that the mutex is locked.
* streamline mutex unlock to remove a useless branch, use a_store to unlockRich Felker2011-03-301-2/+6
| | | | | | | | | this roughly halves the cost of pthread_mutex_unlock, at least for non-robust, normal-type mutexes. the a_store change is in preparation for future support of archs which require a memory barrier or special atomic store operation, and also should prevent the possibility of the compiler misordering writes.
* cheap special-case optimization for normal mutexesRich Felker2011-03-301-0/+4
| | | | | | | cycle-level benchmark on atom cpu showed typical pthread_mutex_lock call dropping from ~120 cycles to ~90 cycles with this change. benefit may vary with compiler options and version, but this optimization is very cheap to make and should always help some.
* revert mutex "optimization" that turned out to be worseRich Felker2011-03-291-1/+1
|
* major improvements to cancellation handlingRich Felker2011-03-292-6/+13
| | | | | | | | | | | | | - there is no longer any risk of spoofing cancellation requests, since the cancel flag is set in pthread_cancel rather than in the signal handler. - cancellation signal is no longer unblocked when running the cancellation handlers. instead, pthread_create will cause any new threads created from a cancellation handler to unblock their own cancellation signal. - various tweaks in preparation for POSIX timer support.
* revert some more spin optimizations that turned out to be pessimizationsRich Felker2011-03-282-3/+2
|
* fix broken spinlock due to miscompilationRich Felker2011-03-281-1/+1
| | | | actually this trick also seems to have made the uncontended case slower.
* match glibc/lsb cancellation abi on i386Rich Felker2011-03-256-0/+36
| | | | | | | | glibc made the ridiculous choice to use pass-by-register calling convention for these functions, which is impossible to duplicate directly on non-gcc compilers. instead, we use ugly asm to wrap and convert the calling convention. presumably this works with every compiler anyone could potentially want to use.
* prepare pthread_spin_unlock for archs that need memory barriersRich Felker2011-03-241-1/+2
|
* optimize contended case for pthread_spin_trylockRich Felker2011-03-241-1/+2
|
* optimize spinlock spinRich Felker2011-03-241-1/+1
|
* overhaul cancellation to fix resource leaks and dangerous behavior with signalsRich Felker2011-03-243-10/+21
| | | | | | | | | | | | | | | | | | | | | | | | | | | | this commit addresses two issues: 1. a race condition, whereby a cancellation request occurring after a syscall returned from kernelspace but before the subsequent CANCELPT_END would cause cancellable resource-allocating syscalls (like open) to leak resources. 2. signal handlers invoked while the thread was blocked at a cancellation point behaved as if asynchronous cancellation mode wer in effect, resulting in potentially dangerous state corruption if a cancellation request occurs. the glibc/nptl implementation of threads shares both of these issues. with this commit, both are fixed. however, cancellation points encountered in a signal handler will not be acted upon if the signal was received while the thread was already at a cancellation point. they will of course be acted upon after the signal handler returns, so in real-world usage where signal handlers quickly return, it should not be a problem. it's possible to solve this problem too by having sigaction() wrap all signal handlers with a function that uses a pthread_cleanup handler to catch cancellation, patch up the saved context, and return into the cancellable function that will catch and act upon the cancellation. however that would be a lot of complexity for minimal if any benefit...
* global cleanup to use the new syscall interfaceRich Felker2011-03-204-6/+5
|
* if returning errno value directly from a syscall, we need to negate it.Rich Felker2011-03-192-2/+2
|
* syscall overhaul part two - unify public and internal syscall interfaceRich Felker2011-03-197-15/+9
| | | | | | | | | | | | | | | | with this patch, the syscallN() functions are no longer needed; a variadic syscall() macro allows syscalls with anywhere from 0 to 6 arguments to be made with a single macro name. also, manually casting each non-integer argument with (long) is no longer necessary; the casts are hidden in the macros. some source files which depended on being able to define the old macro SYSCALL_RETURNS_ERRNO have been modified to directly use __syscall() instead of syscall(). references to SYSCALL_SIGSET_SIZE and SYSCALL_LL have also been changed. x86_64 has not been tested, and may need a follow-up commit to fix any minor bugs/oversights.
* overhaul syscall interfaceRich Felker2011-03-191-2/+2
| | | | | | | | | | | | | | | this commit shuffles around the location of syscall definitions so that we can make a syscall() library function with both SYS_* and __NR_* style syscall names available to user applications, provides the syscall() library function, and optimizes the code that performs the actual inline syscalls in the library itself. previously on i386 when built as PIC (shared library), syscalls were incurring bus lock (lock prefix) overhead at entry and exit, due to the way the ebx register was being loaded (xchg instruction with a memory operand). now the xchg takes place between two registers. further cleanup to arch/$(ARCH)/syscall.h is planned.
* implement robust mutexesRich Felker2011-03-178-10/+78
| | | | | | some of this code should be cleaned up, e.g. using macros for some of the bit flags, masks, etc. nonetheless, the code is believed to be working and correct at this point.
* avoid function call to pthread_self in mutex unlockRich Felker2011-03-171-1/+1
| | | | | | if the mutex was previously locked, we can assume pthread_self was already called at the time of locking, and thus that the thread pointer is initialized.
* unify lock and owner fields of mutex structureRich Felker2011-03-173-7/+6
| | | | | | this change is necessary to free up one slot in the mutex structure so that we can use doubly-linked lists in the implementation of robust mutexes.
* optimize contended normal mutex case; add int compare-and-swap atomicRich Felker2011-03-171-1/+1
|
* simplify logic, slightly optimize contended case for non-default mutex typesRich Felker2011-03-161-4/+2
|
* correct error returns for error-checking mutexesRich Felker2011-03-162-2/+6
|
* cut out a syscall on thread creation in the case where guard size is 0Rich Felker2011-03-161-1/+1
|
* don't expose EAGAIN, etc. from timed futex wait to callerRich Felker2011-03-161-1/+4
|
* implement flockfile api, rework stdio lockingRich Felker2011-03-121-0/+1
|
* implement dummy pthread_attr_[gs]etschedparam functionsRich Felker2011-03-112-0/+14
| | | | | | | | for some reason these functions are not shaded by the PS/TPS option in POSIX, so presumably they are mandatory, even though the functionality they offer is optional. for now, provide them in case any programs depend on their existence, but disallow any priority except the default.
* fix pthread_attr_* implementations to match corrected prototypesRich Felker2011-03-114-4/+4
|
* fix failure behavior of sem_open when sem does not existRich Felker2011-03-101-1/+5
|
* fix some semaphore wait semantics (race condition deadlock and error checking)Rich Felker2011-03-102-1/+7
|
* fix sem_open and sem_close to obey posix semanticsRich Felker2011-03-102-33/+80
| | | | | | | | | multiple opens of the same named semaphore must return the same pointer, and only the last close can unmap it. thus the ugly global state keeping track of mappings. the maximum number of distinct named semaphores that can be opened is limited sufficiently small that the linear searches take trivial time, especially compared to the syscall overhead of these functions.
* optimize pthread termination in the non-detached caseRich Felker2011-03-101-4/+15
| | | | | | | we can avoid blocking signals by simply using a flag to mark that the thread has exited and prevent it from getting counted in the rsyscall signal-pingpong. this restores the original pthread create/join throughput from before the sigprocmask call was added.
* security fix: check that cancel/rsyscall signal was sent by the process itselfRich Felker2011-03-101-0/+3
|
* fix error handling for pthread_sigmaskRich Felker2011-03-091-0/+10
| | | | it must return errno, not -1, and should reject invalud values for how.
* optimize pthread initializationRich Felker2011-03-091-2/+2
| | | | | | the set_tid_address returns the tid (which is also the pid when called from the initial thread) so there is no need to make a separate syscall to get pid/tid.
* simplify and optimize pthread_mutex_trylockRich Felker2011-03-081-17/+16
|
* rwlock trylock functions were wrongly returning EAGAIN instead of EBUSYRich Felker2011-03-082-3/+3
|
* fix major breakage in pthread_once (it was always deadlocking)Rich Felker2011-03-081-8/+7
| | | | | | the issue was a break statement that was breaking only from the switch, not the enclosing for loop, and a failure to set the final success state.