about summary refs log tree commit diff
Commit message (Collapse)AuthorAgeFilesLines
* support vfork on i386Rich Felker2011-10-141-0/+14
|
* make available a namespace-safe vfork, if supportedRich Felker2011-10-141-1/+4
| | | | this may be useful to posix_spawn..?
* add dummy __cxa_finalizeRich Felker2011-10-141-0/+4
| | | | | | musl's dynamic linker does not support unloading dsos, so there's nothing for this function to do. adding the symbol in case anything depends on its presence..
* support __cxa_atexit, and registering atexit functions from atexit handlersRich Felker2011-10-141-7/+26
| | | | | mildly tested; may have bugs. the locking should be updated not to use spinlocks but that's outside the scope of this one module.
* simplify atexit and fflush-on-exit handlingRich Felker2011-10-143-9/+12
|
* fix F_GETOWN return value handlingRich Felker2011-10-091-0/+1
| | | | | | | | the fcntl syscall can return a negative value when the command is F_GETOWN, and this is not an error code but an actual value. thus we must special-case it and avoid calling __syscall_ret to set errno. this fix is better than the glibc fix (using F_GETOWN_EX) which only works on newer kernels and is more complex.
* fix typo in arm clone() asmRich Felker2011-10-091-1/+1
|
* fix fcntl O_* flags for armRich Felker2011-10-051-4/+4
| | | | no idea why these 4 are permuted and the rest are standard/generic
* make [U]INTn_C() macros have the right type...Rich Felker2011-10-041-4/+4
| | | | ...and still be valid in #if directives.
* recovering ownerdead robust mutex must reset recursive lock countRich Felker2011-10-031-0/+1
|
* simplify robust mutex unlock code pathRich Felker2011-10-031-4/+4
| | | | | | | right now it's questionable whether this change is an improvement or not, but if we later want to support priority inheritance mutexes, it will be important to have the code paths unified like this to avoid major code duplication.
* fix crash if pthread_mutex_unlock is called without ever lockingRich Felker2011-10-031-1/+1
| | | | | this is valid for error-checking mutexes; otherwise it invokes UB and would be justified in crashing.
* use count=0 instead of 1 for recursive mutex with only one lock referenceRich Felker2011-10-032-4/+2
| | | | | | | this simplifies the code paths slightly, but perhaps what's nicer is that it makes recursive mutexes fully reentrant, i.e. locking and unlocking from a signal handler works even if the interrupted code was in the middle of locking or unlocking.
* sysconf for PTHREAD_KEYS_MAXRich Felker2011-10-031-1/+1
|
* synchronize cond var destruction with exiting waitsRich Felker2011-10-023-0/+12
|
* fix failure-to-wake in rwlock unlockRich Felker2011-10-011-1/+1
| | | | | | | | | a reader unlocking the lock need only wake one waiter (necessarily a writer, but a writer unlocking the lock must wake all waiters (necessarily readers). if it only wakes one, the remainder can remain blocked indefinitely, or at least until the first reader unlocks (in which case the whole lock becomes serialized and behaves as a mutex rather than a read lock).
* dlsym entry point for armRich Felker2011-10-011-0/+6
|
* dynamic linker entry point for armRich Felker2011-10-011-0/+14
| | | | mildly tested, seems to work
* typo (copy 2) in arm reloc.hRich Felker2011-10-011-1/+1
|
* typo in arm reloc.hRich Felker2011-10-011-1/+1
|
* first attempt at arm dynamic linkingRich Felker2011-10-011-0/+26
|
* fix various bugs in path and error handling in execvp/fexecveRich Felker2011-09-292-18/+29
|
* don't crash on null strings in printfRich Felker2011-09-281-1/+1
| | | | passing null pointer for %s is UB but lots of broken programs do it anyway
* fix excessive/insufficient wakes in __vm_unlockRich Felker2011-09-281-3/+3
| | | | | | there is no need to send a wake when the lock count does not hit zero, but when it does, all waiters must be woken (since all with the same sign are eligible to obtain the lock).
* make getmntent_r discard long lines when it returns error, not seek backRich Felker2011-09-281-2/+1
| | | | | seeking back can be performed by the caller, but if the caller doesn't expect it, it will result in an infinite loop of failures.
* improve pshared barriersRich Felker2011-09-282-12/+14
| | | | | | | | | | | | | | eliminate the sequence number field and instead use the counter as the futex because of the way the lock is held, sequence numbers are completely useless, and this frees up a field in the barrier structure to be used as a waiter count for the count futex, which lets us avoid some syscalls in the best case. as of now, self-synchronized destruction and unmapping should be fully safe. before any thread can return from the barrier, all threads in the barrier have obtained the vm lock, and each holds a shared lock on the barrier. the barrier memory is not inspected after the shared lock count reaches 0, nor after the vm lock is released.
* next step making barrier self-sync'd destruction safeRich Felker2011-09-282-6/+18
| | | | i think this works, but it can be simplified. (next step)
* barrier destroy must also wait for threads in other processes exiting barrierRich Felker2011-09-281-0/+2
| | | | | the vm lock only waits for threads in the same process exiting. actually this fix is not enough, but it's a start...
* stupid typo (caused by rather ugly spelling in POSIX..) in aioRich Felker2011-09-282-2/+2
|
* correctly handle the degenerate barrier in the pshared caseRich Felker2011-09-271-1/+1
|
* fix crash in pthread_cond_wait mutex-locked checkRich Felker2011-09-271-1/+1
| | | | | | it was assuming the result of the condition it was supposed to be checking for, i.e. that the thread ptr had already been initialized by pthread_mutex_lock. use the slower call to be safe.
* fix crash in pthread_testcancel if pthread_self has not been calledRich Felker2011-09-271-1/+1
|
* improve/debloat mutex unlock error checking in pthread_cond_waitRich Felker2011-09-271-3/+3
| | | | | | | we're not required to check this except for error-checking mutexes, but it doesn't hurt. the new test is actually simpler/lighter, and it also eliminates the need to later check that pthread_mutex_unlock succeeds.
* check mutex owner in pthread_cond_waitRich Felker2011-09-271-0/+3
| | | | | | | | | | when used with error-checking mutexes, pthread_cond_wait is required to fail with EPERM if the mutex is not locked by the caller. previously we relied on pthread_mutex_unlock to generate the error, but this is not valid, since in the case of such invalid usage the internal state of the cond variable has already been potentially corrupted (due to access outside the control of the mutex). thus, we have to check first.
* fix pshared barrier wrong return value.Rich Felker2011-09-271-1/+1
| | | | i set the return value but then never used it... oops!
* convert the barrier pshared option back to 0/1 values when getting itRich Felker2011-09-271-1/+1
|
* improve error handling in mntent stuff...Rich Felker2011-09-271-2/+9
| | | | not sure if this is correct/ideal. it needs further attention.
* fix missing va_end in prctl syscall wrapperRich Felker2011-09-271-0/+1
|
* fix clobbering of errno in get(pw|gr)([ug]id|nam) by fcloseRich Felker2011-09-272-0/+12
|
* process-shared barrier support, based on discussion with bdonlanRich Felker2011-09-277-16/+102
| | | | | | | | | | | | | this implementation is rather heavy-weight, but it's the first solution i've found that's actually correct. all waiters actually wait twice at the barrier so that they can synchronize exit, and they hold a "vm lock" that prevents changes to virtual memory mappings (and blocks pthread_barrier_destroy) until all waiters are finished inspecting the barrier. thus, it is safe for any thread to destroy and/or unmap the barrier's memory as soon as pthread_barrier_wait returns, without further synchronization.
* fix incorrect allocation failure check in pthread_createRich Felker2011-09-271-1/+1
| | | | | mmap returns MAP_FAILED not 0 because some idiot thought the ability to mmap the null pointer page would be a good idea...
* fix ctype macros to cast argument to (unsigned) firstRich Felker2011-09-262-7/+7
| | | | | | | issue reported by nsz, but it's actually not just pedantic. the functions can take input of any arithmetic type, including floating point, and the behavior needs to be as if the conversion implicit in the function call took place.
* another cond var fix: requeue count race conditionRich Felker2011-09-262-15/+13
| | | | | | | | | | | | | lock out new waiters during the broadcast. otherwise the wait count added to the mutex might be lower than the actual number of waiters moved, and wakeups may be lost. this issue could also be solved by temporarily setting the mutex waiter count higher than any possible real count, then relying on the kernel to tell us how many waiters were requeued, and updating the counts afterwards. however the logic is more complex, and i don't really trust the kernel. the solution here is also nice in that it replaces some atomic cas loops with simple non-atomic ops under lock.
* fix lost signals in cond varsRich Felker2011-09-263-14/+17
| | | | | | | | | | | due to moving waiters from the cond var to the mutex in bcast, these waiters upon wakeup would steal slots in the count from newer waiters that had not yet been signaled, preventing the signal function from taking any action. to solve the problem, we simply use two separate waiter counts, and so that the original "total" waiters count is undisturbed by broadcast and still available for signal.
* cleanup various minor issues reported by nszRich Felker2011-09-266-9/+9
| | | | | | | | | the changes to syscall_ret are mostly no-ops in the generated code, just cleanup of type issues and removal of some implementation-defined behavior. the one exception is the change in the comparison value, which is fixed so that 0xf...f000 (which in principle could be a valid return value for mmap, although probably never in reality) is not treated as an error return.
* redo cond vars again, use sequence numbersRich Felker2011-09-264-48/+52
| | | | | | | | | | | | | | | | | | | | | | | | testing revealed that the old implementation, while correct, was giving way too many spurious wakeups due to races changing the value of the condition futex. in a test program with 5 threads receiving broadcast signals, the number of returns from pthread_cond_wait was roughly 3 times what it should have been (2 spurious wakeups for every legitimate wakeup). moreover, the magnitude of this effect seems to grow with the number of threads. the old implementation may also have had some nasty race conditions with reuse of the cond var with a new mutex. the new implementation is based on incrementing a sequence number with each signal event. this sequence number has nothing to do with the number of threads intended to be woken; it's only used to provide a value for the futex wait to avoid deadlock. in theory there is a danger of race conditions due to the value wrapping around after 2^32 signals. it would be nice to eliminate that, if there's a way. testing showed no spurious wakeups (though they are of course possible) with the new implementation, as well as slightly improved performance.
* revert previous change in cond var waiter moveRich Felker2011-09-251-2/+6
| | | | | | using swap has a race condition: the waiters must be added to the mutex waiter count *before* they are taken off the cond var waiter count, or wake events can be lost.
* optimize cond waiter move using atomic swap instead of cas loopRich Felker2011-09-251-6/+2
|
* fix logic for when wakeup is not desired on cond bcastRich Felker2011-09-251-3/+4
| | | | somehow i forgot that normal-type mutexes don't store the owner tid.
* new futex-requeue-based pthread_cond_broadcast implementationRich Felker2011-09-254-7/+69
| | | | | | this avoids the "stampede effect" where pthread_cond_broadcast would result in all waiters waking up simultaneously, only to immediately contend for the mutex and go back to sleep.