about summary refs log tree commit diff
path: root/nptl/sysdeps
Commit message (Collapse)AuthorAgeFilesLines
* i386: Remove useless pthread_spin_{init,unlock} wrapper files.Roland McGrath2014-05-142-2/+0
|
* Move remaining nptl/sysdeps/unix/sysv/linux/x86_64/ files.Roland McGrath2014-05-1425-5323/+0
|
* x86: Consolidate NPTL fork.Roland McGrath2014-05-143-61/+1
|
* Consolidate not-cancel.h files.Roland McGrath2014-05-146-109/+0
|
* Update s390 timer_*.c files for x86_64 file moves.Roland McGrath2014-05-145-5/+5
|
* x86_64: Remove useless pthread_spin_{init,unlock} wrapper files.Roland McGrath2014-05-142-2/+0
|
* Move x86_64 compat-timer.h out of nptl/Roland McGrath2014-05-141-45/+0
|
* Move x86_64 timer_*.c out of nptl/Roland McGrath2014-05-1416-240/+10
|
* x86: Consolidate NPTL/non versions of cloneRoland McGrath2014-05-142-18/+0
|
* Move NPTL public ABI headers for x86 to sysdeps/x86/nptl/.Roland McGrath2014-05-142-290/+0
|
* x86: Consolidate NPTL/non versions of vforkRoland McGrath2014-05-144-184/+0
|
* Clean up __exit_thread.Roland McGrath2014-05-137-60/+0
|
* S/390: Port of lock elision to System/zDominik Vogt2014-05-0914-2/+584
| | | | | | | | | | Added support for TX lock elision of pthread mutexes on s390 and s390x. This may improve lock scaling of existing programs on TX capable systems. The lock elision code is only built with --enable-lock-elision=yes and then requires a GCC version supporting the TX builtins. With lock elision default mutexes are elided via __builtin_tbegin, if the cpu supports transactions. By default lock elision is not enabled and the elision code is not built.
* Deconditionalize use of LLL_LOCK_INITIALIZER in bits/libc-lock.h.Roland McGrath2014-05-011-6/+1
|
* Include atomic.h in sem_wait.c and sem_trywait.cSiddhesh Poyarekar2014-04-222-0/+2
|
* PowerPC: Sync pthread_once with default implementationAdhemerval Zanella2014-04-171-110/+0
| | | | | | | | This patch removes the arch specific powerpc implementation and instead uses the linux default one. Although the current powerpc implementation already constains the required memory barriers for correct initialization, the default implementation shows a better performance on newer chips.
* Fixed and unified pthread_once.Torvald Riegel2014-04-112-93/+131
| | | | | | | [BZ #15215] This unifies various pthread_once architecture-specific implementations which were using the same algorithm with slightly different implementations. It also adds missing memory barriers that are required for correctness.
* Move __PTHREAD_SPINS definition to architecture specific headerAdhemerval Zanella2014-04-096-13/+15
| | | | | | | This patch moves the __PTHREAD_SPINS definition to arch specific header since pthread_mutex_t layout is also arch specific. This leads to no need to defining __PTHREAD_MUTEX_HAVE_ELISION and thus removing of the undefined compiler warning.
* Fix dwarf2 unwinding through futex functions.Andi Kleen2014-03-262-336/+40
| | | | | | | | | | | | | | | | | | | When profiling programs with lock problems with perf record -g dwarf, libunwind can currently not backtrace through the futex and unlock functions in pthread. This is because they use out of line sections, and those are not correctly described in dwarf2 (I believe needs dwarf3 or 4). This patch first removes the out of line sections. They only save a single jump, but cause a lot of pain. Then it converts the now inline lock code to use the now standard gas .cfi_* commands. With these changes libunwind/perf can backtrace through the futex functions now. Longer term it would be likely better to just use C futex() functions on x86 like all the other architectures. This would clean the code up even more.
* PowerPC: Revert __PTHREAD_MUTEX_HAVE_ELISION changeAdhemerval Zanella2014-03-251-2/+0
| | | | This patch revert 449282f2e0e850c29f6a9666058503d4734964f0.
* PowerPC: Fix -Wundef warning for __PTHREAD_MUTEX_HAVE_ELISION.Adhemerval Zanella2014-03-251-0/+2
|
* S390: Fix -Wundef warning for __PTHREAD_MUTEX_HAVE_ELISION.Stefan Liebler2014-03-241-0/+2
|
* Revert "Fix __PTHREAD_MUTEX_HAVE_ELISION -Wundef warning"Will Newton2014-03-171-9/+5
| | | | This reverts commit 788bba368c2eaf8aa3fd2ca18d269395d6bc8afb.
* Fix __PTHREAD_MUTEX_HAVE_ELISION -Wundef warningWill Newton2014-03-171-5/+9
| | | | | | | | | | ChangeLog: 2014-03-17 Will Newton <will.newton@linaro.org> * nptl/sysdeps/pthread/pthread.h: Check __PTHREAD_MUTEX_HAVE_ELISION is defined before testing its value.
* Get rid of __LT_SPINLOCK_INITSiddhesh Poyarekar2014-03-171-6/+1
| | | | | | | | | We got rid of LinuxThreads in 2005, but we didn't remove __LT_SPINLOCK_INIT back then. Do it now. * nptl/sysdeps/pthread/bits/libc-lockP.h [defined NOT_IN_libc && !defined IS_IN_libpthread && __LT_SPINNOCK_INIT != 0]: Remove.
* Compile with -Wundef.Roland McGrath2014-03-146-13/+19
|
* Fix two stray cases using #ifdef vs #if for TLS_TCB_AT_TP.Roland McGrath2014-03-141-1/+1
|
* Use glibc_likely instead __builtin_expect.Ondřej Bílka2014-02-1019-44/+44
|
* Fix tst-setgetname for Linux kernels < 2.6.33.Carlos O'Donell2014-02-051-0/+14
| | | | | | | | Support for /proc/self/task/$tid/comm as added in Linux 2.6.33, therefore since the test tst-setgetname relies on this functionality to operate we must skip the test in kernels < 2.6.33. We wrap the checks with __ASSUME_PROC_PID_TASK_COMM such that in the future when we move arch_minimum_kernel to 2.6.33 we can remove this code.
* S/390: Make jmp_buf extendible.Andreas Krebbel2014-01-072-0/+68
|
* Update copyright notices with scripts/update-copyrightsAllan McRae2014-01-01307-307/+307
|
* PowerPC: Add DSO and TAR fields to TLSAdhemerval Zanella2013-12-122-0/+8
|
* Add test case for pthread_[sg]etname_np.Carlos O'Donell2013-12-122-0/+302
| | | | | | | | | This commit adds a testcase for pthread_setname_np and pthread_getname_np. The testcase itself has four tests to validate that these functions work as expected. The test is only enabled for Linux since it requires access to an alternate method for validating the functions work.
* Fix CFI annotations in pthread_cond_timedwait for i486+Andreas Schwab2013-12-091-2/+1
|
* PowerPC64 ELFv2 ABI 4/6: Stack frame layout changesUlrich Weigand2013-12-041-26/+34
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This updates glibc for the changes in the ELFv2 relating to the stack frame layout. These are described in more detail here: http://gcc.gnu.org/ml/gcc-patches/2013-11/msg01149.html http://gcc.gnu.org/ml/gcc-patches/2013-11/msg01146.html Specifically, the "compiler and linker doublewords" were removed, which has the effect that the save slot for the TOC register is now at offset 24 rather than 40 to the stack pointer. In addition, a function may now no longer necessarily assume that its caller has set up a 64-byte register save area its use. To address the first change, the patch goes through all assembler files and replaces immediate offsets in instructions accessing the ABI-defined stack slots by symbolic offsets. Those already were defined in ucontext_i.sym and used in some of the context routines, but that doesn't really seem like the right place for those defines. The patch instead defines those symbolic offsets in sysdeps.h, in two variants for the old and new ABI, and uses them systematically in all assembler files, not just the context routines. The second change only affected a few assembler files that used the save area to temporarily store some registers. In those cases where this happens within a leaf function, this patch changes the code to store those registers to the "red zone" below the stack pointer. Otherwise, the functions already allocate a stack frame, and the patch changes them to add extra space in these frames as temporary space for the ELFv2 ABI.
* PowerPC64: Add __private_ss field to TCB headerUlrich Weigand2013-12-041-0/+2
| | | | | | | | | | | | | | | The TCB header on Intel contains a field __private_ss that is used to efficiently implement the -fsplit-stack GCC feature. In order to prepare for a possible future implementation of that feature on powerpc64, we'd like to reserve a similar field in the TCB header as well. (It would be good if this went in with or before the ELFv2 patches to ensure that this field will be available always in the ELFv2 environment.) The field needs to be added at the front of tcbhead_t structure to avoid changing the ABI; see the recent discussion when adding the EBB fields.
* Use __glibc_reserved instead __unused.Ondřej Bílka2013-11-262-3/+3
|
* rename configure.in to configure.acMike Frysinger2013-10-304-2/+2
| | | | | | | Autoconf has been deprecating configure.in for quite a long time. Rename all our configure.in and preconfigure.in files to .ac. Signed-off-by: Mike Frysinger <vapier@gentoo.org>
* PowerPC ugly symbol versioningAlan Modra2013-10-041-9/+4
| | | | | | | | | | | | | | | | | | | | | | | http://sourceware.org/ml/libc-alpha/2013-08/msg00090.html This patch fixes symbol versioning in setjmp/longjmp. The existing code uses raw versions, which results in wrong symbol versioning when you want to build glibc with a base version of 2.19 for LE. Note that the merging the 64-bit and 32-bit versions in novmx-lonjmp.c and pt-longjmp.c doesn't result in GLIBC_2.0 versions for 64-bit, due to the base in shlib_versions. * sysdeps/powerpc/longjmp.c: Use proper symbol versioning macros. * sysdeps/powerpc/novmx-longjmp.c: Likewise. * sysdeps/powerpc/powerpc32/bsd-_setjmp.S: Likewise. * sysdeps/powerpc/powerpc32/bsd-setjmp.S: Likewise. * sysdeps/powerpc/powerpc32/fpu/__longjmp.S: Likewise. * sysdeps/powerpc/powerpc32/fpu/setjmp.S: Likewise. * sysdeps/powerpc/powerpc32/mcount.c: Likewise. * sysdeps/powerpc/powerpc32/setjmp.S: Likewise. * sysdeps/powerpc/powerpc64/setjmp.S: Likewise. * nptl/sysdeps/unix/sysv/linux/powerpc/pt-longjmp.c: Likewise.
* Use the mutex member of the argumen in __libc_lock_*_recursiveSiddhesh Poyarekar2013-09-271-3/+3
|
* Remove --disable-versioning.Joseph Myers2013-09-041-1/+1
|
* Fix typos.Ondřej Bílka2013-08-302-3/+3
|
* Fix typos.Ondřej Bílka2013-08-211-1/+1
|
* Clean up whitespace in lock elision patches.Dominik Vogt2013-07-1912-28/+34
| | | | Signed-off-by: Carlos O'Donell <carlos@redhat.com>
* Remove remains of rwlock elision which is not implemented yet.Dominik Vogt2013-07-192-11/+0
| | | | | | | | | | | | | | | Signed-off-by: Carlos O'Donell <carlos@redhat.com> --- nptl/ 2013-07-19 Dominik Vogt <vogt@de.ibm.com> * sysdeps/unix/sysv/linux/x86/elision-conf.c: Remove __rwlock_rtm_enabled and __rwlock_rtm_read_retries. (elision_init): Don't set __rwlock_rtm_enabled. * sysdeps/unix/sysv/linux/x86/elision-conf.h: Remove __rwlock_rtm_enabled.
* Add x86 init-arch to nptlH.J. Lu2013-07-032-0/+2
|
* Add a configure option to enable lock elision and disable by defaultAndi Kleen2013-07-021-0/+3
| | | | Can be enabled with --enable-lock-elision=yes at configure time.
* Add elision to pthread_mutex_{try,timed,un}lockAndi Kleen2013-07-028-5/+134
| | | | | | | | | | | | | | | | | | | | | | | Add elision paths to the basic mutex locks. The normal path has a check for RTM and upgrades the lock to RTM when available. Trylocks cannot automatically upgrade, so they check for elision every time. We use a 4 byte value in the mutex to store the lock elision adaptation state. This is separate from the adaptive spin state and uses a separate field. Condition variables currently do not support elision. Recursive mutexes and condition variables may be supported at some point, but are not in the current implementation. Also "trylock" will not automatically enable elision unless some other lock call has been already called on the lock. This version does not use IFUNC, so it means every lock has one additional check for elision. Benchmarking showed the overhead to be negligible.
* Add new internal mutex type flags for elision.Andi Kleen2013-07-021-8/+20
| | | | | | | | | | Add Enable/disable flags used internally Extend the mutex initializers to have the fields needed for elision. The layout stays the same, and this is not visible to programs. These changes are not exposed outside pthread
* Add the low level infrastructure for pthreads lock elision with TSXAndi Kleen2013-07-0210-0/+480
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Lock elision using TSX is a technique to optimize lock scaling It allows to run locks in parallel using hardware support for a transactional execution mode in 4th generation Intel Core CPUs. See http://www.intel.com/software/tsx for more Information. This patch implements a simple adaptive lock elision algorithm based on RTM. It enables elision for the pthread mutexes and rwlocks. The algorithm keeps track whether a mutex successfully elides or not, and stops eliding for some time when it is not. When the CPU supports RTM the elision path is automatically tried, otherwise any elision is disabled. The adaptation algorithm and its tuning is currently preliminary. The code adds some checks to the lock fast paths. Micro-benchmarks show little to no difference without RTM. This patch implements the low level "lll_" code for lock elision. Followon patches hook this into the pthread implementation Changes with the RTM mutexes: ----------------------------- Lock elision in pthreads is generally compatible with existing programs. There are some obscure exceptions, which are expected to be uncommon. See the manual for more details. - A broken program that unlocks a free lock will crash. There are ways around this with some tradeoffs (more code in hot paths) I'm still undecided on what approach to take here; have to wait for testing reports. - pthread_mutex_destroy of a lock mutex will not return EBUSY but 0. - There's also a similar situation with trylock outside the mutex, "knowing" that the mutex must be held due to some other condition. In this case an assert failure cannot be recovered. This situation is usually an existing bug in the program. - Same applies to the rwlocks. Some of the return values changes (for example there is no EDEADLK for an elided lock, unless it aborts. However when elided it will also never deadlock of course) - Timing changes, so broken programs that make assumptions about specific timing may expose already existing latent problems. Note that these broken programs will break in other situations too (loaded system, new faster hardware, compiler optimizations etc.) - Programs with non recursive mutexes that take them recursively in a thread and which would always deadlock without elision may not always see a deadlock. The deadlock will only happen on an early or delayed abort (which typically happens at some point) This only happens for mutexes not explicitely set to PTHREAD_MUTEX_NORMAL or PTHREAD_MUTEX_ADAPTIVE_NP. PTHREAD_MUTEX_NORMAL mutexes do not elide. The elision default can be set at configure time. This patch implements the basic infrastructure for elision.