diff options
author | Adhemerval Zanella <adhemerval.zanella@linaro.org> | 2021-08-16 15:08:27 -0300 |
---|---|---|
committer | Adhemerval Zanella <adhemerval.zanella@linaro.org> | 2021-12-15 17:35:38 -0300 |
commit | 98d5fcb8d099a1a868e032c89891c395a2f365c5 (patch) | |
tree | 13b1a0a98165f832a5ae394b958cb749709dad1d /malloc/arena.c | |
parent | 6cc3ccc67e0dda654fc839377af2818a296f0007 (diff) | |
download | glibc-98d5fcb8d099a1a868e032c89891c395a2f365c5.tar.gz glibc-98d5fcb8d099a1a868e032c89891c395a2f365c5.tar.xz glibc-98d5fcb8d099a1a868e032c89891c395a2f365c5.zip |
malloc: Add Huge Page support for mmap
With the morecore hook removed, there is not easy way to provide huge pages support on with glibc allocator without resorting to transparent huge pages. And some users and programs do prefer to use the huge pages directly instead of THP for multiple reasons: no splitting, re-merging by the VM, no TLB shootdowns for running processes, fast allocation from the reserve pool, no competition with the rest of the processes unlike THP, no swapping all, etc. This patch extends the 'glibc.malloc.hugetlb' tunable: the value '2' means to use huge pages directly with the system default size, while a positive value means and specific page size that is matched against the supported ones by the system. Currently only memory allocated on sysmalloc() is handled, the arenas still uses the default system page size. To test is a new rule is added tests-malloc-hugetlb2, which run the addes tests with the required GLIBC_TUNABLE setting. On systems without a reserved huge pages pool, is just stress the mmap(MAP_HUGETLB) allocation failure. To improve test coverage it is required to create a pool with some allocated pages. Checked on x86_64-linux-gnu. Reviewed-by: DJ Delorie <dj@redhat.com>
Diffstat (limited to 'malloc/arena.c')
-rw-r--r-- | malloc/arena.c | 4 |
1 files changed, 2 insertions, 2 deletions
diff --git a/malloc/arena.c b/malloc/arena.c index cd00c7bef4..9a6e1af2bd 100644 --- a/malloc/arena.c +++ b/malloc/arena.c @@ -230,7 +230,7 @@ TUNABLE_CALLBACK_FNDECL (set_tcache_count, size_t) TUNABLE_CALLBACK_FNDECL (set_tcache_unsorted_limit, size_t) #endif TUNABLE_CALLBACK_FNDECL (set_mxfast, size_t) -TUNABLE_CALLBACK_FNDECL (set_hugetlb, int32_t) +TUNABLE_CALLBACK_FNDECL (set_hugetlb, size_t) #else /* Initialization routine. */ #include <string.h> @@ -331,7 +331,7 @@ ptmalloc_init (void) TUNABLE_CALLBACK (set_tcache_unsorted_limit)); # endif TUNABLE_GET (mxfast, size_t, TUNABLE_CALLBACK (set_mxfast)); - TUNABLE_GET (hugetlb, int32_t, TUNABLE_CALLBACK (set_hugetlb)); + TUNABLE_GET (hugetlb, size_t, TUNABLE_CALLBACK (set_hugetlb)); #else if (__glibc_likely (_environ != NULL)) { |