Message ID | 20231011164356.2218554-1-adhemerval.zanella@linaro.org |
---|---|
State | Accepted |
Commit | 472894d2cfee5751b44c0aaa71ed87df81c8e62e |
Headers | show |
Series | malloc: Use __get_nprocs on arena_get2 (BZ 30945) | expand |
Part of the reason for using get_nprocs_sched() is that cpu affinity can limit the application to fewer processors, and the logic needs to take that into account. If you have a 192-core system but use setaffinity to limit a process to 4 cores, there will be *way* too many arenas created.
On 11/10/23 15:54, DJ Delorie wrote: > > Part of the reason for using get_nprocs_sched() is that cpu affinity can > limit the application to fewer processors, and the logic needs to take > that into account. > > If you have a 192-core system but use setaffinity to limit a process to > 4 cores, there will be *way* too many arenas created. > Yes, that was your rationale on the previous attempt [1]. Maybe adding a tunable, so if you are using affinity you can change thie behavior? In any case, I think we should either close the bug or fix this and work on a better heuristic to handle process that set affinity. [1] https://sourceware.org/pipermail/libc-alpha/2022-June/140161.html
Adhemerval Zanella Netto <adhemerval.zanella@linaro.org> writes: > Yes, that was your rationale on the previous attempt [1]. Maybe adding a > tunable, so if you are using affinity you can change thie behavior? We have a tunable to set the max arenas already. Measuring the number of runnable CPUs seems more of a "correctness" thing to me. This code is only called when we need to create a new arena, so it's not in a performance-critical path. It's more important to get this right so that the rest of the code maximizes performance. > In any case, I think we should either close the bug or fix this and work > on a better heuristic to handle process that set affinity. It sound like we need something between get_nprocs and get_nprocs_sched that counts the number of cores the program is assigned, not the number of cores the thread is assigned? I thought that's what sched_getaffinity() did, but the man page says "process" most of the time, but sometimes "thread".
On 11/10/23 17:34, DJ Delorie wrote: > Adhemerval Zanella Netto <adhemerval.zanella@linaro.org> writes: >> Yes, that was your rationale on the previous attempt [1]. Maybe adding a >> tunable, so if you are using affinity you can change thie behavior? > > We have a tunable to set the max arenas already. > > Measuring the number of runnable CPUs seems more of a "correctness" > thing to me. This code is only called when we need to create a new > arena, so it's not in a performance-critical path. It's more important > to get this right so that the rest of the code maximizes performance. I agree, my main problem with this change it was originally unintentional. My suggestion would to revert it back, and proper change with a meaningful commit message that it is intentional and process that use setaffinity and see contention increase the max arenas. > >> In any case, I think we should either close the bug or fix this and work >> on a better heuristic to handle process that set affinity. > > It sound like we need something between get_nprocs and get_nprocs_sched > that counts the number of cores the program is assigned, not the number > of cores the thread is assigned? I thought that's what > sched_getaffinity() did, but the man page says "process" most of the > time, but sometimes "thread". > The sched_getaffinity on Linux is essentially pthread_getaffinity_np, since the cpu mask is per process attribute. And I couldn't find any procfs way to access the AND mask of all mask withins the thread group. And I see no easy way to accomplish it, iterating over the /proc/self/task is way too slow and do not scale. Best option I can think of is add a global mask and keep track of set bits on each sched_setaffinity successful call; but it would require at least some global locking to avoid racy conditions.
Adhemerval Zanella Netto <adhemerval.zanella@linaro.org> writes: > On 11/10/23 15:54, DJ Delorie wrote: >> >> Part of the reason for using get_nprocs_sched() is that cpu affinity can >> limit the application to fewer processors, and the logic needs to take >> that into account. >> >> If you have a 192-core system but use setaffinity to limit a process to >> 4 cores, there will be *way* too many arenas created. >> > > Yes, that was your rationale on the previous attempt [1]. Maybe adding a > tunable, so if you are using affinity you can change thie behavior? > > In any case, I think we should either close the bug or fix this and work > on a better heuristic to handle process that set affinity. > > [1] https://sourceware.org/pipermail/libc-alpha/2022-June/140161.html Rather than let this issue rot, I'm withdrawing my dissent. I still think we need to resolve this issue with something between "nprocs" and "thread procs" - perhaps store the sched mask at startup and use that? But that could be deferred to some new development. I think the max-arenas tunable is more useful if you err on the side of too many arenas, too.
diff --git a/include/sys/sysinfo.h b/include/sys/sysinfo.h index c490561581..65742b1036 100644 --- a/include/sys/sysinfo.h +++ b/include/sys/sysinfo.h @@ -14,10 +14,6 @@ libc_hidden_proto (__get_nprocs_conf) extern int __get_nprocs (void); libc_hidden_proto (__get_nprocs) -/* Return the number of available processors which the process can - be scheduled. */ -extern int __get_nprocs_sched (void) attribute_hidden; - /* Return number of physical pages of memory in the system. */ extern long int __get_phys_pages (void); libc_hidden_proto (__get_phys_pages) diff --git a/malloc/arena.c b/malloc/arena.c index 6f03955ff2..82b09adb47 100644 --- a/malloc/arena.c +++ b/malloc/arena.c @@ -820,7 +820,7 @@ arena_get2 (size_t size, mstate avoid_arena) narenas_limit = mp_.arena_max; else if (narenas > mp_.arena_test) { - int n = __get_nprocs_sched (); + int n = __get_nprocs (); if (n >= 1) narenas_limit = NARENAS_FROM_NCORES (n); diff --git a/misc/getsysstats.c b/misc/getsysstats.c index 5f36adc0e8..23cc112074 100644 --- a/misc/getsysstats.c +++ b/misc/getsysstats.c @@ -44,12 +44,6 @@ weak_alias (__get_nprocs, get_nprocs) link_warning (get_nprocs, "warning: get_nprocs will always return 1") -int -__get_nprocs_sched (void) -{ - return 1; -} - long int __get_phys_pages (void) { diff --git a/sysdeps/mach/getsysstats.c b/sysdeps/mach/getsysstats.c index 5184e5eee1..d3834f3b69 100644 --- a/sysdeps/mach/getsysstats.c +++ b/sysdeps/mach/getsysstats.c @@ -62,12 +62,6 @@ __get_nprocs (void) libc_hidden_def (__get_nprocs) weak_alias (__get_nprocs, get_nprocs) -int -__get_nprocs_sched (void) -{ - return __get_nprocs (); -} - /* Return the number of physical pages on the system. */ long int __get_phys_pages (void) diff --git a/sysdeps/unix/sysv/linux/getsysstats.c b/sysdeps/unix/sysv/linux/getsysstats.c index b0b6c154ac..1ea7f1f01f 100644 --- a/sysdeps/unix/sysv/linux/getsysstats.c +++ b/sysdeps/unix/sysv/linux/getsysstats.c @@ -29,7 +29,7 @@ #include <sys/sysinfo.h> #include <sysdep.h> -int +static int __get_nprocs_sched (void) { enum