Message ID | 20220719162339.23865-1-tariqt@nvidia.com (mailing list archive) |
---|---|
Headers | show |
Series | Introduce and use NUMA distance metrics | expand |
On 7/19/2022 7:23 PM, Tariq Toukan wrote: > Hi, > > Implement and expose CPU spread API based on the scheduler's > sched_numa_find_closest(). Use it in mlx5 and enic device drivers. This > replaces the binary NUMA preference (local / remote) with an improved one > that minds the actual distances, so that remote NUMAs with short distance > are preferred over farther ones. > > This has significant performance implications when using NUMA-aware > memory allocations, improving the throughput and CPU utilization. > > Regards, > Tariq > > v3: > - Introduce the logic as a common API instead of being mlx5 specific. > - Add implementation to enic device driver. > - Use non-atomic version of __cpumask_clear_cpu. > Comments on V2 were addressed. Please let me now of any other comments on this V3.
On Sun, 24 Jul 2022 15:09:11 +0300 Tariq Toukan wrote: > On 7/19/2022 7:23 PM, Tariq Toukan wrote: > > Hi, > > > > Implement and expose CPU spread API based on the scheduler's > > sched_numa_find_closest(). Use it in mlx5 and enic device drivers. This > > replaces the binary NUMA preference (local / remote) with an improved one > > that minds the actual distances, so that remote NUMAs with short distance > > are preferred over farther ones. > > > > This has significant performance implications when using NUMA-aware > > memory allocations, improving the throughput and CPU utilization. > > Comments on V2 were addressed. > Please let me now of any other comments on this V3. Good enough from my POV, just waiting for PeterZ's re-ack since patch 1 has changed.