Message ID | 8581ca937d4064b3cd138845d5bd418580d177da.1685506205.git.raghavendra.kt@amd.com (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | sched/numa: Fix disjoint set vma scan regression | expand |
On 5/31/2023 9:55 AM, Raghavendra K T wrote: > With the numa scan enhancements [1], only the threads which had previously > accessed vma are allowed to scan. > > While this had improved significant system time overhead, there were corner > cases, which genuinely need some relaxation. For e.g., > > 1) Concern raised by PeterZ, where if there are N partition sets of vmas > belonging to tasks, then unfairness in allowing these threads to scan could > potentially amplify the side effect of some of the vmas being left > unscanned. > > 2) Below reports of LKP numa01 benchmark regression. > > Currently this is handled by allowing first two scanning unconditional > as indicated by mm->numa_scan_seq. This is imprecise since for some > benchmark vma scanning might itself start at numa_scan_seq > 2. > > Solution: > Allow unconditional scanning of vmas of tasks depending on vma size. This > is achieved by maintaining a per vma scan counter, where > > f(allowed_to_scan) = f(scan_counter < vma_size / scan_size) > > Result: > numa01_THREAD_ALLOC result on 6.4.0-rc2 (that has numascan enhancement) > base-numascan base base+fix > real 1m1.507s 1m23.259s 1m2.632s > user 213m51.336s 251m46.363s 220m35.528s > sys 3m3.397s 0m12.492s 2m41.393s > > numa_hit 5615517 4560123 4963875 > numa_local 5615505 4560024 4963700 > numa_other 12 99 175 > numa_pte_updates 1822797 493 1559111 > numa_hint_faults 1307113 523 1469031 > numa_hint_faults_local 612617 488 884829 > numa_pages_migrated 694370 35 584202 > > Summary: Regression in base is recovered by allowing scanning as required. > > [1] https://lore.kernel.org/lkml/cover.1677672277.git.raghavendra.kt@amd.com/T/#t > > Fixes: fc137c0ddab2 ("sched/numa: enhance vma scanning logic") > regression. > Reported-by: Aithal Srikanth <sraithal@amd.com> > Reported-by: kernel test robot <oliver.sang@intel.com> > Closes: https://lore.kernel.org/lkml/db995c11-08ba-9abf-812f-01407f70a5d4@amd.com/T/ > Signed-off-by: Raghavendra K T <raghavendra.kt@amd.com> Hello kernel test robot, Gentle ping to check if the patch has helped your regression report. Thanks - Raghu
hi, Raghavendra K T, On Wed, Jul 05, 2023 at 11:18:37AM +0530, Raghavendra K T wrote: > On 5/31/2023 9:55 AM, Raghavendra K T wrote: > > With the numa scan enhancements [1], only the threads which had previously > > accessed vma are allowed to scan. > > > > While this had improved significant system time overhead, there were corner > > cases, which genuinely need some relaxation. For e.g., > > > > 1) Concern raised by PeterZ, where if there are N partition sets of vmas > > belonging to tasks, then unfairness in allowing these threads to scan could > > potentially amplify the side effect of some of the vmas being left > > unscanned. > > > > 2) Below reports of LKP numa01 benchmark regression. > > > > Currently this is handled by allowing first two scanning unconditional > > as indicated by mm->numa_scan_seq. This is imprecise since for some > > benchmark vma scanning might itself start at numa_scan_seq > 2. > > > > Solution: > > Allow unconditional scanning of vmas of tasks depending on vma size. This > > is achieved by maintaining a per vma scan counter, where > > > > f(allowed_to_scan) = f(scan_counter < vma_size / scan_size) > > > > Result: > > numa01_THREAD_ALLOC result on 6.4.0-rc2 (that has numascan enhancement) > > base-numascan base base+fix > > real 1m1.507s 1m23.259s 1m2.632s > > user 213m51.336s 251m46.363s 220m35.528s > > sys 3m3.397s 0m12.492s 2m41.393s > > > > numa_hit 5615517 4560123 4963875 > > numa_local 5615505 4560024 4963700 > > numa_other 12 99 175 > > numa_pte_updates 1822797 493 1559111 > > numa_hint_faults 1307113 523 1469031 > > numa_hint_faults_local 612617 488 884829 > > numa_pages_migrated 694370 35 584202 > > > > Summary: Regression in base is recovered by allowing scanning as required. > > > > [1] https://lore.kernel.org/lkml/cover.1677672277.git.raghavendra.kt@amd.com/T/#t > > > > Fixes: fc137c0ddab2 ("sched/numa: enhance vma scanning logic") > > regression. > > Reported-by: Aithal Srikanth <sraithal@amd.com> > > Reported-by: kernel test robot <oliver.sang@intel.com> > > Closes: https://lore.kernel.org/lkml/db995c11-08ba-9abf-812f-01407f70a5d4@amd.com/T/ > > Signed-off-by: Raghavendra K T <raghavendra.kt@amd.com> > > Hello kernel test robot, > > Gentle ping to check if the patch has helped your regression report. sorry for late. previously we found a 118.9% regression of autonuma-benchmark.numa01.seconds on a Cascade Lake, which happened to be converted for other test purposes, so we cannot test your patch on it again. however, we also found a 39.3% regression on a Sapphire Rapids test machine: ========================================================================================= compiler/cpufreq_governor/iterations/kconfig/rootfs/tbox_group/test/testcase: gcc-11/performance/4x/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-spr-r02/numa02_SMT/autonuma-benchmark ef6a22b70f6d9044 fc137c0ddab29b591db6a091dc6 ---------------- --------------------------- %stddev %change %stddev \ | \ 193.14 +39.2% 268.84 autonuma-benchmark.numa01.seconds 8.14 -0.7% 8.09 autonuma-benchmark.numa02.seconds now we tested your v3 patch on it, found regression mostly recovered (55fd15913b18d6a790c17d947df is just [RFC PATCH V3 1/1] sched/numa: Fix disjoint set vma scan regression) ========================================================================================= compiler/cpufreq_governor/iterations/kconfig/rootfs/tbox_group/test/testcase: gcc-11/performance/4x/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-spr-r02/numa01_THREAD_ALLOC/autonuma-benchmark ef6a22b70f6d9044 55fd15913b18d6a790c17d947df ---------------- --------------------------- %stddev %change %stddev \ | \ 193.14 +5.8% 204.37 ± 3% autonuma-benchmark.numa01.seconds 8.14 -0.9% 8.06 autonuma-benchmark.numa02.seconds detail comparison as below: ========================================================================================= compiler/cpufreq_governor/iterations/kconfig/rootfs/tbox_group/test/testcase: gcc-11/performance/4x/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-spr-r02/numa01_THREAD_ALLOC/autonuma-benchmark ef6a22b70f6d9044 55fd15913b18d6a790c17d947df ---------------- --------------------------- %stddev %change %stddev \ | \ 1.185e+10 ± 6% +43.9% 1.705e+10 ± 6% cpuidle..time 12184642 ± 6% +44.0% 17542134 ± 6% cpuidle..usage 5.00 ±100% -100.0% 0.00 dmesg.bootstage:last 189.35 ± 11% +32.1% 250.07 ± 20% dmesg.timestamp:last 1021 +12.6% 1149 ± 3% uptime.boot 21665 ± 6% +50.5% 32596 ± 22% uptime.idle 47.15 ± 16% +84.9% 87.19 ± 53% boot-time.boot 42.12 ± 17% +88.2% 79.25 ± 58% boot-time.dhcp 9448 ± 18% +60.5% 15162 ± 51% boot-time.idle 6.93 +267.2% 25.45 ± 76% boot-time.smp_boot 5.43 ± 5% +1.8 7.21 ± 5% mpstat.cpu.all.idle% 0.02 ± 15% -0.0 0.01 ± 10% mpstat.cpu.all.iowait% 1.08 -0.1 1.03 mpstat.cpu.all.irq% 0.06 ± 2% +0.0 0.08 ± 2% mpstat.cpu.all.soft% 1.99 -0.2 1.81 ± 2% mpstat.cpu.all.sys% 91.41 -1.6 89.86 mpstat.cpu.all.usr% 0.00 -100.0% 0.00 numa-numastat.node0.interleave_hit 1912287 ± 6% -3.9% 1837158 ± 13% numa-numastat.node0.local_node 2592838 ± 5% -9.8% 2338452 ± 13% numa-numastat.node0.numa_hit 680538 ± 6% -26.3% 501281 ± 30% numa-numastat.node0.other_node 0.00 -100.0% 0.00 numa-numastat.node1.interleave_hit 1890172 ± 6% +11.2% 2101474 ± 14% numa-numastat.node1.local_node 2436852 ± 5% +15.4% 2811758 ± 13% numa-numastat.node1.numa_hit 546653 ± 6% +29.9% 710290 ± 23% numa-numastat.node1.other_node 970.73 +9.1% 1059 ± 2% time.elapsed_time 970.73 +9.1% 1059 ± 2% time.elapsed_time.max 624931 +4.9% 655561 ± 2% time.involuntary_context_switches 25.00 ± 21% +8.7% 27.17 ± 19% time.major_page_faults 3150370 +0.0% 3150530 time.maximum_resident_set_size 2514815 -0.6% 2498715 ± 2% time.minor_page_faults 4096 +0.0% 4096 time.page_size 20893 -1.9% 20499 time.percent_of_cpu_this_job_got 3931 -0.8% 3899 time.system_time 198882 +7.2% 213298 ± 2% time.user_time 37974 ± 13% -21.0% 29982 ± 7% time.voluntary_context_switches 5.17 ± 7% +35.5% 7.00 ± 8% vmstat.cpu.id 2.50 ± 20% -20.0% 2.00 vmstat.cpu.sy 90.83 -1.8% 89.17 vmstat.cpu.us 0.00 -100.0% 0.00 vmstat.cpu.wa 1.67 ± 44% -40.0% 1.00 ±100% vmstat.io.bi 4.00 +0.0% 4.00 vmstat.memory.buff 3256207 ± 3% +3.6% 3373341 vmstat.memory.cache 2.524e+08 -0.0% 2.523e+08 vmstat.memory.free 0.00 -100.0% 0.00 vmstat.procs.b 212.17 -1.6% 208.83 vmstat.procs.r 2330 ± 7% -7.8% 2148 vmstat.system.cs 233033 -1.3% 229952 vmstat.system.in 193.14 +5.8% 204.37 ± 3% autonuma-benchmark.numa01.seconds 40.49 ± 3% +27.3% 51.53 ± 4% autonuma-benchmark.numa01_THREAD_ALLOC.seconds 8.14 -0.9% 8.06 autonuma-benchmark.numa02.seconds 970.73 +9.1% 1059 ± 2% autonuma-benchmark.time.elapsed_time 970.73 +9.1% 1059 ± 2% autonuma-benchmark.time.elapsed_time.max 624931 +4.9% 655561 ± 2% autonuma-benchmark.time.involuntary_context_switches 25.00 ± 21% +8.7% 27.17 ± 19% autonuma-benchmark.time.major_page_faults 3150370 +0.0% 3150530 autonuma-benchmark.time.maximum_resident_set_size 2514815 -0.6% 2498715 ± 2% autonuma-benchmark.time.minor_page_faults 4096 +0.0% 4096 autonuma-benchmark.time.page_size 20893 -1.9% 20499 autonuma-benchmark.time.percent_of_cpu_this_job_got 3931 -0.8% 3899 autonuma-benchmark.time.system_time 198882 +7.2% 213298 ± 2% autonuma-benchmark.time.user_time 37974 ± 13% -21.0% 29982 ± 7% autonuma-benchmark.time.voluntary_context_switches 2737 -1.8% 2687 turbostat.Avg_MHz 94.39 -1.7 92.66 turbostat.Busy% 2900 +0.0% 2900 turbostat.Bzy_MHz 12957 ± 36% +15.9% 15015 ± 25% turbostat.C1 5540462 ± 6% +46.1% 8093574 ± 6% turbostat.C1E 2.48 ± 5% +0.8 3.30 ± 6% turbostat.C1E% 6453445 ± 6% +43.6% 9265716 ± 6% turbostat.C6 2.92 ± 5% +0.9 3.84 ± 6% turbostat.C6% 5.61 ± 5% +30.8% 7.34 ± 5% turbostat.CPU%c1 43.17 ± 7% +6.2% 45.83 ± 5% turbostat.CoreTmp 2.307e+08 +6.0% 2.446e+08 ± 2% turbostat.IRQ 7.90 ± 31% +5.7 13.56 ± 32% turbostat.PKG_% 28989 ± 36% -32.3% 19619 ± 11% turbostat.POLL 43.17 ± 7% +6.6% 46.00 ± 6% turbostat.PkgTmp 666.52 -0.1% 665.80 turbostat.PkgWatt 70.96 -2.9% 68.89 turbostat.RAMWatt 2000 +0.0% 2000 turbostat.TSC_MHz 237031 ± 49% +47.9% 350630 ± 3% meminfo.Active 236935 ± 49% +47.9% 350502 ± 3% meminfo.Active(anon) 96.00 ± 33% +33.3% 128.00 meminfo.Active(file) 6155451 +1.0% 6215315 meminfo.AnonHugePages 6257767 +0.4% 6284278 meminfo.AnonPages 4.00 +0.0% 4.00 meminfo.Buffers 3088930 ± 3% +4.0% 3211320 meminfo.Cached 1.318e+08 -0.0% 1.318e+08 meminfo.CommitLimit 8554815 +1.6% 8690945 meminfo.Committed_AS 2.567e+08 -0.1% 2.566e+08 meminfo.DirectMap1G 10951680 ± 3% +17.0% 12811605 ± 5% meminfo.DirectMap2M 2437198 -69.1% 752036 ± 14% meminfo.DirectMap4k 2048 +0.0% 2048 meminfo.Hugepagesize 6294588 +0.5% 6326486 meminfo.Inactive 6294432 +0.5% 6326322 meminfo.Inactive(anon) 156.00 ± 5% +5.0% 163.83 meminfo.Inactive(file) 167387 -3.1% 162160 meminfo.KReclaimable 39732 +0.0% 39751 meminfo.KernelStack 87203 ± 6% +12.8% 98392 ± 14% meminfo.Mapped 2.514e+08 -0.0% 2.513e+08 meminfo.MemAvailable 2.524e+08 -0.0% 2.523e+08 meminfo.MemFree 2.637e+08 -0.0% 2.637e+08 meminfo.MemTotal 11248567 +0.8% 11333611 meminfo.Memused 1604 +4.3% 1673 meminfo.Mlocked 16758 +0.3% 16803 meminfo.PageTables 110655 -7.3% 102584 meminfo.Percpu 167387 -3.1% 162160 meminfo.SReclaimable 430222 -0.4% 428319 meminfo.SUnreclaim 273371 ± 42% +43.6% 392445 meminfo.Shmem 597610 -1.2% 590480 meminfo.Slab 2815544 +0.1% 2818830 meminfo.Unevictable 1.374e+13 +0.0% 1.374e+13 meminfo.VmallocTotal 338296 -17.2% 280268 meminfo.VmallocUsed 12911235 -0.2% 12891540 meminfo.max_used_kB 29565 ±175% +18.3% 34976 ±214% numa-meminfo.node0.Active 29522 ±175% +18.3% 34912 ±214% numa-meminfo.node0.Active(anon) 42.67 ±111% +49.6% 63.83 ±100% numa-meminfo.node0.Active(file) 3202041 ± 2% -8.3% 2935773 ± 10% numa-meminfo.node0.AnonHugePages 3256105 ± 2% -8.9% 2967039 ± 10% numa-meminfo.node0.AnonPages 6450186 -0.6% 6408448 numa-meminfo.node0.AnonPages.max 2728164 ± 2% -13.5% 2360218 ± 41% numa-meminfo.node0.FilePages 3266857 ± 2% -8.9% 2976525 ± 10% numa-meminfo.node0.Inactive 3266783 ± 2% -8.9% 2976443 ± 10% numa-meminfo.node0.Inactive(anon) 74.00 ±100% +10.6% 81.83 ± 93% numa-meminfo.node0.Inactive(file) 101285 ± 9% -9.0% 92138 ± 27% numa-meminfo.node0.KReclaimable 20573 ± 3% -2.1% 20149 ± 8% numa-meminfo.node0.KernelStack 51364 ± 7% -5.0% 48807 ± 35% numa-meminfo.node0.Mapped 1.246e+08 +0.7% 1.254e+08 numa-meminfo.node0.MemFree 1.316e+08 +0.0% 1.316e+08 numa-meminfo.node0.MemTotal 7033196 ± 2% -11.9% 6193901 ± 15% numa-meminfo.node0.MemUsed 1487 ± 7% -6.6% 1388 ± 37% numa-meminfo.node0.Mlocked 9628 ± 12% -26.6% 7070 ± 36% numa-meminfo.node0.PageTables 101285 ± 9% -9.0% 92138 ± 27% numa-meminfo.node0.SReclaimable 237810 ± 7% -6.4% 222602 ± 5% numa-meminfo.node0.SUnreclaim 39839 ±139% +11.9% 44566 ±194% numa-meminfo.node0.Shmem 339095 ± 5% -7.2% 314741 ± 10% numa-meminfo.node0.Slab 2688324 ± 2% -13.9% 2315704 ± 41% numa-meminfo.node0.Unevictable 207469 ± 53% +52.2% 315756 ± 25% numa-meminfo.node1.Active 207416 ± 53% +52.2% 315692 ± 25% numa-meminfo.node1.Active(anon) 53.33 ±107% +20.0% 64.00 ±100% numa-meminfo.node1.Active(file) 2949624 ± 3% +11.1% 3278227 ± 9% numa-meminfo.node1.AnonHugePages 2997887 ± 2% +10.6% 3315853 ± 9% numa-meminfo.node1.AnonPages 6474808 +0.0% 6477523 numa-meminfo.node1.AnonPages.max 360720 ± 40% +136.0% 851138 ±114% numa-meminfo.node1.FilePages 3024066 ± 2% +10.7% 3348368 ± 9% numa-meminfo.node1.Inactive 3023983 ± 2% +10.7% 3348286 ± 9% numa-meminfo.node1.Inactive(anon) 82.00 ± 93% -0.2% 81.83 ± 93% numa-meminfo.node1.Inactive(file) 66102 ± 14% +5.9% 70021 ± 36% numa-meminfo.node1.KReclaimable 19151 ± 3% +2.3% 19600 ± 8% numa-meminfo.node1.KernelStack 35805 ± 13% +38.0% 49423 ± 17% numa-meminfo.node1.Mapped 1.278e+08 -0.7% 1.269e+08 numa-meminfo.node1.MemFree 1.32e+08 -0.0% 1.32e+08 numa-meminfo.node1.MemTotal 4211780 ± 6% +22.0% 5137530 ± 18% numa-meminfo.node1.MemUsed 116.67 ±100% +145.4% 286.33 ±183% numa-meminfo.node1.Mlocked 7122 ± 16% +36.6% 9729 ± 27% numa-meminfo.node1.PageTables 66102 ± 14% +5.9% 70021 ± 36% numa-meminfo.node1.SReclaimable 192408 ± 9% +6.9% 205718 ± 5% numa-meminfo.node1.SUnreclaim 233482 ± 47% +49.0% 347910 ± 24% numa-meminfo.node1.Shmem 258511 ± 6% +6.7% 275740 ± 11% numa-meminfo.node1.Slab 127219 ± 57% +295.5% 503125 ±189% numa-meminfo.node1.Unevictable 7390 ±175% +18.1% 8730 ±214% numa-vmstat.node0.nr_active_anon 10.67 ±111% +48.4% 15.83 ±100% numa-vmstat.node0.nr_active_file 812998 ± 2% -8.8% 741650 ± 10% numa-vmstat.node0.nr_anon_pages 1560 ± 2% -8.2% 1432 ± 10% numa-vmstat.node0.nr_anon_transparent_hugepages 682054 ± 2% -13.5% 590056 ± 41% numa-vmstat.node0.nr_file_pages 31146566 +0.7% 31355317 numa-vmstat.node0.nr_free_pages 815598 ± 2% -8.8% 744048 ± 10% numa-vmstat.node0.nr_inactive_anon 18.50 ±100% +9.9% 20.33 ± 94% numa-vmstat.node0.nr_inactive_file 40.50 ± 14% -62.1% 15.33 ± 39% numa-vmstat.node0.nr_isolated_anon 20572 ± 3% -2.1% 20150 ± 8% numa-vmstat.node0.nr_kernel_stack 12953 ± 7% -4.8% 12333 ± 35% numa-vmstat.node0.nr_mapped 371.33 ± 7% -6.6% 346.67 ± 37% numa-vmstat.node0.nr_mlock 2406 ± 12% -26.6% 1767 ± 36% numa-vmstat.node0.nr_page_table_pages 9973 ±139% +11.7% 11143 ±194% numa-vmstat.node0.nr_shmem 25321 ± 9% -9.0% 23034 ± 27% numa-vmstat.node0.nr_slab_reclaimable 59452 ± 7% -6.4% 55651 ± 5% numa-vmstat.node0.nr_slab_unreclaimable 672080 ± 2% -13.9% 578925 ± 41% numa-vmstat.node0.nr_unevictable 7390 ±175% +18.1% 8730 ±214% numa-vmstat.node0.nr_zone_active_anon 10.67 ±111% +48.4% 15.83 ±100% numa-vmstat.node0.nr_zone_active_file 815599 ± 2% -8.8% 744047 ± 10% numa-vmstat.node0.nr_zone_inactive_anon 18.50 ±100% +9.9% 20.33 ± 94% numa-vmstat.node0.nr_zone_inactive_file 672080 ± 2% -13.9% 578925 ± 41% numa-vmstat.node0.nr_zone_unevictable 2593013 ± 5% -9.8% 2338261 ± 13% numa-vmstat.node0.numa_hit 0.00 -100.0% 0.00 numa-vmstat.node0.numa_interleave 1912463 ± 6% -3.9% 1836967 ± 13% numa-vmstat.node0.numa_local 680538 ± 6% -26.3% 501281 ± 30% numa-vmstat.node0.numa_other 51877 ± 53% +52.2% 78947 ± 25% numa-vmstat.node1.nr_active_anon 13.33 ±107% +20.0% 16.00 ±100% numa-vmstat.node1.nr_active_file 750017 ± 2% +10.6% 829613 ± 9% numa-vmstat.node1.nr_anon_pages 1440 ± 3% +11.1% 1601 ± 9% numa-vmstat.node1.nr_anon_transparent_hugepages 90208 ± 40% +135.9% 212803 ±114% numa-vmstat.node1.nr_file_pages 31957700 -0.7% 31726112 numa-vmstat.node1.nr_free_pages 756559 ± 2% +10.7% 837739 ± 9% numa-vmstat.node1.nr_inactive_anon 20.50 ± 93% -0.8% 20.33 ± 93% numa-vmstat.node1.nr_inactive_file 37.00 ± 9% -26.6% 27.17 ± 26% numa-vmstat.node1.nr_isolated_anon 19147 ± 3% +2.3% 19596 ± 8% numa-vmstat.node1.nr_kernel_stack 9012 ± 13% +37.6% 12401 ± 17% numa-vmstat.node1.nr_mapped 29.00 ±100% +146.6% 71.50 ±184% numa-vmstat.node1.nr_mlock 1779 ± 16% +36.7% 2433 ± 27% numa-vmstat.node1.nr_page_table_pages 58399 ± 47% +49.0% 86996 ± 24% numa-vmstat.node1.nr_shmem 16525 ± 14% +5.9% 17505 ± 36% numa-vmstat.node1.nr_slab_reclaimable 48100 ± 9% +6.9% 51428 ± 5% numa-vmstat.node1.nr_slab_unreclaimable 31804 ± 57% +295.5% 125781 ±189% numa-vmstat.node1.nr_unevictable 51877 ± 53% +52.2% 78946 ± 25% numa-vmstat.node1.nr_zone_active_anon 13.33 ±107% +20.0% 16.00 ±100% numa-vmstat.node1.nr_zone_active_file 756560 ± 2% +10.7% 837741 ± 9% numa-vmstat.node1.nr_zone_inactive_anon 20.50 ± 93% -0.8% 20.33 ± 93% numa-vmstat.node1.nr_zone_inactive_file 31804 ± 57% +295.5% 125781 ±189% numa-vmstat.node1.nr_zone_unevictable 2436786 ± 5% +15.4% 2811280 ± 13% numa-vmstat.node1.numa_hit 0.00 -100.0% 0.00 numa-vmstat.node1.numa_interleave 1890106 ± 6% +11.2% 2100996 ± 14% numa-vmstat.node1.numa_local 546653 ± 6% +29.9% 710290 ± 23% numa-vmstat.node1.numa_other 1159 -70.9% 336.83 ± 15% proc-vmstat.direct_map_level2_splits 6.17 ± 17% -2.7% 6.00 ± 19% proc-vmstat.direct_map_level3_splits 59245 ± 49% +48.0% 87655 ± 3% proc-vmstat.nr_active_anon 24.00 ± 33% +33.3% 32.00 proc-vmstat.nr_active_file 1562608 +0.5% 1570691 proc-vmstat.nr_anon_pages 3001 +1.1% 3033 proc-vmstat.nr_anon_transparent_hugepages 6272807 -0.0% 6270545 proc-vmstat.nr_dirty_background_threshold 12560952 -0.0% 12556422 proc-vmstat.nr_dirty_threshold 772234 ± 3% +4.0% 802849 proc-vmstat.nr_file_pages 63104657 -0.0% 63082004 proc-vmstat.nr_free_pages 1571739 +0.6% 1581204 proc-vmstat.nr_inactive_anon 39.00 ± 5% +4.7% 40.83 proc-vmstat.nr_inactive_file 77.67 ± 6% -50.9% 38.17 ± 15% proc-vmstat.nr_isolated_anon 39726 +0.1% 39748 proc-vmstat.nr_kernel_stack 21963 ± 6% +12.8% 24764 ± 14% proc-vmstat.nr_mapped 400.67 +4.4% 418.33 proc-vmstat.nr_mlock 4186 +0.3% 4199 proc-vmstat.nr_page_table_pages 68343 ± 42% +43.6% 98129 proc-vmstat.nr_shmem 41846 -3.1% 40539 proc-vmstat.nr_slab_reclaimable 107554 -0.4% 107079 proc-vmstat.nr_slab_unreclaimable 703885 +0.1% 704707 proc-vmstat.nr_unevictable 59245 ± 49% +48.0% 87655 ± 3% proc-vmstat.nr_zone_active_anon 24.00 ± 33% +33.3% 32.00 proc-vmstat.nr_zone_active_file 1571739 +0.6% 1581204 proc-vmstat.nr_zone_inactive_anon 39.00 ± 5% +4.7% 40.83 proc-vmstat.nr_zone_inactive_file 703885 +0.1% 704707 proc-vmstat.nr_zone_unevictable 200540 ± 6% -18.2% 164021 ± 7% proc-vmstat.numa_hint_faults 120494 ± 5% -10.8% 107425 ± 10% proc-vmstat.numa_hint_faults_local 5032020 +2.4% 5152018 proc-vmstat.numa_hit 112746 ± 2% -46.1% 60803 ± 4% proc-vmstat.numa_huge_pte_updates 0.00 -100.0% 0.00 proc-vmstat.numa_interleave 3804791 +3.6% 3940440 proc-vmstat.numa_local 1227192 -1.3% 1211572 ± 2% proc-vmstat.numa_other 11866841 ± 4% -40.8% 7027448 ± 2% proc-vmstat.numa_pages_migrated 57894552 ± 2% -45.9% 31298933 ± 4% proc-vmstat.numa_pte_updates 74510 ± 14% +10.3% 82160 ± 18% proc-vmstat.pgactivate 0.00 -100.0% 0.00 proc-vmstat.pgalloc_dma32 4.537e+08 -1.7% 4.462e+08 proc-vmstat.pgalloc_normal 4907690 +3.1% 5060927 proc-vmstat.pgfault 4.537e+08 -1.7% 4.461e+08 proc-vmstat.pgfree 86.17 ±221% -94.6% 4.67 ± 91% proc-vmstat.pgmigrate_fail 11866841 ± 4% -40.8% 7027448 ± 2% proc-vmstat.pgmigrate_success 2376 ± 44% -40.0% 1426 ±100% proc-vmstat.pgpgin 172880 +8.8% 188160 ± 2% proc-vmstat.pgreuse 98.67 ± 18% +24.8% 123.17 ± 9% proc-vmstat.thp_collapse_alloc 14.67 ± 35% +0.0% 14.67 ± 28% proc-vmstat.thp_deferred_split_page 26654 +0.0% 26654 proc-vmstat.thp_fault_alloc 0.17 ±223% -100.0% 0.00 proc-vmstat.thp_migration_fail 23128 ± 4% -40.8% 13682 ± 2% proc-vmstat.thp_migration_success 14.67 ± 35% +0.0% 14.67 ± 28% proc-vmstat.thp_split_pmd 0.00 -100.0% 0.00 proc-vmstat.thp_zero_page_alloc 1005 ± 12% -12.2% 883.33 proc-vmstat.unevictable_pgs_culled 607.00 -0.1% 606.67 proc-vmstat.unevictable_pgs_mlocked 4.00 ± 14% -8.3% 3.67 ± 20% proc-vmstat.unevictable_pgs_munlocked 3.83 ± 17% -13.0% 3.33 ± 14% proc-vmstat.unevictable_pgs_rescued 7602432 +8.8% 8273408 ± 3% proc-vmstat.unevictable_pgs_scanned 1459346 -100.0% 0.00 proc-vmstat.vma_lock_abort 2043 ± 47% -100.0% 0.00 proc-vmstat.vma_lock_retry 1755990 -100.0% 0.00 proc-vmstat.vma_lock_success 54.87 ± 10% -28.5% 39.24 ± 5% perf-stat.i.MPKI 4.041e+08 -4.1% 3.877e+08 perf-stat.i.branch-instructions 0.62 -0.0 0.61 perf-stat.i.branch-miss-rate% 1799541 -2.6% 1752689 perf-stat.i.branch-misses 29.53 ± 2% -3.0 26.50 ± 2% perf-stat.i.cache-miss-rate% 52797131 ± 10% -27.3% 38359610 ± 4% perf-stat.i.cache-misses 1.172e+08 ± 6% -20.3% 93439529 ± 2% perf-stat.i.cache-references 2294 ± 7% -7.8% 2116 perf-stat.i.context-switches 541.44 -2.3% 529.18 perf-stat.i.cpi 224169 -0.0% 224149 perf-stat.i.cpu-clock 6.06e+11 -1.7% 5.954e+11 perf-stat.i.cpu-cycles 233.65 -1.7% 229.68 perf-stat.i.cpu-migrations 94785 ± 2% +16.9% 110768 ± 3% perf-stat.i.cycles-between-cache-misses 0.07 ± 9% +0.0 0.10 ± 8% perf-stat.i.dTLB-load-miss-rate% 224367 ± 12% +51.5% 339894 ± 8% perf-stat.i.dTLB-load-misses 4.949e+08 -3.1% 4.795e+08 perf-stat.i.dTLB-loads 0.60 -0.0 0.58 perf-stat.i.dTLB-store-miss-rate% 1041912 -1.8% 1022937 perf-stat.i.dTLB-store-misses 1.832e+08 +0.6% 1.844e+08 perf-stat.i.dTLB-stores 1.893e+09 -3.3% 1.83e+09 perf-stat.i.instructions 0.00 ± 3% -4.0% 0.00 perf-stat.i.ipc 0.11 ± 19% -22.4% 0.09 ± 12% perf-stat.i.major-faults 2.70 -1.7% 2.66 perf-stat.i.metric.GHz 1516 ± 2% +0.7% 1526 perf-stat.i.metric.K/sec 3.79 ± 2% -6.3% 3.56 perf-stat.i.metric.M/sec 5088 -5.0% 4834 ± 2% perf-stat.i.minor-faults 49.22 -1.9 47.29 perf-stat.i.node-load-miss-rate% 965573 ± 3% -25.0% 724436 ± 3% perf-stat.i.node-load-misses 740879 ± 8% -2.0% 725851 ± 9% perf-stat.i.node-loads 5088 -5.0% 4835 ± 2% perf-stat.i.page-faults 224169 -0.0% 224149 perf-stat.i.task-clock 59.37 ± 5% -17.5% 48.98 perf-stat.overall.MPKI 0.45 +0.0 0.46 perf-stat.overall.branch-miss-rate% 44.29 ± 4% -3.9 40.39 ± 2% perf-stat.overall.cache-miss-rate% 332.08 +1.9% 338.53 perf-stat.overall.cpi 12720 ± 10% +34.7% 17135 ± 4% perf-stat.overall.cycles-between-cache-misses 0.05 ± 18% +0.0 0.07 ± 7% perf-stat.overall.dTLB-load-miss-rate% 0.57 -0.0 0.56 perf-stat.overall.dTLB-store-miss-rate% 0.00 -1.9% 0.00 perf-stat.overall.ipc 55.05 ± 6% -6.7 48.39 ± 4% perf-stat.overall.node-load-miss-rate% 3.897e+08 -4.4% 3.726e+08 perf-stat.ps.branch-instructions 1760709 -2.7% 1713424 perf-stat.ps.branch-misses 48311108 ± 9% -27.6% 34960590 ± 4% perf-stat.ps.cache-misses 1.089e+08 ± 5% -20.5% 86529746 perf-stat.ps.cache-references 2280 ± 7% -7.7% 2105 perf-stat.ps.context-switches 223740 +0.0% 223765 perf-stat.ps.cpu-clock 6.088e+11 -1.8% 5.98e+11 perf-stat.ps.cpu-cycles 230.49 -1.6% 226.70 perf-stat.ps.cpu-migrations 223824 ± 18% +49.2% 333910 ± 7% perf-stat.ps.dTLB-load-misses 4.8e+08 -3.4% 4.638e+08 perf-stat.ps.dTLB-loads 1041413 -2.0% 1020849 perf-stat.ps.dTLB-store-misses 1.812e+08 +0.6% 1.823e+08 perf-stat.ps.dTLB-stores 1.833e+09 -3.6% 1.767e+09 perf-stat.ps.instructions 0.11 ± 18% -21.9% 0.08 ± 13% perf-stat.ps.major-faults 4861 -5.2% 4608 ± 2% perf-stat.ps.minor-faults 971538 ± 5% -25.0% 728715 ± 5% perf-stat.ps.node-load-misses 795675 ± 10% -1.8% 781081 ± 10% perf-stat.ps.node-loads 4861 -5.2% 4608 ± 2% perf-stat.ps.page-faults 223740 +0.0% 223765 perf-stat.ps.task-clock 1.781e+12 +5.2% 1.873e+12 perf-stat.total.instructions 32005 ± 86% +12.8% 36117 ±135% sched_debug.cfs_rq:/.MIN_vruntime.avg 6653945 ± 94% +15.2% 7664854 ±135% sched_debug.cfs_rq:/.MIN_vruntime.max 0.00 +0.0% 0.00 sched_debug.cfs_rq:/.MIN_vruntime.min 456998 ± 91% +14.7% 524024 ±135% sched_debug.cfs_rq:/.MIN_vruntime.stddev 0.92 ± 2% -1.1% 0.91 ± 5% sched_debug.cfs_rq:/.h_nr_running.avg 1.90 ± 4% -6.2% 1.78 ± 3% sched_debug.cfs_rq:/.h_nr_running.max 0.67 ± 18% -5.0% 0.64 ± 20% sched_debug.cfs_rq:/.h_nr_running.min 0.15 ± 12% +0.4% 0.15 ± 24% sched_debug.cfs_rq:/.h_nr_running.stddev 5652 ± 6% -3.0% 5483 ± 13% sched_debug.cfs_rq:/.load.avg 250428 ± 13% -14.4% 214489 ± 49% sched_debug.cfs_rq:/.load.max 1548 ± 26% -8.3% 1420 ± 20% sched_debug.cfs_rq:/.load.min 18136 ± 14% -12.5% 15867 ± 46% sched_debug.cfs_rq:/.load.stddev 7.26 ± 9% -0.6% 7.22 ± 12% sched_debug.cfs_rq:/.load_avg.avg 215.19 ± 20% +3.0% 221.67 ± 22% sched_debug.cfs_rq:/.load_avg.max 1.55 ± 18% -9.0% 1.41 ± 21% sched_debug.cfs_rq:/.load_avg.min 22.20 ± 15% +1.2% 22.46 ± 13% sched_debug.cfs_rq:/.load_avg.stddev 32005 ± 86% +12.8% 36117 ±135% sched_debug.cfs_rq:/.max_vruntime.avg 6653945 ± 94% +15.2% 7664854 ±135% sched_debug.cfs_rq:/.max_vruntime.max 0.00 +0.0% 0.00 sched_debug.cfs_rq:/.max_vruntime.min 456998 ± 91% +14.7% 524024 ±135% sched_debug.cfs_rq:/.max_vruntime.stddev 1.203e+08 ± 5% +11.8% 1.345e+08 ± 2% sched_debug.cfs_rq:/.min_vruntime.avg 1.231e+08 ± 5% +12.5% 1.385e+08 ± 2% sched_debug.cfs_rq:/.min_vruntime.max 99360296 ± 6% +9.4% 1.087e+08 ± 4% sched_debug.cfs_rq:/.min_vruntime.min 2529598 ± 9% +32.7% 3357930 ± 15% sched_debug.cfs_rq:/.min_vruntime.stddev 0.91 ± 2% -0.9% 0.90 ± 5% sched_debug.cfs_rq:/.nr_running.avg 1.10 ± 7% -3.8% 1.06 ± 6% sched_debug.cfs_rq:/.nr_running.max 0.67 ± 18% -5.0% 0.64 ± 20% sched_debug.cfs_rq:/.nr_running.min 0.08 ± 39% +16.1% 0.09 ± 51% sched_debug.cfs_rq:/.nr_running.stddev 1.68 ± 48% -5.5% 1.59 ± 64% sched_debug.cfs_rq:/.removed.load_avg.avg 76.51 ± 34% -0.4% 76.20 ± 34% sched_debug.cfs_rq:/.removed.load_avg.max 10.52 ± 17% -6.9% 9.79 ± 26% sched_debug.cfs_rq:/.removed.load_avg.stddev 0.82 ± 50% -7.1% 0.76 ± 65% sched_debug.cfs_rq:/.removed.runnable_avg.avg 37.12 ± 24% +1.1% 37.52 ± 31% sched_debug.cfs_rq:/.removed.runnable_avg.max 5.08 ± 19% -7.5% 4.70 ± 28% sched_debug.cfs_rq:/.removed.runnable_avg.stddev 0.81 ± 52% -7.1% 0.76 ± 66% sched_debug.cfs_rq:/.removed.util_avg.avg 35.30 ± 14% +1.5% 35.83 ± 30% sched_debug.cfs_rq:/.removed.util_avg.max 4.96 ± 22% -7.5% 4.59 ± 30% sched_debug.cfs_rq:/.removed.util_avg.stddev 944.71 ± 2% -1.1% 934.57 ± 5% sched_debug.cfs_rq:/.runnable_avg.avg 1881 ± 3% -6.7% 1755 ± 6% sched_debug.cfs_rq:/.runnable_avg.max 643.11 ± 13% -3.7% 619.05 ± 23% sched_debug.cfs_rq:/.runnable_avg.min 141.77 ± 13% +3.3% 146.48 ± 25% sched_debug.cfs_rq:/.runnable_avg.stddev 0.00 ±100% +78.5% 0.00 ±153% sched_debug.cfs_rq:/.spread.avg 0.38 ±100% +78.5% 0.68 ±153% sched_debug.cfs_rq:/.spread.max 0.03 ±100% +78.5% 0.05 ±153% sched_debug.cfs_rq:/.spread.stddev 8911395 ± 18% +77.5% 15821663 ± 33% sched_debug.cfs_rq:/.spread0.avg 11635928 ± 17% +70.0% 19782673 ± 27% sched_debug.cfs_rq:/.spread0.max -12145667 -17.2% -10050665 sched_debug.cfs_rq:/.spread0.min 2510183 ± 10% +33.3% 3345922 ± 15% sched_debug.cfs_rq:/.spread0.stddev 934.78 ± 2% -1.1% 924.53 ± 5% sched_debug.cfs_rq:/.util_avg.avg 1278 ± 5% -6.2% 1198 ± 2% sched_debug.cfs_rq:/.util_avg.max 515.26 ± 12% -5.7% 485.68 ± 22% sched_debug.cfs_rq:/.util_avg.min 95.53 ± 24% +10.5% 105.53 ± 40% sched_debug.cfs_rq:/.util_avg.stddev 441.09 ± 8% -21.8% 345.05 ± 11% sched_debug.cfs_rq:/.util_est_enqueued.avg 1272 ± 6% -10.8% 1135 ± 9% sched_debug.cfs_rq:/.util_est_enqueued.max 280.15 ± 9% +14.2% 319.98 ± 9% sched_debug.cfs_rq:/.util_est_enqueued.stddev 1308353 ± 3% +22.3% 1600451 ± 3% sched_debug.cpu.avg_idle.avg 2922459 ± 2% +10.5% 3230277 ± 5% sched_debug.cpu.avg_idle.max 448661 ± 15% +36.4% 611757 ± 13% sched_debug.cpu.avg_idle.min 412548 ± 5% +38.1% 569757 ± 7% sched_debug.cpu.avg_idle.stddev 505223 ± 4% +20.0% 606292 ± 5% sched_debug.cpu.clock.avg 505890 ± 4% +20.0% 606955 ± 5% sched_debug.cpu.clock.max 504409 ± 4% +20.1% 605556 ± 5% sched_debug.cpu.clock.min 415.73 ± 25% -4.3% 397.71 ± 28% sched_debug.cpu.clock.stddev 499837 ± 4% +19.8% 598837 ± 5% sched_debug.cpu.clock_task.avg 501130 ± 4% +19.9% 600703 ± 5% sched_debug.cpu.clock_task.max 478575 ± 4% +12.9% 540404 ± 2% sched_debug.cpu.clock_task.min 1619 ± 20% +150.3% 4053 ± 73% sched_debug.cpu.clock_task.stddev 17029 ± 6% +5.7% 17999 ± 7% sched_debug.cpu.curr->pid.avg 19932 ± 3% +7.9% 21502 sched_debug.cpu.curr->pid.max 9932 ± 24% +3.7% 10302 ± 27% sched_debug.cpu.curr->pid.min 1540 ± 44% +20.8% 1861 ± 55% sched_debug.cpu.curr->pid.stddev 591935 ± 2% +13.8% 673677 ± 4% sched_debug.cpu.max_idle_balance_cost.avg 1240997 ± 5% +12.9% 1400895 ± 4% sched_debug.cpu.max_idle_balance_cost.max 500000 +0.0% 500000 sched_debug.cpu.max_idle_balance_cost.min 143032 ± 7% +41.5% 202328 ± 5% sched_debug.cpu.max_idle_balance_cost.stddev 4295 +0.0% 4295 sched_debug.cpu.next_balance.avg 4295 +0.0% 4295 sched_debug.cpu.next_balance.max 4295 +0.0% 4295 sched_debug.cpu.next_balance.min 0.00 ± 25% -4.1% 0.00 ± 27% sched_debug.cpu.next_balance.stddev 0.91 ± 3% -2.1% 0.89 ± 6% sched_debug.cpu.nr_running.avg 1.92 ± 5% -5.3% 1.82 ± 2% sched_debug.cpu.nr_running.max 0.68 ± 16% -5.2% 0.64 ± 19% sched_debug.cpu.nr_running.min 0.16 ± 16% +4.0% 0.16 ± 26% sched_debug.cpu.nr_running.stddev 5837 ± 5% +4.8% 6118 sched_debug.cpu.nr_switches.avg 49744 ± 20% +13.8% 56608 ± 24% sched_debug.cpu.nr_switches.max 2002 ± 6% +14.9% 2301 ± 2% sched_debug.cpu.nr_switches.min 5880 ± 11% -1.9% 5769 ± 10% sched_debug.cpu.nr_switches.stddev 2.039e+09 ± 3% -100.0% 0.00 ± 76% sched_debug.cpu.nr_uninterruptible.avg 4.295e+09 -100.0% 46.24 ± 39% sched_debug.cpu.nr_uninterruptible.max 0.00 -6.3e+103% -63.50 sched_debug.cpu.nr_uninterruptible.min 2.14e+09 -100.0% 8.63 ± 48% sched_debug.cpu.nr_uninterruptible.stddev 504394 ± 4% +20.1% 605544 ± 5% sched_debug.cpu_clk 996147 +0.0% 996147 sched_debug.dl_rq:.dl_bw->bw.avg 996147 +0.0% 996147 sched_debug.dl_rq:.dl_bw->bw.max 996147 +0.0% 996147 sched_debug.dl_rq:.dl_bw->bw.min 4.295e+09 +0.0% 4.295e+09 sched_debug.jiffies 499768 ± 4% +20.9% 604335 ± 5% sched_debug.ktime 0.00 ± 19% -4.0% 0.00 ± 2% sched_debug.rt_rq:.rt_nr_migratory.avg 0.12 ± 19% -4.0% 0.11 ± 2% sched_debug.rt_rq:.rt_nr_migratory.max 0.01 ± 19% -4.0% 0.01 ± 2% sched_debug.rt_rq:.rt_nr_migratory.stddev 0.00 ± 19% -4.0% 0.00 ± 2% sched_debug.rt_rq:.rt_nr_running.avg 0.12 ± 19% -4.0% 0.11 ± 2% sched_debug.rt_rq:.rt_nr_running.max 0.01 ± 19% -4.0% 0.01 ± 2% sched_debug.rt_rq:.rt_nr_running.stddev 950.00 +0.0% 950.00 sched_debug.rt_rq:.rt_runtime.avg 950.00 +0.0% 950.00 sched_debug.rt_rq:.rt_runtime.max 950.00 +0.0% 950.00 sched_debug.rt_rq:.rt_runtime.min 0.00 +4.1e+97% 0.00 ±223% sched_debug.rt_rq:.rt_throttled.avg 0.00 +9.3e+99% 0.01 ±223% sched_debug.rt_rq:.rt_throttled.max 0.00 +6.2e+98% 0.00 ±223% sched_debug.rt_rq:.rt_throttled.stddev 0.20 ± 55% +57.2% 0.31 ± 35% sched_debug.rt_rq:.rt_time.avg 44.78 ± 55% +57.2% 70.39 ± 35% sched_debug.rt_rq:.rt_time.max 2.99 ± 55% +57.2% 4.69 ± 35% sched_debug.rt_rq:.rt_time.stddev 505282 ± 3% +11.1% 561614 ± 2% sched_debug.sched_clk 1.00 +0.0% 1.00 sched_debug.sched_clock_stable() 58611259 +0.0% 58611259 sched_debug.sysctl_sched.sysctl_sched_features 0.75 +0.0% 0.75 sched_debug.sysctl_sched.sysctl_sched_idle_min_granularity 24.00 +0.0% 24.00 sched_debug.sysctl_sched.sysctl_sched_latency 3.00 +0.0% 3.00 sched_debug.sysctl_sched.sysctl_sched_min_granularity 1.00 +0.0% 1.00 sched_debug.sysctl_sched.sysctl_sched_tunable_scaling 4.00 +0.0% 4.00 sched_debug.sysctl_sched.sysctl_sched_wakeup_granularity 11.97 ± 60% -8.6 3.38 ± 66% perf-profile.calltrace.cycles-pp.asm_exc_page_fault 11.86 ± 60% -8.6 3.27 ± 68% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault 11.84 ± 60% -8.6 3.25 ± 69% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 11.59 ± 61% -8.5 3.08 ± 71% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 11.58 ± 64% -8.4 3.17 ± 67% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 9.33 ± 69% -8.3 1.06 ±191% perf-profile.calltrace.cycles-pp.do_huge_pmd_numa_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 9.19 ± 69% -8.2 1.01 ±192% perf-profile.calltrace.cycles-pp.migrate_misplaced_page.do_huge_pmd_numa_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 9.19 ± 69% -8.2 1.01 ±192% perf-profile.calltrace.cycles-pp.migrate_pages.migrate_misplaced_page.do_huge_pmd_numa_page.__handle_mm_fault.handle_mm_fault 9.18 ± 69% -8.2 1.01 ±192% perf-profile.calltrace.cycles-pp.migrate_pages_batch.migrate_pages.migrate_misplaced_page.do_huge_pmd_numa_page.__handle_mm_fault 8.62 ± 69% -7.7 0.93 ±191% perf-profile.calltrace.cycles-pp.move_to_new_folio.migrate_pages_batch.migrate_pages.migrate_misplaced_page.do_huge_pmd_numa_page 8.61 ± 69% -7.7 0.93 ±191% perf-profile.calltrace.cycles-pp.migrate_folio_extra.move_to_new_folio.migrate_pages_batch.migrate_pages.migrate_misplaced_page 8.59 ± 69% -7.7 0.93 ±191% perf-profile.calltrace.cycles-pp.folio_copy.migrate_folio_extra.move_to_new_folio.migrate_pages_batch.migrate_pages 8.48 ± 69% -7.6 0.92 ±191% perf-profile.calltrace.cycles-pp.copy_page.folio_copy.migrate_folio_extra.move_to_new_folio.migrate_pages_batch 4.67 ±111% -3.8 0.88 ±223% perf-profile.calltrace.cycles-pp.perf_session__process_user_event.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record 4.66 ±111% -3.8 0.88 ±223% perf-profile.calltrace.cycles-pp.__ordered_events__flush.perf_session__process_user_event.reader__read_event.perf_session__process_events.record__finish_output 3.99 ±113% -3.3 0.71 ±223% perf-profile.calltrace.cycles-pp.perf_session__deliver_event.__ordered_events__flush.perf_session__process_user_event.reader__read_event.perf_session__process_events 2.20 ± 53% -2.2 0.00 perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copy_page_from_iter_atomic.generic_perform_write.__generic_file_write_iter.generic_file_write_iter 2.72 ±109% -2.2 0.55 ±223% perf-profile.calltrace.cycles-pp.evlist__parse_sample.perf_session__deliver_event.__ordered_events__flush.perf_session__process_user_event.reader__read_event 1.40 ± 28% -1.4 0.00 perf-profile.calltrace.cycles-pp.drm_fbdev_fb_dirty.drm_fb_helper_damage_work.process_one_work.worker_thread.kthread 1.00 ±131% -0.9 0.14 ±223% perf-profile.calltrace.cycles-pp.build_id__mark_dso_hit.perf_session__deliver_event.__ordered_events__flush.perf_session__process_user_event.reader__read_event 0.90 ±108% -0.8 0.12 ±223% perf-profile.calltrace.cycles-pp.evsel__parse_sample.evlist__parse_sample.perf_session__deliver_event.__ordered_events__flush.perf_session__process_user_event 0.85 ± 53% -0.6 0.26 ±142% perf-profile.calltrace.cycles-pp.__x64_sys_poll.do_syscall_64.entry_SYSCALL_64_after_hwframe.__poll.__cmd_record 0.85 ± 53% -0.6 0.26 ±142% perf-profile.calltrace.cycles-pp.do_sys_poll.__x64_sys_poll.do_syscall_64.entry_SYSCALL_64_after_hwframe.__poll 0.86 ± 52% -0.6 0.28 ±142% perf-profile.calltrace.cycles-pp.__poll.__cmd_record.cmd_record.run_builtin.main 0.85 ± 53% -0.6 0.27 ±142% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__poll.__cmd_record.cmd_record.run_builtin 0.85 ± 53% -0.6 0.27 ±142% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__poll.__cmd_record.cmd_record 7.65 ± 55% -0.6 7.09 ± 57% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.main.__libc_start_main 7.65 ± 55% -0.6 7.09 ± 57% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.main.__libc_start_main 0.53 ± 49% -0.5 0.00 perf-profile.calltrace.cycles-pp.__fxstat64 0.52 ± 49% -0.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__fxstat64 0.51 ± 49% -0.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__fxstat64 1.90 ± 23% -0.5 1.41 ± 13% perf-profile.calltrace.cycles-pp.process_one_work.worker_thread.kthread.ret_from_fork 0.66 ± 52% -0.5 0.21 ±144% perf-profile.calltrace.cycles-pp.do_poll.do_sys_poll.__x64_sys_poll.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.44 ±223% -0.4 0.00 perf-profile.calltrace.cycles-pp.thread__find_map.build_id__mark_dso_hit.perf_session__deliver_event.__ordered_events__flush.perf_session__process_user_event 0.43 ± 76% -0.4 0.00 perf-profile.calltrace.cycles-pp.drm_atomic_helper_dirtyfb.drm_fbdev_fb_dirty.drm_fb_helper_damage_work.process_one_work.worker_thread 0.43 ±223% -0.4 0.00 perf-profile.calltrace.cycles-pp.map__load.thread__find_map.build_id__mark_dso_hit.perf_session__deliver_event.__ordered_events__flush 0.43 ±223% -0.4 0.00 perf-profile.calltrace.cycles-pp.dso__load.map__load.thread__find_map.build_id__mark_dso_hit.perf_session__deliver_event 1.50 ± 54% -0.4 1.07 ±103% perf-profile.calltrace.cycles-pp.cmd_stat 1.50 ± 54% -0.4 1.07 ±103% perf-profile.calltrace.cycles-pp.dispatch_events.cmd_stat 1.49 ± 55% -0.4 1.06 ±103% perf-profile.calltrace.cycles-pp.process_interval.dispatch_events.cmd_stat 0.43 ±223% -0.4 0.00 perf-profile.calltrace.cycles-pp.__dso__load_kallsyms.dso__load.map__load.thread__find_map.build_id__mark_dso_hit 5.41 ± 57% -0.4 5.00 ± 58% perf-profile.calltrace.cycles-pp.__generic_file_write_iter.generic_file_write_iter.vfs_write.ksys_write.do_syscall_64 0.41 ± 76% -0.4 0.00 perf-profile.calltrace.cycles-pp.drm_atomic_commit.drm_atomic_helper_dirtyfb.drm_fbdev_fb_dirty.drm_fb_helper_damage_work.process_one_work 5.37 ± 56% -0.4 4.96 ± 58% perf-profile.calltrace.cycles-pp.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.vfs_write.ksys_write 5.44 ± 56% -0.4 5.04 ± 58% perf-profile.calltrace.cycles-pp.generic_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe 5.51 ± 57% -0.4 5.11 ± 58% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_write 1.41 ± 55% -0.4 1.01 ±104% perf-profile.calltrace.cycles-pp.read_counters.process_interval.dispatch_events.cmd_stat 5.56 ± 57% -0.4 5.16 ± 58% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_write.writen 5.57 ± 57% -0.4 5.20 ± 58% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_write.writen.record__pushfn 0.56 ±165% -0.4 0.19 ±223% perf-profile.calltrace.cycles-pp.evlist__parse_sample_timestamp.process_simple.reader__read_event.perf_session__process_events.record__finish_output 5.58 ± 57% -0.4 5.20 ± 58% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_write.writen.record__pushfn.perf_mmap__push 5.60 ± 57% -0.4 5.24 ± 58% perf-profile.calltrace.cycles-pp.__libc_write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist 1.94 ± 23% -0.4 1.58 ± 21% perf-profile.calltrace.cycles-pp.worker_thread.kthread.ret_from_fork 0.72 ± 57% -0.3 0.41 ± 71% perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 1.40 ± 28% -0.3 1.08 ± 15% perf-profile.calltrace.cycles-pp.drm_fb_helper_damage_work.process_one_work.worker_thread.kthread.ret_from_fork 0.31 ±105% -0.3 0.00 perf-profile.calltrace.cycles-pp.drm_atomic_helper_commit.drm_atomic_commit.drm_atomic_helper_dirtyfb.drm_fbdev_fb_dirty.drm_fb_helper_damage_work 0.31 ±103% -0.3 0.00 perf-profile.calltrace.cycles-pp.__do_sys_newfstat.do_syscall_64.entry_SYSCALL_64_after_hwframe.__fxstat64 0.30 ±103% -0.3 0.00 perf-profile.calltrace.cycles-pp.irq_work_run_list.irq_work_run.__sysvec_irq_work.sysvec_irq_work.asm_sysvec_irq_work 0.30 ±103% -0.3 0.00 perf-profile.calltrace.cycles-pp.irq_work_single.irq_work_run_list.irq_work_run.__sysvec_irq_work.sysvec_irq_work 0.30 ±103% -0.3 0.00 perf-profile.calltrace.cycles-pp._printk.irq_work_single.irq_work_run_list.irq_work_run.__sysvec_irq_work 0.30 ±103% -0.3 0.00 perf-profile.calltrace.cycles-pp.vprintk_emit._printk.irq_work_single.irq_work_run_list.irq_work_run 0.30 ±103% -0.3 0.00 perf-profile.calltrace.cycles-pp.console_unlock.vprintk_emit._printk.irq_work_single.irq_work_run_list 0.30 ±103% -0.3 0.00 perf-profile.calltrace.cycles-pp.console_flush_all.console_unlock.vprintk_emit._printk.irq_work_single 0.30 ±103% -0.3 0.00 perf-profile.calltrace.cycles-pp.serial8250_console_write.console_flush_all.console_unlock.vprintk_emit._printk 0.40 ±154% -0.3 0.13 ±223% perf-profile.calltrace.cycles-pp.machine__findnew_thread.build_id__mark_dso_hit.perf_session__deliver_event.__ordered_events__flush.perf_session__process_user_event 0.28 ±100% -0.3 0.00 perf-profile.calltrace.cycles-pp.sched_setaffinity.__x64_sys_sched_setaffinity.do_syscall_64.entry_SYSCALL_64_after_hwframe.sched_setaffinity 0.27 ±100% -0.3 0.00 perf-profile.calltrace.cycles-pp.cpu_stopper_thread.smpboot_thread_fn.kthread.ret_from_fork 1.03 ± 68% -0.3 0.77 ± 78% perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_write_begin.generic_perform_write.__generic_file_write_iter.generic_file_write_iter 0.26 ±100% -0.3 0.00 perf-profile.calltrace.cycles-pp.__sched_setaffinity.sched_setaffinity.__x64_sys_sched_setaffinity.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.26 ±100% -0.3 0.00 perf-profile.calltrace.cycles-pp.migration_cpu_stop.cpu_stopper_thread.smpboot_thread_fn.kthread.ret_from_fork 1.05 ± 69% -0.3 0.79 ± 78% perf-profile.calltrace.cycles-pp.shmem_write_begin.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.vfs_write 5.76 ± 58% -0.3 5.50 ± 61% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record 0.35 ±103% -0.3 0.10 ±223% perf-profile.calltrace.cycles-pp.perf_poll.do_poll.do_sys_poll.__x64_sys_poll.do_syscall_64 0.24 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.kallsyms__parse.__dso__load_kallsyms.dso__load.map__load.thread__find_map 0.24 ±150% -0.2 0.00 perf-profile.calltrace.cycles-pp.evsel__parse_sample_timestamp.process_simple.reader__read_event.perf_session__process_events.record__finish_output 0.82 ± 31% -0.2 0.58 ± 45% perf-profile.calltrace.cycles-pp.link_path_walk.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat 0.44 ± 75% -0.2 0.20 ±141% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm 5.88 ± 57% -0.2 5.65 ± 61% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record 0.44 ± 75% -0.2 0.20 ±141% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput 2.83 ± 29% -0.2 2.60 ± 21% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt 0.23 ±142% -0.2 0.00 perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 0.23 ±142% -0.2 0.00 perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 0.32 ±100% -0.2 0.09 ±223% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.sched_setaffinity.evlist_cpu_iterator__next.read_counters.process_interval 0.32 ±100% -0.2 0.09 ±223% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.sched_setaffinity.evlist_cpu_iterator__next.read_counters 0.31 ±100% -0.2 0.09 ±223% perf-profile.calltrace.cycles-pp.__x64_sys_sched_setaffinity.do_syscall_64.entry_SYSCALL_64_after_hwframe.sched_setaffinity.evlist_cpu_iterator__next 0.22 ±142% -0.2 0.00 perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_read 0.22 ±144% -0.2 0.00 perf-profile.calltrace.cycles-pp.asm_sysvec_call_function 0.22 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.perf_session__delete.__cmd_record 3.04 ± 37% -0.2 2.84 ± 21% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64 0.20 ±141% -0.2 0.00 perf-profile.calltrace.cycles-pp.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.19 ±141% -0.2 0.00 perf-profile.calltrace.cycles-pp.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64 0.52 ± 80% -0.2 0.32 ±103% perf-profile.calltrace.cycles-pp.__percpu_counter_init.mm_init.alloc_bprm.do_execveat_common.__x64_sys_execve 0.19 ±141% -0.2 0.00 perf-profile.calltrace.cycles-pp.drm_gem_vmap_unlocked.drm_client_buffer_vmap.drm_fbdev_fb_dirty.drm_fb_helper_damage_work.process_one_work 0.19 ±141% -0.2 0.00 perf-profile.calltrace.cycles-pp.drm_gem_vmap.drm_gem_vmap_unlocked.drm_client_buffer_vmap.drm_fbdev_fb_dirty.drm_fb_helper_damage_work 0.19 ±141% -0.2 0.00 perf-profile.calltrace.cycles-pp.drm_gem_shmem_vmap.drm_gem_vmap.drm_gem_vmap_unlocked.drm_client_buffer_vmap.drm_fbdev_fb_dirty 0.19 ±141% -0.2 0.00 perf-profile.calltrace.cycles-pp.drm_client_buffer_vmap.drm_fbdev_fb_dirty.drm_fb_helper_damage_work.process_one_work.worker_thread 0.19 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.__open64 0.55 ± 47% -0.2 0.36 ± 70% perf-profile.calltrace.cycles-pp.sched_setaffinity.evlist_cpu_iterator__next.read_counters.process_interval.dispatch_events 0.19 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__open64 0.19 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__open64 2.24 ± 53% -0.2 2.05 ± 57% perf-profile.calltrace.cycles-pp.copy_page_from_iter_atomic.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.vfs_write 0.18 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.__open64 0.18 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.__open64 0.87 ± 57% -0.2 0.70 ± 18% perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 0.26 ±145% -0.2 0.09 ±223% perf-profile.calltrace.cycles-pp.__do_softirq.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork 0.26 ±145% -0.2 0.09 ±223% perf-profile.calltrace.cycles-pp.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork 0.25 ±145% -0.2 0.08 ±223% perf-profile.calltrace.cycles-pp.rcu_do_batch.rcu_core.__do_softirq.run_ksoftirqd.smpboot_thread_fn 0.25 ±145% -0.2 0.08 ±223% perf-profile.calltrace.cycles-pp.rcu_core.__do_softirq.run_ksoftirqd.smpboot_thread_fn.kthread 0.16 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.cmd_record 0.16 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.evlist__delete.cmd_record 0.16 ±223% -0.2 0.00 perf-profile.calltrace.cycles-pp.migrate_folio_unmap.migrate_pages_batch.migrate_pages.migrate_misplaced_page.do_huge_pmd_numa_page 1.21 ± 30% -0.2 1.05 ± 19% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit 0.35 ±103% -0.2 0.19 ±141% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap 1.22 ± 30% -0.2 1.06 ± 20% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group 0.97 ± 69% -0.1 0.82 ± 77% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.fault_in_readable.fault_in_iov_iter_readable.generic_perform_write.__generic_file_write_iter 1.24 ± 30% -0.1 1.09 ± 20% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 0.14 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.io__get_hex.kallsyms__parse.__dso__load_kallsyms.dso__load.map__load 0.14 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__ordered_events__flush.perf_session__process_events.record__finish_output.__cmd_record 0.23 ±146% -0.1 0.08 ±223% perf-profile.calltrace.cycles-pp.perf_evsel__read.read_counters.process_interval.dispatch_events.cmd_stat 0.14 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.select_task_rq_fair.wake_up_new_task.kernel_clone.__x64_sys_vfork.do_syscall_64 0.14 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.find_idlest_cpu.select_task_rq_fair.wake_up_new_task.kernel_clone.__x64_sys_vfork 0.31 ±104% -0.1 0.17 ±141% perf-profile.calltrace.cycles-pp.__close_nocancel 6.66 ± 58% -0.1 6.52 ± 62% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.main 0.14 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.rmap_walk_anon.try_to_migrate.migrate_folio_unmap.migrate_pages_batch.migrate_pages 0.14 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.try_to_migrate.migrate_folio_unmap.migrate_pages_batch.migrate_pages.migrate_misplaced_page 0.14 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.try_to_migrate_one.rmap_walk_anon.try_to_migrate.migrate_folio_unmap.migrate_pages_batch 0.54 ± 57% -0.1 0.41 ± 72% perf-profile.calltrace.cycles-pp.vmstat_start.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read 6.37 ± 58% -0.1 6.24 ± 62% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin 0.22 ±145% -0.1 0.09 ±223% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__close_nocancel 0.13 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.perf_session__do_write_header.__cmd_record 0.22 ±145% -0.1 0.09 ±223% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__close_nocancel 0.13 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.find_idlest_group.find_idlest_cpu.select_task_rq_fair.wake_up_new_task.kernel_clone 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.update_sg_wakeup_stats.find_idlest_group.find_idlest_cpu.select_task_rq_fair.wake_up_new_task 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.asm_sysvec_irq_work 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.sysvec_irq_work.asm_sysvec_irq_work 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__sysvec_irq_work.sysvec_irq_work.asm_sysvec_irq_work 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.irq_work_run.__sysvec_irq_work.sysvec_irq_work.asm_sysvec_irq_work 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_read.io__get_hex.kallsyms__parse.__dso__load_kallsyms 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_read.io__get_hex.kallsyms__parse 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__libc_read.io__get_hex.kallsyms__parse.__dso__load_kallsyms.dso__load 0.82 ± 70% -0.1 0.70 ± 79% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.fault_in_readable.fault_in_iov_iter_readable 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_read.io__get_hex 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.schedule.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone 0.39 ±103% -0.1 0.27 ±100% perf-profile.calltrace.cycles-pp.walk_component.link_path_walk.path_openat.do_filp_open.do_sys_openat2 1.30 ± 33% -0.1 1.18 ± 9% perf-profile.calltrace.cycles-pp.smpboot_thread_fn.kthread.ret_from_fork 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.machine__exit.perf_session__delete.__cmd_record 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__schedule.schedule.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.readn.perf_evsel__read.read_counters.process_interval.dispatch_events 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__libc_read.readn.perf_evsel__read.read_counters.process_interval 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.wait_for_lsr.serial8250_console_write.console_flush_all.console_unlock.vprintk_emit 3.05 ± 37% -0.1 2.93 ± 16% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.seq_read_iter.seq_read.proc_reg_read.vfs_read.ksys_read 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.seq_read.proc_reg_read.vfs_read.ksys_read.do_syscall_64 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.proc_reg_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_read.readn.perf_evsel__read.read_counters 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_read.readn.perf_evsel__read 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_read.readn 0.38 ±103% -0.1 0.27 ±100% perf-profile.calltrace.cycles-pp.lookup_fast.walk_component.link_path_walk.path_openat.do_filp_open 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.dso__delete.machine__exit.perf_session__delete.__cmd_record 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.gather_pte_stats.walk_pmd_range.walk_pud_range.walk_p4d_range.walk_pgd_range 0.12 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.evlist__close.evlist__delete.cmd_record 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__schedule.schedule.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.schedule.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.finish_task_switch.__schedule.schedule.exit_to_user_mode_loop.exit_to_user_mode_prepare 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.commit_tail.drm_atomic_helper_commit.drm_atomic_commit.drm_atomic_helper_dirtyfb.drm_fbdev_fb_dirty 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.ast_mode_config_helper_atomic_commit_tail.commit_tail.drm_atomic_helper_commit.drm_atomic_commit.drm_atomic_helper_dirtyfb 0.55 ± 53% -0.1 0.44 ± 75% perf-profile.calltrace.cycles-pp.seq_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.machine__process_fork_event.perf_session__deliver_event.__ordered_events__flush.perf_session__process_user_event.reader__read_event 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.perf_session__deliver_event.__ordered_events__flush.perf_session__process_events.record__finish_output.__cmd_record 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.set_pmd_migration_entry.try_to_migrate_one.rmap_walk_anon.try_to_migrate.migrate_folio_unmap 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.rb_erase.symbols__delete.dso__delete.machine__exit.perf_session__delete 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.symbols__delete.dso__delete.machine__exit.perf_session__delete.__cmd_record 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.next_uptodate_page.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.pick_next_task_fair.__schedule.schedule.io_schedule.migration_entry_wait_on_locked 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.proc_fill_cache.proc_pid_readdir.iterate_dir.__x64_sys_getdents64.do_syscall_64 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.setlocale 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.perf_evsel__close_cpu.evlist__close.evlist__delete.cmd_record 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.newidle_balance.pick_next_task_fair.__schedule.schedule.io_schedule 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.drm_fbdev_damage_blit_real.drm_fbdev_fb_dirty.drm_fb_helper_damage_work.process_one_work.worker_thread 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.pid_revalidate.lookup_fast.walk_component.link_path_walk.path_openat 0.20 ±144% -0.1 0.10 ±223% perf-profile.calltrace.cycles-pp.__collapse_huge_page_copy.collapse_huge_page.hpage_collapse_scan_pmd.khugepaged_scan_mm_slot.khugepaged 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__mmdrop.finish_task_switch.__schedule.schedule.exit_to_user_mode_loop 0.11 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.memcpy_erms.drm_fbdev_damage_blit_real.drm_fbdev_fb_dirty.drm_fb_helper_damage_work.process_one_work 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__close.perf_evsel__close_cpu.evlist__close.evlist__delete 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__close.perf_evsel__close_cpu.evlist__close 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__close.perf_evsel__close_cpu.evlist__close.evlist__delete.cmd_record 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.proc_pid_cmdline_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.flush_tlb_mm_range.pmdp_invalidate.set_pmd_migration_entry.try_to_migrate_one.rmap_walk_anon 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.pmdp_invalidate.set_pmd_migration_entry.try_to_migrate_one.rmap_walk_anon.try_to_migrate 0.84 ± 70% -0.1 0.74 ± 78% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.fault_in_readable.fault_in_iov_iter_readable.generic_perform_write 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.write_cache.perf_session__do_write_header.__cmd_record 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.get_mm_cmdline.proc_pid_cmdline_read.vfs_read.ksys_read.do_syscall_64 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.dispatch_events 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.process_interval.dispatch_events 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.pmdp_invalidate.set_pmd_migration_entry 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.on_each_cpu_cond_mask.flush_tlb_mm_range.pmdp_invalidate.set_pmd_migration_entry.try_to_migrate_one 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.vfs_fstat.__do_sys_newfstat.do_syscall_64.entry_SYSCALL_64_after_hwframe.__fxstat64 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.all_vm_events.vmstat_start.seq_read_iter.proc_reg_read_iter.vfs_read 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.machine__delete_threads.perf_session__delete.__cmd_record 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.sysvec_call_function.asm_sysvec_call_function 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.__close 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.__close.perf_evsel__close_cpu 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__fput.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.__close_nocancel 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.__close_nocancel 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.drm_atomic_helper_commit_tail_rpm.ast_mode_config_helper_atomic_commit_tail.commit_tail.drm_atomic_helper_commit.drm_atomic_commit 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.drm_atomic_helper_commit_planes.drm_atomic_helper_commit_tail_rpm.ast_mode_config_helper_atomic_commit_tail.commit_tail.drm_atomic_helper_commit 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.ast_primary_plane_helper_atomic_update.drm_atomic_helper_commit_planes.drm_atomic_helper_commit_tail_rpm.ast_mode_config_helper_atomic_commit_tail.commit_tail 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.drm_fb_memcpy.ast_primary_plane_helper_atomic_update.drm_atomic_helper_commit_planes.drm_atomic_helper_commit_tail_rpm.ast_mode_config_helper_atomic_commit_tail 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.memcpy_toio.drm_fb_memcpy.ast_primary_plane_helper_atomic_update.drm_atomic_helper_commit_planes.drm_atomic_helper_commit_tail_rpm 0.48 ± 75% -0.1 0.39 ± 72% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.thread__delete.machine__delete_threads.perf_session__delete.__cmd_record 0.10 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.read_counters.process_interval.dispatch_events 0.96 ± 32% -0.1 0.86 ± 21% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64 0.96 ± 32% -0.1 0.86 ± 21% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.wake_up_new_task.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__folio_alloc.vma_alloc_folio.shmem_alloc_folio.shmem_alloc_and_acct_folio.shmem_get_folio_gfp 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp._dl_addr 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.shmem_add_to_page_cache.shmem_get_folio_gfp.shmem_write_begin.generic_perform_write.__generic_file_write_iter 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.____machine__findnew_thread.machine__process_fork_event.perf_session__deliver_event.__ordered_events__flush.perf_session__process_user_event 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.exit_to_user_mode_loop 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.copy_page.__collapse_huge_page_copy.collapse_huge_page.hpage_collapse_scan_pmd.khugepaged_scan_mm_slot 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__alloc_pages.__folio_alloc.vma_alloc_folio.shmem_alloc_folio.shmem_alloc_and_acct_folio 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.select_task_rq_fair.wake_up_new_task.kernel_clone.__do_sys_clone.do_syscall_64 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.find_idlest_cpu.select_task_rq_fair.wake_up_new_task.kernel_clone.__do_sys_clone 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.walk_component.path_lookupat.filename_lookup.vfs_statx.vfs_fstatat 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__x64_sys_mprotect.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.rb_next.symbols__fixup_end.__dso__load_kallsyms.dso__load.map__load 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.rb_erase.__dso__load_kallsyms.dso__load.map__load.thread__find_map 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.symbols__fixup_end.__dso__load_kallsyms.dso__load.map__load.thread__find_map 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__set_cpus_allowed_ptr.__sched_setaffinity.sched_setaffinity.__x64_sys_sched_setaffinity.do_syscall_64 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__check_object_size.get_mm_cmdline.proc_pid_cmdline_read.vfs_read.ksys_read 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.asm_sysvec_irq_work.__virt_addr_valid.__check_object_size.get_mm_cmdline.proc_pid_cmdline_read 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.sysvec_irq_work.asm_sysvec_irq_work.__virt_addr_valid.__check_object_size.get_mm_cmdline 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__sysvec_irq_work.sysvec_irq_work.asm_sysvec_irq_work.__virt_addr_valid.__check_object_size 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.irq_work_run.__sysvec_irq_work.sysvec_irq_work.asm_sysvec_irq_work.__virt_addr_valid 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__virt_addr_valid.__check_object_size.get_mm_cmdline.proc_pid_cmdline_read.vfs_read 0.09 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.diskstats_show.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.move_queued_task.migration_cpu_stop.cpu_stopper_thread.smpboot_thread_fn.kthread 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.rb_free.rcu_do_batch.rcu_core.__do_softirq.run_ksoftirqd 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.asm_sysvec_irq_work.queue_event.ordered_events__queue.process_simple.reader__read_event 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.sysvec_irq_work.asm_sysvec_irq_work.queue_event.ordered_events__queue.process_simple 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.__sysvec_irq_work.sysvec_irq_work.asm_sysvec_irq_work.queue_event.ordered_events__queue 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.irq_work_run.__sysvec_irq_work.sysvec_irq_work.asm_sysvec_irq_work.queue_event 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.task_dump_owner.pid_revalidate.lookup_fast.walk_component.link_path_walk 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_read_mapping_page_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages_locked.drm_gem_shmem_vmap 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.filemap_get_entry.shmem_get_folio_gfp.shmem_read_mapping_page_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages_locked 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.drm_gem_shmem_get_pages_locked.drm_gem_shmem_vmap.drm_gem_vmap.drm_gem_vmap_unlocked.drm_client_buffer_vmap 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.drm_gem_get_pages.drm_gem_shmem_get_pages_locked.drm_gem_shmem_vmap.drm_gem_vmap.drm_gem_vmap_unlocked 0.08 ±223% -0.1 0.00 perf-profile.calltrace.cycles-pp.shmem_read_mapping_page_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages_locked.drm_gem_shmem_vmap.drm_gem_vmap 1.57 ± 30% -0.1 1.49 ± 12% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe 1.57 ± 30% -0.1 1.49 ± 12% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe 1.57 ± 30% -0.1 1.49 ± 12% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.50 ± 80% -0.1 0.42 ± 74% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 0.28 ±151% -0.1 0.20 ±144% perf-profile.calltrace.cycles-pp.shmem_alloc_and_acct_folio.shmem_get_folio_gfp.shmem_write_begin.generic_perform_write.__generic_file_write_iter 0.70 ± 37% -0.1 0.63 ± 47% perf-profile.calltrace.cycles-pp.sched_exec.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64 0.23 ±149% -0.1 0.18 ±141% perf-profile.calltrace.cycles-pp.wake_up_new_task.kernel_clone.__x64_sys_vfork.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.84 ± 40% -0.1 0.79 ± 18% perf-profile.calltrace.cycles-pp.path_lookupat.filename_lookup.vfs_statx.vfs_fstatat.__do_sys_newstat 0.66 ± 37% -0.1 0.61 ± 47% perf-profile.calltrace.cycles-pp.select_task_rq_fair.sched_exec.bprm_execve.do_execveat_common.__x64_sys_execve 2.88 ± 30% -0.0 2.84 ± 18% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64 2.89 ± 30% -0.0 2.84 ± 18% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64 0.13 ±223% -0.0 0.09 ±223% perf-profile.calltrace.cycles-pp.pcpu_alloc.__percpu_counter_init.mm_init.alloc_bprm.do_execveat_common 0.85 ± 40% -0.0 0.80 ± 18% perf-profile.calltrace.cycles-pp.filename_lookup.vfs_statx.vfs_fstatat.__do_sys_newstat.do_syscall_64 2.98 ± 30% -0.0 2.94 ± 18% perf-profile.calltrace.cycles-pp.open64 0.87 ± 40% -0.0 0.83 ± 17% perf-profile.calltrace.cycles-pp.vfs_statx.vfs_fstatat.__do_sys_newstat.do_syscall_64.entry_SYSCALL_64_after_hwframe 2.91 ± 30% -0.0 2.88 ± 18% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.open64 2.91 ± 30% -0.0 2.87 ± 18% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64 0.40 ± 77% -0.0 0.37 ± 71% perf-profile.calltrace.cycles-pp.show_numa_map.seq_read_iter.seq_read.vfs_read.ksys_read 0.33 ±107% -0.0 0.29 ±100% perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault 0.12 ±223% -0.0 0.09 ±223% perf-profile.calltrace.cycles-pp.step_into.path_lookupat.filename_lookup.vfs_statx.vfs_fstatat 0.13 ±223% -0.0 0.09 ±223% perf-profile.calltrace.cycles-pp.shmem_alloc_folio.shmem_alloc_and_acct_folio.shmem_get_folio_gfp.shmem_write_begin.generic_perform_write 0.98 ± 39% -0.0 0.94 ± 17% perf-profile.calltrace.cycles-pp.__xstat64 0.96 ± 39% -0.0 0.93 ± 17% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__xstat64 0.12 ±223% -0.0 0.09 ±223% perf-profile.calltrace.cycles-pp.pick_link.step_into.path_lookupat.filename_lookup.vfs_statx 0.12 ±223% -0.0 0.08 ±223% perf-profile.calltrace.cycles-pp.walk_p4d_range.walk_pgd_range.__walk_page_range.walk_page_vma.show_numa_map 0.12 ±223% -0.0 0.08 ±223% perf-profile.calltrace.cycles-pp.walk_pud_range.walk_p4d_range.walk_pgd_range.__walk_page_range.walk_page_vma 0.95 ± 40% -0.0 0.92 ± 17% perf-profile.calltrace.cycles-pp.__do_sys_newstat.do_syscall_64.entry_SYSCALL_64_after_hwframe.__xstat64 0.96 ± 39% -0.0 0.92 ± 17% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__xstat64 0.12 ±223% -0.0 0.08 ±223% perf-profile.calltrace.cycles-pp.walk_pmd_range.walk_pud_range.walk_p4d_range.walk_pgd_range.__walk_page_range 0.93 ± 39% -0.0 0.91 ± 17% perf-profile.calltrace.cycles-pp.vfs_fstatat.__do_sys_newstat.do_syscall_64.entry_SYSCALL_64_after_hwframe.__xstat64 0.11 ±223% -0.0 0.09 ±223% perf-profile.calltrace.cycles-pp.vma_alloc_folio.shmem_alloc_folio.shmem_alloc_and_acct_folio.shmem_get_folio_gfp.shmem_write_begin 0.12 ±223% -0.0 0.10 ±223% perf-profile.calltrace.cycles-pp.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 0.39 ±114% -0.0 0.37 ±103% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.fault_in_readable 0.24 ±143% -0.0 0.22 ±141% perf-profile.calltrace.cycles-pp.khugepaged.kthread.ret_from_fork 0.22 ±144% -0.0 0.20 ±141% perf-profile.calltrace.cycles-pp.collapse_huge_page.hpage_collapse_scan_pmd.khugepaged_scan_mm_slot.khugepaged.kthread 1.58 ± 57% -0.0 1.56 ± 59% perf-profile.calltrace.cycles-pp.fault_in_iov_iter_readable.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.vfs_write 0.23 ±144% -0.0 0.21 ±141% perf-profile.calltrace.cycles-pp.hpage_collapse_scan_pmd.khugepaged_scan_mm_slot.khugepaged.kthread.ret_from_fork 0.23 ±144% -0.0 0.21 ±141% perf-profile.calltrace.cycles-pp.khugepaged_scan_mm_slot.khugepaged.kthread.ret_from_fork 0.45 ± 75% +0.0 0.45 ± 71% perf-profile.calltrace.cycles-pp.lookup_fast.open_last_lookups.path_openat.do_filp_open.do_sys_openat2 5.79 ± 27% +0.0 5.80 ± 15% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 0.47 ± 75% +0.0 0.47 ± 71% perf-profile.calltrace.cycles-pp.open_last_lookups.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat 0.64 ± 48% +0.0 0.65 ± 7% perf-profile.calltrace.cycles-pp.evlist_cpu_iterator__next.read_counters.process_interval.dispatch_events.cmd_stat 5.74 ± 27% +0.0 5.75 ± 15% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 0.68 ± 58% +0.0 0.69 ± 47% perf-profile.calltrace.cycles-pp.kernel_clone.__x64_sys_vfork.do_syscall_64.entry_SYSCALL_64_after_hwframe.__vfork 0.68 ± 58% +0.0 0.69 ± 47% perf-profile.calltrace.cycles-pp.__x64_sys_vfork.do_syscall_64.entry_SYSCALL_64_after_hwframe.__vfork 1.45 ± 57% +0.0 1.47 ± 60% perf-profile.calltrace.cycles-pp.fault_in_readable.fault_in_iov_iter_readable.generic_perform_write.__generic_file_write_iter.generic_file_write_iter 0.57 ± 59% +0.0 0.60 ± 47% perf-profile.calltrace.cycles-pp.find_idlest_cpu.select_task_rq_fair.sched_exec.bprm_execve.do_execveat_common 0.50 ± 77% +0.0 0.53 ± 44% perf-profile.calltrace.cycles-pp.iterate_dir.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe.getdents64 0.08 ±223% +0.0 0.12 ±223% perf-profile.calltrace.cycles-pp.schedule.smpboot_thread_fn.kthread.ret_from_fork 0.50 ± 77% +0.0 0.54 ± 44% perf-profile.calltrace.cycles-pp.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe.getdents64 0.09 ±223% +0.0 0.13 ±223% perf-profile.calltrace.cycles-pp.__percpu_counter_sum.__mmdrop.finish_task_switch.__schedule.schedule 0.49 ± 77% +0.0 0.53 ± 44% perf-profile.calltrace.cycles-pp.proc_pid_readdir.iterate_dir.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.50 ± 77% +0.0 0.54 ± 45% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.getdents64 0.50 ± 77% +0.0 0.54 ± 45% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.getdents64 0.50 ± 77% +0.0 0.54 ± 45% perf-profile.calltrace.cycles-pp.getdents64 0.91 ± 34% +0.0 0.95 ± 13% perf-profile.calltrace.cycles-pp.do_open.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat 0.55 ± 58% +0.0 0.60 ± 47% perf-profile.calltrace.cycles-pp.find_idlest_group.find_idlest_cpu.select_task_rq_fair.sched_exec.bprm_execve 3.33 ± 32% +0.0 3.38 ± 13% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe 0.78 ± 39% +0.0 0.83 ± 20% perf-profile.calltrace.cycles-pp.__vfork 0.12 ±223% +0.1 0.17 ±141% perf-profile.calltrace.cycles-pp.walk_page_vma.show_numa_map.seq_read_iter.seq_read.vfs_read 0.12 ±223% +0.1 0.17 ±141% perf-profile.calltrace.cycles-pp.__walk_page_range.walk_page_vma.show_numa_map.seq_read_iter.seq_read 0.12 ±223% +0.1 0.17 ±141% perf-profile.calltrace.cycles-pp.walk_pgd_range.__walk_page_range.walk_page_vma.show_numa_map.seq_read_iter 3.32 ± 32% +0.1 3.38 ± 13% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.12 ±223% +0.1 0.17 ±141% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exec_mmap.begin_new_exec.load_elf_binary 0.12 ±223% +0.1 0.17 ±141% perf-profile.calltrace.cycles-pp.__mmput.exec_mmap.begin_new_exec.load_elf_binary.search_binary_handler 0.13 ±223% +0.1 0.18 ±141% perf-profile.calltrace.cycles-pp.mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64 7.93 ± 55% +0.1 7.99 ± 51% perf-profile.calltrace.cycles-pp.__libc_start_main 7.92 ± 55% +0.1 7.98 ± 51% perf-profile.calltrace.cycles-pp.main.__libc_start_main 7.92 ± 55% +0.1 7.98 ± 51% perf-profile.calltrace.cycles-pp.run_builtin.main.__libc_start_main 3.54 ± 26% +0.1 3.60 ± 15% perf-profile.calltrace.cycles-pp.seq_read_iter.seq_read.vfs_read.ksys_read.do_syscall_64 0.15 ±223% +0.1 0.22 ±146% perf-profile.calltrace.cycles-pp.shmem_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.vfs_write 5.95 ± 27% +0.1 6.02 ± 15% perf-profile.calltrace.cycles-pp.read 3.55 ± 26% +0.1 3.62 ± 15% perf-profile.calltrace.cycles-pp.seq_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.1 0.08 ±223% perf-profile.calltrace.cycles-pp.copy_mc_enhanced_fast_string.__collapse_huge_page_copy.collapse_huge_page.hpage_collapse_scan_pmd.khugepaged_scan_mm_slot 5.87 ± 27% +0.1 5.95 ± 16% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read 0.00 +0.1 0.08 ±223% perf-profile.calltrace.cycles-pp.mm_init.dup_mm.copy_process.kernel_clone.__do_sys_clone 5.85 ± 27% +0.1 5.94 ± 16% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 0.00 +0.1 0.09 ±223% perf-profile.calltrace.cycles-pp.newidle_balance.pick_next_task_fair.__schedule.schedule.pipe_read 0.00 +0.1 0.09 ±223% perf-profile.calltrace.cycles-pp.pick_next_task_fair.__schedule.schedule.pipe_read.vfs_read 0.00 +0.1 0.09 ±223% perf-profile.calltrace.cycles-pp.show_stat.seq_read_iter.vfs_read.ksys_read.do_syscall_64 0.00 +0.1 0.09 ±223% perf-profile.calltrace.cycles-pp.__schedule.schedule.pipe_read.vfs_read.ksys_read 0.00 +0.1 0.09 ±223% perf-profile.calltrace.cycles-pp.schedule.pipe_read.vfs_read.ksys_read.do_syscall_64 0.00 +0.1 0.09 ±223% perf-profile.calltrace.cycles-pp.update_sg_lb_stats.update_sd_lb_stats.find_busiest_group.load_balance.rebalance_domains 0.00 +0.1 0.09 ±223% perf-profile.calltrace.cycles-pp.__memset.pcpu_alloc.__percpu_counter_init.mm_init.alloc_bprm 2.63 ± 26% +0.1 2.72 ± 15% perf-profile.calltrace.cycles-pp.proc_single_show.seq_read_iter.seq_read.vfs_read.ksys_read 2.61 ± 27% +0.1 2.70 ± 15% perf-profile.calltrace.cycles-pp.do_task_stat.proc_single_show.seq_read_iter.seq_read.vfs_read 0.00 +0.1 0.10 ±223% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__x64_sys_vfork.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.1 0.10 ±223% perf-profile.calltrace.cycles-pp.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt 0.00 +0.1 0.10 ±223% perf-profile.calltrace.cycles-pp.pipe_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.68 ± 59% +0.1 0.78 ± 22% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__vfork 0.68 ± 59% +0.1 0.78 ± 22% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__vfork 0.09 ±223% +0.1 0.20 ±141% perf-profile.calltrace.cycles-pp.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 0.00 +0.1 0.11 ±223% perf-profile.calltrace.cycles-pp.update_sd_lb_stats.find_busiest_group.load_balance.rebalance_domains.__do_softirq 0.63 ± 55% +0.1 0.74 ± 12% perf-profile.calltrace.cycles-pp.do_dentry_open.do_open.path_openat.do_filp_open.do_sys_openat2 0.00 +0.1 0.11 ±223% perf-profile.calltrace.cycles-pp.find_busiest_group.load_balance.rebalance_domains.__do_softirq.__irq_exit_rcu 1.12 ± 36% +0.1 1.23 ± 19% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_fork 1.12 ± 36% +0.1 1.23 ± 19% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork 1.12 ± 36% +0.1 1.23 ± 19% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork 1.12 ± 36% +0.1 1.23 ± 19% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork 0.00 +0.1 0.11 ±223% perf-profile.calltrace.cycles-pp.newidle_balance.pick_next_task_fair.__schedule.schedule.worker_thread 0.00 +0.1 0.11 ±223% perf-profile.calltrace.cycles-pp.pick_next_task_fair.__schedule.schedule.worker_thread.kthread 0.00 +0.1 0.12 ±223% perf-profile.calltrace.cycles-pp.__schedule.schedule.smpboot_thread_fn.kthread.ret_from_fork 0.80 ± 36% +0.1 0.92 ± 17% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.1 0.12 ±223% perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt 0.00 +0.1 0.12 ±223% perf-profile.calltrace.cycles-pp.task_mm_cid_work.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode 0.00 +0.1 0.12 ±223% perf-profile.calltrace.cycles-pp.__schedule.schedule.worker_thread.kthread.ret_from_fork 0.00 +0.1 0.12 ±223% perf-profile.calltrace.cycles-pp.schedule.worker_thread.kthread.ret_from_fork 0.00 +0.1 0.13 ±223% perf-profile.calltrace.cycles-pp.__pthread_disable_asynccancel.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist 0.46 ± 80% +0.1 0.59 ± 47% perf-profile.calltrace.cycles-pp.update_sg_wakeup_stats.find_idlest_group.find_idlest_cpu.select_task_rq_fair.sched_exec 0.00 +0.1 0.13 ±223% perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt 0.00 +0.1 0.13 ±223% perf-profile.calltrace.cycles-pp.__mmdrop.finish_task_switch.__schedule.schedule.do_wait 0.00 +0.1 0.14 ±223% perf-profile.calltrace.cycles-pp.finish_task_switch.__schedule.schedule.do_wait.kernel_wait4 0.00 +0.1 0.14 ±223% perf-profile.calltrace.cycles-pp.load_balance.rebalance_domains.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt 0.14 ±223% +0.1 0.28 ±100% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.14 ±223% +0.1 0.28 ±100% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.14 ±223% +0.1 0.28 ±101% perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.1 0.14 ±223% perf-profile.calltrace.cycles-pp.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt 3.77 ± 23% +0.1 3.91 ± 35% perf-profile.calltrace.cycles-pp.ret_from_fork 3.77 ± 23% +0.1 3.91 ± 35% perf-profile.calltrace.cycles-pp.kthread.ret_from_fork 0.00 +0.1 0.15 ±223% perf-profile.calltrace.cycles-pp.alloc_empty_file.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat 0.00 +0.1 0.15 ±223% perf-profile.calltrace.cycles-pp.__alloc_file.alloc_empty_file.path_openat.do_filp_open.do_sys_openat2 0.00 +0.1 0.15 ±223% perf-profile.calltrace.cycles-pp.menu_select.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary 0.00 +0.2 0.15 ±223% perf-profile.calltrace.cycles-pp.__schedule.schedule.do_wait.kernel_wait4.do_syscall_64 0.00 +0.2 0.15 ±223% perf-profile.calltrace.cycles-pp.schedule.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.2 0.16 ±223% perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt 0.14 ±223% +0.2 0.30 ±146% perf-profile.calltrace.cycles-pp.ring_buffer_read_head.perf_mmap__read_head.perf_mmap__push.record__mmap_read_evlist.__cmd_record 0.14 ±223% +0.2 0.30 ±146% perf-profile.calltrace.cycles-pp.perf_mmap__read_head.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record 0.11 ±223% +0.2 0.28 ±223% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 0.00 +0.2 0.17 ±141% perf-profile.calltrace.cycles-pp.drm_atomic_helper_commit.drm_atomic_commit.drm_atomic_helper_dirtyfb.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work 0.00 +0.2 0.17 ±223% perf-profile.calltrace.cycles-pp.rebalance_domains.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 0.00 +0.2 0.17 ±141% perf-profile.calltrace.cycles-pp.drm_atomic_commit.drm_atomic_helper_dirtyfb.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work.process_one_work 0.00 +0.2 0.17 ±223% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init 0.56 ± 61% +0.2 0.74 ± 18% perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64 0.00 +0.2 0.18 ±223% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init 0.00 +0.2 0.19 ±141% perf-profile.calltrace.cycles-pp.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 0.00 +0.2 0.19 ±223% perf-profile.calltrace.cycles-pp.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel.x86_64_start_reservations 0.00 +0.2 0.19 ±223% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel 0.00 +0.2 0.19 ±223% perf-profile.calltrace.cycles-pp.x86_64_start_kernel.secondary_startup_64_no_verify 0.00 +0.2 0.19 ±223% perf-profile.calltrace.cycles-pp.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify 0.00 +0.2 0.19 ±223% perf-profile.calltrace.cycles-pp.start_kernel.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify 0.00 +0.2 0.19 ±223% perf-profile.calltrace.cycles-pp.arch_call_rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify 0.00 +0.2 0.19 ±223% perf-profile.calltrace.cycles-pp.rest_init.arch_call_rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel 0.00 +0.2 0.19 ±141% perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 1.18 ± 36% +0.2 1.37 ± 19% perf-profile.calltrace.cycles-pp.__libc_fork 0.66 ± 28% +0.2 0.90 ± 25% perf-profile.calltrace.cycles-pp.exec_mmap.begin_new_exec.load_elf_binary.search_binary_handler.exec_binprm 0.32 ±107% +0.3 0.58 ± 13% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.32 ±108% +0.3 0.58 ± 14% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.3 0.26 ±100% perf-profile.calltrace.cycles-pp.drm_atomic_helper_dirtyfb.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work.process_one_work.worker_thread 0.00 +0.3 0.27 ±148% perf-profile.calltrace.cycles-pp.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe.wait4 0.74 ± 30% +0.3 1.01 ± 24% perf-profile.calltrace.cycles-pp.begin_new_exec.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve 0.62 ± 59% +0.3 0.91 ± 24% perf-profile.calltrace.cycles-pp.mm_init.alloc_bprm.do_execveat_common.__x64_sys_execve.do_syscall_64 0.55 ± 84% +0.3 0.84 ± 48% perf-profile.calltrace.cycles-pp.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt 0.55 ± 84% +0.3 0.84 ± 48% perf-profile.calltrace.cycles-pp.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt 1.65 ± 33% +0.3 1.95 ± 22% perf-profile.calltrace.cycles-pp.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve 1.65 ± 34% +0.3 1.96 ± 22% perf-profile.calltrace.cycles-pp.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64 1.64 ± 33% +0.3 1.94 ± 22% perf-profile.calltrace.cycles-pp.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common 0.67 ± 59% +0.3 0.97 ± 24% perf-profile.calltrace.cycles-pp.alloc_bprm.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.3 0.30 ±223% perf-profile.calltrace.cycles-pp.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state 0.00 +0.3 0.32 ±223% perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter 0.00 +0.3 0.32 ±151% perf-profile.calltrace.cycles-pp.queue_event 0.00 +0.3 0.33 ±223% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state 0.00 +0.3 0.34 ±223% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter 0.09 ±223% +0.3 0.43 ±110% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.wait4 0.09 ±223% +0.3 0.43 ±109% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.wait4 0.09 ±223% +0.3 0.43 ±110% perf-profile.calltrace.cycles-pp.wait4 2.44 ± 35% +0.3 2.78 ± 21% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.3 0.34 ±223% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle 0.00 +0.4 0.36 ±110% perf-profile.calltrace.cycles-pp.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe.wait4 0.09 ±223% +0.4 0.48 ±223% perf-profile.calltrace.cycles-pp.update_sg_lb_stats.update_sd_lb_stats.find_busiest_group.load_balance.newidle_balance 0.09 ±223% +0.4 0.49 ±223% perf-profile.calltrace.cycles-pp.find_busiest_group.load_balance.newidle_balance.pick_next_task_fair.__schedule 0.09 ±223% +0.4 0.49 ±223% perf-profile.calltrace.cycles-pp.update_sd_lb_stats.find_busiest_group.load_balance.newidle_balance.pick_next_task_fair 0.30 ±146% +0.4 0.75 ± 48% perf-profile.calltrace.cycles-pp.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt 0.00 +0.4 0.45 ±223% perf-profile.calltrace.cycles-pp.newidle_balance.pick_next_task_fair.__schedule.schedule.schedule_timeout 0.00 +0.4 0.45 ±223% perf-profile.calltrace.cycles-pp.pick_next_task_fair.__schedule.schedule.schedule_timeout.rcu_gp_fqs_loop 0.00 +0.5 0.46 ±223% perf-profile.calltrace.cycles-pp.__schedule.schedule.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread 0.00 +0.5 0.46 ±223% perf-profile.calltrace.cycles-pp.schedule.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 0.00 +0.5 0.46 ±223% perf-profile.calltrace.cycles-pp.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread.ret_from_fork 0.00 +0.5 0.48 ±100% perf-profile.calltrace.cycles-pp.evsel__read_counter.read_counters.process_interval.dispatch_events.cmd_stat 0.00 +0.5 0.49 ±223% perf-profile.calltrace.cycles-pp.rcu_gp_fqs_loop.rcu_gp_kthread.kthread.ret_from_fork 0.11 ±223% +0.5 0.64 ±223% perf-profile.calltrace.cycles-pp.load_balance.newidle_balance.pick_next_task_fair.__schedule.schedule 0.00 +0.5 0.54 ±223% perf-profile.calltrace.cycles-pp.rcu_gp_kthread.kthread.ret_from_fork 3.57 ± 36% +0.5 4.11 ± 22% perf-profile.calltrace.cycles-pp.execve 3.55 ± 36% +0.5 4.10 ± 22% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve 3.56 ± 36% +0.6 4.11 ± 22% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve 3.56 ± 36% +0.6 4.11 ± 22% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve 3.56 ± 35% +0.6 4.11 ± 22% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve 0.00 +0.6 0.57 ±223% perf-profile.calltrace.cycles-pp.intel_idle_xstate.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle 0.25 ±223% +0.6 0.84 ±106% perf-profile.calltrace.cycles-pp.read_counters.process_interval.dispatch_events.cmd_stat.run_builtin 0.27 ±223% +0.6 0.89 ±107% perf-profile.calltrace.cycles-pp.process_interval.dispatch_events.cmd_stat.run_builtin.main 0.27 ±223% +0.6 0.90 ±107% perf-profile.calltrace.cycles-pp.cmd_stat.run_builtin.main.__libc_start_main 0.27 ±223% +0.6 0.90 ±107% perf-profile.calltrace.cycles-pp.dispatch_events.cmd_stat.run_builtin.main.__libc_start_main 0.00 +0.6 0.63 ± 14% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.queue_event.ordered_events__queue.process_simple.reader__read_event 0.00 +0.8 0.80 ±223% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call 0.00 +1.0 0.98 ±223% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle 0.00 +1.1 1.08 ± 15% perf-profile.calltrace.cycles-pp.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work.process_one_work.worker_thread.kthread 0.10 ±223% +1.8 1.93 ±209% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary 0.10 ±223% +1.9 2.00 ±210% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry 0.00 +2.0 2.02 ± 57% perf-profile.calltrace.cycles-pp.copyin.copy_page_from_iter_atomic.generic_perform_write.__generic_file_write_iter.generic_file_write_iter 0.11 ±223% +2.1 2.21 ±198% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify 0.15 ±223% +2.2 2.37 ±191% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify 0.15 ±223% +2.2 2.37 ±191% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify 0.15 ±223% +2.2 2.37 ±191% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify 0.16 ±223% +2.5 2.63 ±188% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify 45.48 ± 45% +6.5 51.98 ± 8% perf-profile.calltrace.cycles-pp.__cmd_record 45.08 ± 45% +6.9 51.98 ± 8% perf-profile.calltrace.cycles-pp.record__finish_output.__cmd_record 45.07 ± 45% +6.9 51.98 ± 8% perf-profile.calltrace.cycles-pp.perf_session__process_events.record__finish_output.__cmd_record 44.86 ± 45% +7.1 51.95 ± 8% perf-profile.calltrace.cycles-pp.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record 39.39 ± 47% +11.2 50.61 ± 10% perf-profile.calltrace.cycles-pp.process_simple.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record 38.05 ± 48% +11.6 49.63 ± 11% perf-profile.calltrace.cycles-pp.queue_event.ordered_events__queue.process_simple.reader__read_event.perf_session__process_events 38.22 ± 48% +11.6 49.86 ± 11% perf-profile.calltrace.cycles-pp.ordered_events__queue.process_simple.reader__read_event.perf_session__process_events.record__finish_output 14.24 ± 54% -8.8 5.45 ± 40% perf-profile.children.cycles-pp.asm_exc_page_fault 13.84 ± 55% -8.8 5.07 ± 42% perf-profile.children.cycles-pp.do_user_addr_fault 13.89 ± 55% -8.8 5.14 ± 42% perf-profile.children.cycles-pp.exc_page_fault 13.27 ± 56% -8.7 4.62 ± 47% perf-profile.children.cycles-pp.handle_mm_fault 12.85 ± 57% -8.5 4.32 ± 51% perf-profile.children.cycles-pp.__handle_mm_fault 9.37 ± 68% -8.2 1.21 ±161% perf-profile.children.cycles-pp.do_huge_pmd_numa_page 9.28 ± 68% -8.1 1.18 ±159% perf-profile.children.cycles-pp.migrate_misplaced_page 9.27 ± 68% -8.1 1.17 ±159% perf-profile.children.cycles-pp.migrate_pages 9.26 ± 68% -8.1 1.17 ±159% perf-profile.children.cycles-pp.migrate_pages_batch 9.01 ± 66% -7.8 1.18 ±147% perf-profile.children.cycles-pp.copy_page 8.65 ± 68% -7.6 1.07 ±160% perf-profile.children.cycles-pp.move_to_new_folio 8.65 ± 68% -7.6 1.07 ±160% perf-profile.children.cycles-pp.migrate_folio_extra 8.63 ± 68% -7.6 1.07 ±160% perf-profile.children.cycles-pp.folio_copy 4.81 ±112% -3.9 0.88 ±223% perf-profile.children.cycles-pp.__ordered_events__flush 4.67 ±111% -3.8 0.88 ±223% perf-profile.children.cycles-pp.perf_session__process_user_event 4.11 ±113% -3.4 0.71 ±223% perf-profile.children.cycles-pp.perf_session__deliver_event 2.40 ± 50% -2.4 0.00 perf-profile.children.cycles-pp.copy_user_enhanced_fast_string 2.81 ±110% -2.3 0.55 ±223% perf-profile.children.cycles-pp.evlist__parse_sample 1.40 ± 28% -1.4 0.00 perf-profile.children.cycles-pp.drm_fbdev_fb_dirty 1.02 ±130% -0.9 0.14 ±223% perf-profile.children.cycles-pp.build_id__mark_dso_hit 0.92 ±109% -0.8 0.12 ±223% perf-profile.children.cycles-pp.evsel__parse_sample 30.39 ± 33% -0.8 29.60 ± 15% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 30.33 ± 33% -0.8 29.56 ± 15% perf-profile.children.cycles-pp.do_syscall_64 7.81 ± 54% -0.7 7.09 ± 57% perf-profile.children.cycles-pp.cmd_record 0.55 ± 34% -0.6 0.00 perf-profile.children.cycles-pp.memcpy_erms 0.51 ±191% -0.5 0.01 ±223% perf-profile.children.cycles-pp.thread__find_map 1.90 ± 23% -0.5 1.41 ± 13% perf-profile.children.cycles-pp.process_one_work 0.49 ± 34% -0.5 0.00 perf-profile.children.cycles-pp.drm_gem_shmem_vmap 0.44 ± 43% -0.4 0.00 perf-profile.children.cycles-pp.memset_erms 0.43 ±223% -0.4 0.00 perf-profile.children.cycles-pp.map__load 0.43 ±223% -0.4 0.00 perf-profile.children.cycles-pp.dso__load 0.43 ±223% -0.4 0.00 perf-profile.children.cycles-pp.__dso__load_kallsyms 1.06 ± 28% -0.4 0.64 ± 31% perf-profile.children.cycles-pp.do_sys_poll 1.06 ± 28% -0.4 0.65 ± 31% perf-profile.children.cycles-pp.__x64_sys_poll 5.43 ± 56% -0.4 5.02 ± 58% perf-profile.children.cycles-pp.__generic_file_write_iter 5.38 ± 56% -0.4 4.98 ± 58% perf-profile.children.cycles-pp.generic_perform_write 5.46 ± 56% -0.4 5.05 ± 58% perf-profile.children.cycles-pp.generic_file_write_iter 0.53 ± 49% -0.4 0.13 ± 79% perf-profile.children.cycles-pp.smp_call_function_many_cond 0.47 ± 56% -0.4 0.08 ±138% perf-profile.children.cycles-pp.on_each_cpu_cond_mask 1.07 ± 28% -0.4 0.68 ± 30% perf-profile.children.cycles-pp.__poll 5.73 ± 55% -0.4 5.34 ± 56% perf-profile.children.cycles-pp.vfs_write 5.79 ± 55% -0.4 5.40 ± 56% perf-profile.children.cycles-pp.ksys_write 0.50 ±128% -0.4 0.13 ±223% perf-profile.children.cycles-pp.machine__findnew_thread 5.64 ± 56% -0.4 5.27 ± 58% perf-profile.children.cycles-pp.__libc_write 1.94 ± 23% -0.4 1.58 ± 21% perf-profile.children.cycles-pp.worker_thread 0.50 ± 59% -0.3 0.15 ± 82% perf-profile.children.cycles-pp.flush_tlb_mm_range 0.34 ± 42% -0.3 0.00 perf-profile.children.cycles-pp.drm_fbdev_damage_blit_real 0.39 ± 77% -0.3 0.06 ±223% perf-profile.children.cycles-pp.migrate_folio_unmap 0.41 ± 55% -0.3 0.08 ±106% perf-profile.children.cycles-pp.asm_sysvec_call_function 0.83 ± 27% -0.3 0.52 ± 31% perf-profile.children.cycles-pp.do_poll 1.40 ± 28% -0.3 1.08 ± 15% perf-profile.children.cycles-pp.drm_fb_helper_damage_work 0.36 ± 74% -0.3 0.06 ±189% perf-profile.children.cycles-pp.rmap_walk_anon 0.34 ± 74% -0.3 0.05 ±223% perf-profile.children.cycles-pp.try_to_migrate 0.66 ± 26% -0.3 0.37 ± 11% perf-profile.children.cycles-pp.__fxstat64 1.43 ± 53% -0.3 1.14 ± 41% perf-profile.children.cycles-pp.shmem_get_folio_gfp 0.34 ± 74% -0.3 0.05 ±223% perf-profile.children.cycles-pp.try_to_migrate_one 0.57 ± 24% -0.3 0.30 ± 11% perf-profile.children.cycles-pp.__do_sys_newfstat 1.42 ± 31% -0.3 1.15 ± 12% perf-profile.children.cycles-pp.link_path_walk 0.71 ±119% -0.3 0.45 ± 67% perf-profile.children.cycles-pp.evlist__parse_sample_timestamp 5.78 ± 58% -0.3 5.52 ± 61% perf-profile.children.cycles-pp.writen 0.29 ± 70% -0.3 0.04 ±223% perf-profile.children.cycles-pp.set_pmd_migration_entry 0.24 ±223% -0.2 0.00 perf-profile.children.cycles-pp.kallsyms__parse 0.31 ± 60% -0.2 0.06 ±107% perf-profile.children.cycles-pp.sysvec_call_function 0.28 ± 69% -0.2 0.04 ±223% perf-profile.children.cycles-pp.pmdp_invalidate 0.63 ± 28% -0.2 0.39 ± 36% perf-profile.children.cycles-pp.perf_poll 0.24 ±148% -0.2 0.00 perf-profile.children.cycles-pp.rb_erase 5.88 ± 57% -0.2 5.65 ± 61% perf-profile.children.cycles-pp.record__pushfn 1.71 ± 31% -0.2 1.48 ± 15% perf-profile.children.cycles-pp.lookup_fast 5.27 ± 26% -0.2 5.05 ± 17% perf-profile.children.cycles-pp.seq_read_iter 0.24 ± 93% -0.2 0.02 ±223% perf-profile.children.cycles-pp.migration_entry_wait_on_locked 0.22 ±223% -0.2 0.00 perf-profile.children.cycles-pp.evlist__delete 0.24 ± 94% -0.2 0.02 ±223% perf-profile.children.cycles-pp.io_schedule 0.22 ±223% -0.2 0.00 perf-profile.children.cycles-pp.perf_session__delete 3.71 ± 30% -0.2 3.49 ± 15% perf-profile.children.cycles-pp.do_filp_open 3.69 ± 30% -0.2 3.48 ± 15% perf-profile.children.cycles-pp.path_openat 1.08 ± 63% -0.2 0.86 ± 62% perf-profile.children.cycles-pp.shmem_write_begin 1.18 ± 32% -0.2 0.97 ± 7% perf-profile.children.cycles-pp.sched_setaffinity 1.35 ± 32% -0.2 1.14 ± 18% perf-profile.children.cycles-pp.do_read_fault 3.89 ± 30% -0.2 3.69 ± 15% perf-profile.children.cycles-pp.do_sys_openat2 3.90 ± 31% -0.2 3.70 ± 15% perf-profile.children.cycles-pp.__x64_sys_openat 1.28 ± 32% -0.2 1.08 ± 13% perf-profile.children.cycles-pp.walk_component 1.58 ± 33% -0.2 1.38 ± 18% perf-profile.children.cycles-pp.do_fault 0.54 ± 56% -0.2 0.35 ± 21% perf-profile.children.cycles-pp.__fput 0.19 ±223% -0.2 0.00 perf-profile.children.cycles-pp.__open64 0.64 ± 46% -0.2 0.45 ± 23% perf-profile.children.cycles-pp._raw_spin_lock 1.67 ± 28% -0.2 1.49 ± 19% perf-profile.children.cycles-pp.__mmput 1.66 ± 29% -0.2 1.48 ± 19% perf-profile.children.cycles-pp.exit_mmap 0.24 ± 57% -0.2 0.06 ±106% perf-profile.children.cycles-pp.__sysvec_call_function 0.72 ± 25% -0.2 0.54 ± 12% perf-profile.children.cycles-pp.serial8250_console_write 0.72 ± 26% -0.2 0.54 ± 12% perf-profile.children.cycles-pp.console_unlock 0.72 ± 26% -0.2 0.54 ± 12% perf-profile.children.cycles-pp.console_flush_all 0.44 ± 25% -0.2 0.26 ± 14% perf-profile.children.cycles-pp.vfs_fstat 0.24 ± 54% -0.2 0.06 ±104% perf-profile.children.cycles-pp.__flush_smp_call_function_queue 2.24 ± 53% -0.2 2.06 ± 58% perf-profile.children.cycles-pp.copy_page_from_iter_atomic 6.39 ± 27% -0.2 6.22 ± 15% perf-profile.children.cycles-pp.vfs_read 0.73 ± 25% -0.2 0.56 ± 12% perf-profile.children.cycles-pp.vprintk_emit 0.94 ± 27% -0.2 0.77 ± 20% perf-profile.children.cycles-pp.unmap_vmas 0.72 ± 24% -0.2 0.56 ± 12% perf-profile.children.cycles-pp.irq_work_run_list 0.72 ± 24% -0.2 0.56 ± 12% perf-profile.children.cycles-pp.irq_work_single 0.72 ± 23% -0.2 0.56 ± 12% perf-profile.children.cycles-pp.sysvec_irq_work 0.72 ± 23% -0.2 0.56 ± 12% perf-profile.children.cycles-pp.__sysvec_irq_work 0.72 ± 23% -0.2 0.56 ± 12% perf-profile.children.cycles-pp.irq_work_run 0.72 ± 23% -0.2 0.56 ± 12% perf-profile.children.cycles-pp._printk 0.72 ± 23% -0.2 0.56 ± 12% perf-profile.children.cycles-pp.asm_sysvec_irq_work 0.85 ± 29% -0.2 0.68 ± 21% perf-profile.children.cycles-pp.unmap_page_range 0.70 ± 40% -0.2 0.54 ± 22% perf-profile.children.cycles-pp.__libc_read 0.81 ± 30% -0.2 0.65 ± 20% perf-profile.children.cycles-pp.zap_pte_range 0.82 ± 30% -0.2 0.66 ± 21% perf-profile.children.cycles-pp.zap_pmd_range 0.16 ±223% -0.2 0.00 perf-profile.children.cycles-pp.evlist__close 0.93 ± 22% -0.2 0.78 ± 19% perf-profile.children.cycles-pp.filemap_map_pages 0.18 ± 92% -0.2 0.03 ±223% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath 6.51 ± 27% -0.2 6.36 ± 16% perf-profile.children.cycles-pp.ksys_read 0.16 ± 39% -0.1 0.01 ±223% perf-profile.children.cycles-pp.security_inode_getattr 0.22 ±149% -0.1 0.07 ± 49% perf-profile.children.cycles-pp.__close 0.15 ±223% -0.1 0.00 perf-profile.children.cycles-pp.io__get_hex 1.24 ± 30% -0.1 1.09 ± 20% perf-profile.children.cycles-pp.exit_mm 0.22 ± 66% -0.1 0.08 ± 87% perf-profile.children.cycles-pp.task_numa_work 0.20 ± 79% -0.1 0.06 ± 88% perf-profile.children.cycles-pp.change_prot_numa 0.14 ± 27% -0.1 0.00 perf-profile.children.cycles-pp.__mem_cgroup_flush_stats 0.14 ± 27% -0.1 0.00 perf-profile.children.cycles-pp.cgroup_rstat_flush_irqsafe 0.66 ± 22% -0.1 0.52 ± 23% perf-profile.children.cycles-pp.next_uptodate_page 1.09 ± 35% -0.1 0.95 ± 16% perf-profile.children.cycles-pp.path_lookupat 0.21 ± 48% -0.1 0.07 ± 34% perf-profile.children.cycles-pp.free_unref_page 1.09 ± 35% -0.1 0.95 ± 17% perf-profile.children.cycles-pp.__xstat64 0.50 ± 48% -0.1 0.36 ± 27% perf-profile.children.cycles-pp.tlb_finish_mmu 1.10 ± 35% -0.1 0.96 ± 15% perf-profile.children.cycles-pp.filename_lookup 6.66 ± 58% -0.1 6.52 ± 62% perf-profile.children.cycles-pp.record__mmap_read_evlist 1.11 ± 36% -0.1 0.98 ± 16% perf-profile.children.cycles-pp.__do_sys_newstat 1.01 ± 37% -0.1 0.88 ± 16% perf-profile.children.cycles-pp.vfs_statx 0.13 ±223% -0.1 0.00 perf-profile.children.cycles-pp.perf_session__do_write_header 0.13 ±190% -0.1 0.00 perf-profile.children.cycles-pp.machine__process_fork_event 6.37 ± 58% -0.1 6.24 ± 62% perf-profile.children.cycles-pp.perf_mmap__push 0.88 ± 38% -0.1 0.76 ± 24% perf-profile.children.cycles-pp.__percpu_counter_init 0.80 ± 30% -0.1 0.68 ± 6% perf-profile.children.cycles-pp.evlist_cpu_iterator__next 0.30 ±109% -0.1 0.18 ± 59% perf-profile.children.cycles-pp.evsel__parse_sample_timestamp 0.12 ± 44% -0.1 0.00 perf-profile.children.cycles-pp.common_perm_cond 0.48 ± 44% -0.1 0.36 ± 25% perf-profile.children.cycles-pp.perf_evsel__read 0.12 ± 20% -0.1 0.00 perf-profile.children.cycles-pp.update_rq_clock 0.49 ± 34% -0.1 0.37 ± 29% perf-profile.children.cycles-pp.drm_gem_vmap_unlocked 0.49 ± 34% -0.1 0.37 ± 29% perf-profile.children.cycles-pp.drm_gem_vmap 0.12 ±223% -0.1 0.00 perf-profile.children.cycles-pp.perf_evsel__close_cpu 1.09 ± 35% -0.1 0.97 ± 16% perf-profile.children.cycles-pp.vfs_fstatat 1.30 ± 33% -0.1 1.18 ± 9% perf-profile.children.cycles-pp.smpboot_thread_fn 0.12 ±223% -0.1 0.00 perf-profile.children.cycles-pp.machine__exit 0.12 ±223% -0.1 0.00 perf-profile.children.cycles-pp.proc_reg_read 0.12 ±223% -0.1 0.00 perf-profile.children.cycles-pp.dso__delete 0.58 ± 27% -0.1 0.47 ± 14% perf-profile.children.cycles-pp.drm_atomic_helper_dirtyfb 0.13 ± 40% -0.1 0.01 ±223% perf-profile.children.cycles-pp.llist_add_batch 0.11 ±190% -0.1 0.00 perf-profile.children.cycles-pp.____machine__findnew_thread 0.13 ± 26% -0.1 0.02 ±142% perf-profile.children.cycles-pp.cp_new_stat 0.14 ± 53% -0.1 0.03 ±111% perf-profile.children.cycles-pp.free_unref_page_prepare 0.38 ± 27% -0.1 0.27 ± 12% perf-profile.children.cycles-pp.__fget_light 0.11 ±185% -0.1 0.00 perf-profile.children.cycles-pp.rb_next 0.11 ±223% -0.1 0.00 perf-profile.children.cycles-pp.symbols__delete 0.15 ± 69% -0.1 0.04 ±113% perf-profile.children.cycles-pp.task_numa_fault 0.16 ± 80% -0.1 0.06 ± 56% perf-profile.children.cycles-pp.vmstat_shepherd 0.18 ± 89% -0.1 0.08 ± 46% perf-profile.children.cycles-pp.do_numa_page 0.38 ± 38% -0.1 0.28 ± 33% perf-profile.children.cycles-pp.drm_client_buffer_vmap 0.10 ±223% -0.1 0.00 perf-profile.children.cycles-pp.write_cache 0.77 ± 35% -0.1 0.67 ± 16% perf-profile.children.cycles-pp.kmem_cache_alloc 0.38 ± 43% -0.1 0.28 ± 21% perf-profile.children.cycles-pp.__list_del_entry_valid 0.24 ± 60% -0.1 0.14 ± 18% perf-profile.children.cycles-pp.mutex_lock 0.10 ±223% -0.1 0.00 perf-profile.children.cycles-pp.machine__delete_threads 0.34 ± 40% -0.1 0.24 ± 21% perf-profile.children.cycles-pp.___perf_sw_event 0.22 ± 33% -0.1 0.12 ± 40% perf-profile.children.cycles-pp.poll_freewait 0.57 ± 32% -0.1 0.47 ± 8% perf-profile.children.cycles-pp.__x64_sys_sched_setaffinity 0.55 ± 26% -0.1 0.45 ± 13% perf-profile.children.cycles-pp.drm_atomic_commit 0.10 ±223% -0.1 0.00 perf-profile.children.cycles-pp.rb_free 0.10 ±223% -0.1 0.00 perf-profile.children.cycles-pp.perf_release 0.10 ±223% -0.1 0.00 perf-profile.children.cycles-pp.perf_event_release_kernel 0.22 ± 83% -0.1 0.12 ± 23% perf-profile.children.cycles-pp.get_mm_cmdline 0.12 ± 83% -0.1 0.03 ±155% perf-profile.children.cycles-pp.task_numa_migrate 0.10 ±223% -0.1 0.00 perf-profile.children.cycles-pp.thread__delete 0.96 ± 32% -0.1 0.86 ± 21% perf-profile.children.cycles-pp.proc_reg_read_iter 0.21 ± 27% -0.1 0.12 ± 25% perf-profile.children.cycles-pp.task_tick_fair 0.14 ± 35% -0.1 0.05 ± 47% perf-profile.children.cycles-pp.prepare_task_switch 0.55 ± 26% -0.1 0.46 ± 21% perf-profile.children.cycles-pp.__close_nocancel 0.46 ± 50% -0.1 0.37 ± 27% perf-profile.children.cycles-pp.run_ksoftirqd 0.51 ± 47% -0.1 0.42 ± 18% perf-profile.children.cycles-pp.wait_for_lsr 0.45 ± 41% -0.1 0.36 ± 28% perf-profile.children.cycles-pp.native_irq_return_iret 0.45 ± 74% -0.1 0.36 ± 67% perf-profile.children.cycles-pp.shmem_alloc_and_acct_folio 0.37 ± 24% -0.1 0.28 ± 22% perf-profile.children.cycles-pp._dl_addr 0.53 ± 25% -0.1 0.44 ± 14% perf-profile.children.cycles-pp.drm_atomic_helper_commit 0.20 ± 95% -0.1 0.11 ± 19% perf-profile.children.cycles-pp.__virt_addr_valid 0.09 ±223% -0.1 0.00 perf-profile.children.cycles-pp.symbols__fixup_end 0.50 ± 28% -0.1 0.41 ± 21% perf-profile.children.cycles-pp.fold_vm_numa_events 0.09 ±189% -0.1 0.00 perf-profile.children.cycles-pp.thread__new 0.30 ± 44% -0.1 0.21 ± 26% perf-profile.children.cycles-pp.__perf_sw_event 0.09 ±223% -0.1 0.00 perf-profile.children.cycles-pp._free_event 0.21 ± 72% -0.1 0.12 ± 15% perf-profile.children.cycles-pp.task_dump_owner 0.27 ± 28% -0.1 0.18 ± 15% perf-profile.children.cycles-pp.do_readlinkat 0.27 ± 24% -0.1 0.19 ± 14% perf-profile.children.cycles-pp.sysfs_kf_seq_show 0.27 ± 31% -0.1 0.18 ± 76% perf-profile.children.cycles-pp.io_serial_out 0.78 ± 32% -0.1 0.70 ± 10% perf-profile.children.cycles-pp.io_serial_in 0.08 ±188% -0.1 0.00 perf-profile.children.cycles-pp.machine__process_mmap2_event 0.08 ± 27% -0.1 0.00 perf-profile.children.cycles-pp.__mod_memcg_state 0.27 ± 21% -0.1 0.18 ± 14% perf-profile.children.cycles-pp.dev_attr_show 1.63 ± 31% -0.1 1.55 ± 11% perf-profile.children.cycles-pp.__x64_sys_exit_group 0.44 ± 29% -0.1 0.36 ± 19% perf-profile.children.cycles-pp.cpu_stopper_thread 0.08 ±187% -0.1 0.00 perf-profile.children.cycles-pp.nsinfo__new 0.42 ± 26% -0.1 0.34 ± 23% perf-profile.children.cycles-pp._raw_spin_lock_irqsave 0.24 ± 56% -0.1 0.16 ± 24% perf-profile.children.cycles-pp.__vm_munmap 0.24 ± 76% -0.1 0.16 ± 23% perf-profile.children.cycles-pp.proc_pid_cmdline_read 0.27 ± 28% -0.1 0.19 ± 16% perf-profile.children.cycles-pp.__x64_sys_readlink 1.63 ± 31% -0.1 1.55 ± 11% perf-profile.children.cycles-pp.do_group_exit 0.09 ±106% -0.1 0.02 ±223% perf-profile.children.cycles-pp.delay_halt 0.42 ± 29% -0.1 0.34 ± 14% perf-profile.children.cycles-pp.ast_mode_config_helper_atomic_commit_tail 0.46 ± 41% -0.1 0.39 ± 18% perf-profile.children.cycles-pp.filemap_get_entry 0.14 ± 42% -0.1 0.06 ± 76% perf-profile.children.cycles-pp.flush_tlb_func 0.33 ± 60% -0.1 0.25 ± 15% perf-profile.children.cycles-pp.__check_object_size 1.62 ± 31% -0.1 1.55 ± 12% perf-profile.children.cycles-pp.do_exit 0.42 ± 29% -0.1 0.34 ± 17% perf-profile.children.cycles-pp.migration_cpu_stop 0.52 ± 35% -0.1 0.44 ± 21% perf-profile.children.cycles-pp.do_vmi_align_munmap 0.39 ± 36% -0.1 0.31 ± 29% perf-profile.children.cycles-pp.drm_gem_shmem_get_pages_locked 0.17 ± 33% -0.1 0.10 ± 39% perf-profile.children.cycles-pp.remove_wait_queue 0.13 ± 68% -0.1 0.05 ± 76% perf-profile.children.cycles-pp.free_pcppages_bulk 0.42 ± 29% -0.1 0.35 ± 13% perf-profile.children.cycles-pp.commit_tail 0.67 ± 30% -0.1 0.60 ± 17% perf-profile.children.cycles-pp.pid_revalidate 0.54 ± 34% -0.1 0.46 ± 21% perf-profile.children.cycles-pp.do_vmi_munmap 0.38 ± 36% -0.1 0.31 ± 29% perf-profile.children.cycles-pp.drm_gem_get_pages 0.38 ± 36% -0.1 0.31 ± 29% perf-profile.children.cycles-pp.shmem_read_mapping_page_gfp 0.15 ± 13% -0.1 0.08 ± 24% perf-profile.children.cycles-pp.__lock_task_sighand 0.07 ±223% -0.1 0.00 perf-profile.children.cycles-pp.cpu_cache_level__read 0.20 ± 55% -0.1 0.13 ± 19% perf-profile.children.cycles-pp.unmap_region 0.09 ±155% -0.1 0.02 ±146% perf-profile.children.cycles-pp.__munmap 0.07 ± 83% -0.1 0.00 perf-profile.children.cycles-pp.task_numa_find_cpu 0.24 ± 27% -0.1 0.17 ± 14% perf-profile.children.cycles-pp.readlink 0.07 ±192% -0.1 0.00 perf-profile.children.cycles-pp.map__new 0.27 ± 26% -0.1 0.20 ± 19% perf-profile.children.cycles-pp.update_load_avg 0.07 ± 82% -0.1 0.00 perf-profile.children.cycles-pp.task_numa_compare 0.07 ±223% -0.1 0.00 perf-profile.children.cycles-pp.maps__put 0.78 ± 40% -0.1 0.72 ± 26% perf-profile.children.cycles-pp.wake_up_new_task 0.60 ± 37% -0.1 0.54 ± 23% perf-profile.children.cycles-pp.vmstat_start 0.47 ± 34% -0.1 0.40 ± 11% perf-profile.children.cycles-pp.__sched_setaffinity 0.11 ± 30% -0.1 0.05 ± 75% perf-profile.children.cycles-pp.__get_vm_area_node 0.11 ± 28% -0.1 0.04 ± 75% perf-profile.children.cycles-pp.alloc_vmap_area 0.06 ±223% -0.1 0.00 perf-profile.children.cycles-pp.sw_perf_event_destroy 0.16 ± 85% -0.1 0.09 ± 33% perf-profile.children.cycles-pp.__x64_sys_munmap 0.60 ± 37% -0.1 0.53 ± 17% perf-profile.children.cycles-pp.step_into 0.38 ± 30% -0.1 0.32 ± 17% perf-profile.children.cycles-pp.move_queued_task 0.44 ± 35% -0.1 0.38 ± 13% perf-profile.children.cycles-pp.__set_cpus_allowed_ptr 0.06 ± 79% -0.1 0.00 perf-profile.children.cycles-pp.from_kuid_munged 0.06 ±223% -0.1 0.00 perf-profile.children.cycles-pp.swevent_hlist_put_cpu 0.14 ± 33% -0.1 0.08 ± 64% perf-profile.children.cycles-pp.add_wait_queue 0.08 ± 85% -0.1 0.02 ±142% perf-profile.children.cycles-pp.terminate_walk 0.32 ± 49% -0.1 0.26 ± 57% perf-profile.children.cycles-pp.shmem_add_to_page_cache 0.06 ±223% -0.1 0.00 perf-profile.children.cycles-pp.map__process_kallsym_symbol 0.06 ± 93% -0.1 0.00 perf-profile.children.cycles-pp.llist_reverse_order 0.06 ±223% -0.1 0.00 perf-profile.children.cycles-pp.queue_delayed_work_on 0.16 ± 27% -0.1 0.10 ± 40% perf-profile.children.cycles-pp.drm_gem_vunmap_unlocked 0.07 ± 63% -0.1 0.01 ±223% perf-profile.children.cycles-pp.generic_permission 0.06 ± 75% -0.1 0.00 perf-profile.children.cycles-pp.map_id_up 0.06 ±223% -0.1 0.00 perf-profile.children.cycles-pp.__free_pages 0.16 ± 28% -0.1 0.10 ± 40% perf-profile.children.cycles-pp.drm_gem_shmem_vunmap 0.21 ± 48% -0.1 0.15 ± 24% perf-profile.children.cycles-pp.folio_batch_move_lru 0.16 ± 28% -0.1 0.10 ± 39% perf-profile.children.cycles-pp.drm_gem_vunmap 0.12 ± 55% -0.1 0.06 ± 51% perf-profile.children.cycles-pp.folio_add_lru 0.76 ± 29% -0.1 0.70 ± 18% perf-profile.children.cycles-pp.__d_lookup_rcu 0.22 ± 29% -0.1 0.17 ± 34% perf-profile.children.cycles-pp.__count_memcg_events 0.25 ± 30% -0.1 0.20 ± 13% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state 0.08 ± 43% -0.1 0.02 ±142% perf-profile.children.cycles-pp.folio_unlock 0.22 ± 43% -0.1 0.16 ± 18% perf-profile.children.cycles-pp.vsnprintf 0.35 ± 34% -0.1 0.29 ± 12% perf-profile.children.cycles-pp.drm_atomic_helper_commit_planes 0.34 ± 36% -0.1 0.28 ± 13% perf-profile.children.cycles-pp.memcpy_toio 0.34 ± 36% -0.1 0.28 ± 13% perf-profile.children.cycles-pp.drm_fb_memcpy 0.08 ± 35% -0.1 0.03 ±100% perf-profile.children.cycles-pp.irqtime_account_irq 0.35 ± 34% -0.1 0.30 ± 11% perf-profile.children.cycles-pp.drm_atomic_helper_commit_tail_rpm 0.10 ± 29% -0.1 0.05 ± 47% perf-profile.children.cycles-pp.vmap 0.34 ± 36% -0.1 0.29 ± 13% perf-profile.children.cycles-pp.ast_primary_plane_helper_atomic_update 0.27 ± 36% -0.1 0.22 ± 21% perf-profile.children.cycles-pp.mod_objcg_state 0.28 ± 36% -0.0 0.23 ± 18% perf-profile.children.cycles-pp.affine_move_task 0.32 ± 37% -0.0 0.27 ± 23% perf-profile.children.cycles-pp.tlb_batch_pages_flush 0.18 ± 14% -0.0 0.13 ± 29% perf-profile.children.cycles-pp.single_release 1.39 ± 38% -0.0 1.34 ± 23% perf-profile.children.cycles-pp.find_idlest_cpu 0.28 ± 37% -0.0 0.23 ± 19% perf-profile.children.cycles-pp.release_pages 1.44 ± 38% -0.0 1.40 ± 22% perf-profile.children.cycles-pp.select_task_rq_fair 0.47 ± 34% -0.0 0.42 ± 19% perf-profile.children.cycles-pp.inode_permission 0.30 ± 35% -0.0 0.25 ± 17% perf-profile.children.cycles-pp.__d_lookup 0.05 ±223% -0.0 0.00 perf-profile.children.cycles-pp.sysfs__read_int 2.98 ± 30% -0.0 2.94 ± 18% perf-profile.children.cycles-pp.open64 0.10 ± 34% -0.0 0.05 ± 77% perf-profile.children.cycles-pp.acpi_ps_parse_loop 0.11 ± 57% -0.0 0.07 ± 50% perf-profile.children.cycles-pp.__mod_node_page_state 0.05 ±223% -0.0 0.00 perf-profile.children.cycles-pp.s_next 0.06 ± 86% -0.0 0.02 ±223% perf-profile.children.cycles-pp.change_pmd_range 0.13 ± 78% -0.0 0.08 ± 20% perf-profile.children.cycles-pp.seq_printf 0.07 ± 64% -0.0 0.02 ±152% perf-profile.children.cycles-pp.change_protection_range 0.34 ± 38% -0.0 0.29 ± 22% perf-profile.children.cycles-pp.all_vm_events 3.67 ± 26% -0.0 3.62 ± 15% perf-profile.children.cycles-pp.seq_read 0.56 ± 26% -0.0 0.51 ± 14% perf-profile.children.cycles-pp.show_numa_map 0.38 ± 34% -0.0 0.33 ± 25% perf-profile.children.cycles-pp.setlocale 0.11 ± 38% -0.0 0.07 ± 50% perf-profile.children.cycles-pp.acpi_os_execute_deferred 0.04 ±148% -0.0 0.00 perf-profile.children.cycles-pp.asm_sysvec_call_function_single 0.06 ± 84% -0.0 0.01 ±223% perf-profile.children.cycles-pp.change_pte_range 0.12 ± 39% -0.0 0.07 ± 50% perf-profile.children.cycles-pp.__perf_event_read_value 0.46 ± 26% -0.0 0.41 ± 18% perf-profile.children.cycles-pp.walk_pud_range 0.13 ± 19% -0.0 0.09 ± 21% perf-profile.children.cycles-pp.node_read_vmstat 0.26 ± 28% -0.0 0.22 ± 27% perf-profile.children.cycles-pp.__mod_lruvec_page_state 0.31 ± 33% -0.0 0.27 ± 59% perf-profile.children.cycles-pp.scheduler_tick 0.14 ± 27% -0.0 0.10 ± 32% perf-profile.children.cycles-pp.flush_memcg_stats_dwork 0.14 ± 27% -0.0 0.10 ± 32% perf-profile.children.cycles-pp.cgroup_rstat_flush_locked 0.11 ± 35% -0.0 0.06 ± 53% perf-profile.children.cycles-pp.acpi_ns_evaluate 0.04 ±151% -0.0 0.00 perf-profile.children.cycles-pp.thread__get 0.06 ± 14% -0.0 0.02 ±142% perf-profile.children.cycles-pp.need_update 0.16 ± 26% -0.0 0.12 ± 25% perf-profile.children.cycles-pp.dput 0.20 ± 26% -0.0 0.16 ± 27% perf-profile.children.cycles-pp.lock_vma_under_rcu 0.19 ± 37% -0.0 0.15 ± 17% perf-profile.children.cycles-pp.cpu_stop_queue_work 0.20 ± 33% -0.0 0.15 ± 24% perf-profile.children.cycles-pp.shift_arg_pages 0.46 ± 25% -0.0 0.42 ± 17% perf-profile.children.cycles-pp.walk_page_vma 0.46 ± 25% -0.0 0.42 ± 17% perf-profile.children.cycles-pp.__walk_page_range 0.46 ± 25% -0.0 0.42 ± 17% perf-profile.children.cycles-pp.walk_pgd_range 0.46 ± 26% -0.0 0.42 ± 17% perf-profile.children.cycles-pp.walk_p4d_range 0.46 ± 25% -0.0 0.41 ± 18% perf-profile.children.cycles-pp.walk_pmd_range 0.11 ± 35% -0.0 0.06 ± 52% perf-profile.children.cycles-pp.acpi_ev_asynch_execute_gpe_method 0.06 ± 14% -0.0 0.02 ±142% perf-profile.children.cycles-pp.memchr_inv 0.07 ±101% -0.0 0.03 ±101% perf-profile.children.cycles-pp.__free_one_page 0.10 ± 36% -0.0 0.06 ± 52% perf-profile.children.cycles-pp.acpi_ps_execute_method 0.10 ± 36% -0.0 0.06 ± 52% perf-profile.children.cycles-pp.acpi_ps_parse_aml 0.08 ± 29% -0.0 0.04 ± 75% perf-profile.children.cycles-pp.syscall_return_via_sysret 0.16 ± 35% -0.0 0.12 ± 15% perf-profile.children.cycles-pp.wake_up_q 0.10 ± 43% -0.0 0.06 ± 51% perf-profile.children.cycles-pp.__vmalloc_node_range 0.65 ± 34% -0.0 0.61 ± 11% perf-profile.children.cycles-pp.iterate_dir 0.44 ± 26% -0.0 0.40 ± 17% perf-profile.children.cycles-pp.gather_pte_stats 0.04 ±223% -0.0 0.00 perf-profile.children.cycles-pp.update_iter 0.07 ± 21% -0.0 0.03 ±100% perf-profile.children.cycles-pp.drm_atomic_helper_cleanup_planes 0.07 ± 21% -0.0 0.03 ±100% perf-profile.children.cycles-pp.drm_gem_fb_vunmap 0.04 ± 79% -0.0 0.00 perf-profile.children.cycles-pp._compound_head 0.06 ± 47% -0.0 0.02 ±142% perf-profile.children.cycles-pp.scnprintf 0.04 ±223% -0.0 0.00 perf-profile.children.cycles-pp.s_show 0.09 ± 23% -0.0 0.05 ± 51% perf-profile.children.cycles-pp.vfs_readlink 0.04 ±104% -0.0 0.00 perf-profile.children.cycles-pp.vunmap 0.04 ± 72% -0.0 0.00 perf-profile.children.cycles-pp.complete_walk 0.06 ± 83% -0.0 0.02 ±152% perf-profile.children.cycles-pp.change_p4d_range 0.19 ± 34% -0.0 0.15 ± 24% perf-profile.children.cycles-pp.vma_interval_tree_remove 0.07 ± 16% -0.0 0.03 ±100% perf-profile.children.cycles-pp.proc_pid_readlink 0.04 ±223% -0.0 0.00 perf-profile.children.cycles-pp.io__get_char 0.04 ±112% -0.0 0.00 perf-profile.children.cycles-pp.rw_verify_area 0.05 ± 87% -0.0 0.01 ±223% perf-profile.children.cycles-pp.lru_add_fn 0.07 ± 83% -0.0 0.04 ± 73% perf-profile.children.cycles-pp.security_vm_enough_memory_mm 0.08 ± 26% -0.0 0.04 ±103% perf-profile.children.cycles-pp.check_move_unevictable_pages 0.20 ± 14% -0.0 0.16 ± 19% perf-profile.children.cycles-pp.sync_regs 0.22 ± 36% -0.0 0.18 ± 21% perf-profile.children.cycles-pp.setup_arg_pages 0.10 ± 39% -0.0 0.07 ± 52% perf-profile.children.cycles-pp.free_unref_page_list 0.65 ± 34% -0.0 0.62 ± 11% perf-profile.children.cycles-pp.__x64_sys_getdents64 0.20 ± 27% -0.0 0.17 ± 23% perf-profile.children.cycles-pp.vma_prepare 0.11 ± 22% -0.0 0.08 ± 62% perf-profile.children.cycles-pp.drm_gem_shmem_put_pages_locked 0.11 ± 22% -0.0 0.08 ± 62% perf-profile.children.cycles-pp.drm_gem_put_pages 0.04 ±109% -0.0 0.00 perf-profile.children.cycles-pp.percpu_counter_destroy 0.04 ±152% -0.0 0.00 perf-profile.children.cycles-pp.alloc_misplaced_dst_page 0.12 ± 29% -0.0 0.08 ± 29% perf-profile.children.cycles-pp.mem_cgroup_css_rstat_flush 0.08 ± 24% -0.0 0.04 ± 71% perf-profile.children.cycles-pp.dentry_kill 0.03 ± 70% -0.0 0.00 perf-profile.children.cycles-pp.get_cpu_idle_time_us 0.16 ± 48% -0.0 0.13 ± 48% perf-profile.children.cycles-pp.do_set_pte 0.12 ± 34% -0.0 0.09 ± 28% perf-profile.children.cycles-pp.node_read_numastat 0.65 ± 34% -0.0 0.62 ± 11% perf-profile.children.cycles-pp.getdents64 0.11 ± 44% -0.0 0.08 ± 49% perf-profile.children.cycles-pp.free_pages_and_swap_cache 0.03 ±223% -0.0 0.00 perf-profile.children.cycles-pp.machine__findnew_vdso 0.08 ± 29% -0.0 0.05 ± 71% perf-profile.children.cycles-pp.__dentry_kill 0.09 ± 66% -0.0 0.06 ± 48% perf-profile.children.cycles-pp.perf_exclude_event 0.03 ±155% -0.0 0.00 perf-profile.children.cycles-pp.machines__deliver_event 0.03 ±223% -0.0 0.00 perf-profile.children.cycles-pp.kernfs_fop_open 0.12 ± 58% -0.0 0.08 ± 29% perf-profile.children.cycles-pp.lru_add_drain_cpu 0.09 ± 28% -0.0 0.06 ± 47% perf-profile.children.cycles-pp.fput 0.21 ± 31% -0.0 0.18 ± 22% perf-profile.children.cycles-pp.switch_fpu_return 0.18 ± 34% -0.0 0.15 ± 21% perf-profile.children.cycles-pp.kmalloc_trace 0.07 ± 10% -0.0 0.04 ± 73% perf-profile.children.cycles-pp.set_task_cpu 0.03 ±105% -0.0 0.00 perf-profile.children.cycles-pp.get_vm_area_caller 0.03 ±223% -0.0 0.00 perf-profile.children.cycles-pp.kernfs_dop_revalidate 0.10 ± 23% -0.0 0.07 ± 16% perf-profile.children.cycles-pp.__switch_to 0.79 ± 30% -0.0 0.76 ± 16% perf-profile.children.cycles-pp.open_last_lookups 0.13 ± 83% -0.0 0.10 ± 52% perf-profile.children.cycles-pp.finish_fault 0.10 ± 53% -0.0 0.07 ± 26% perf-profile.children.cycles-pp.lru_add_drain 0.08 ± 26% -0.0 0.05 ± 80% perf-profile.children.cycles-pp.kstat_irqs_usr 0.04 ±156% -0.0 0.01 ±223% perf-profile.children.cycles-pp.page_vma_mapped_walk 0.03 ±158% -0.0 0.00 perf-profile.children.cycles-pp.sysvec_call_function_single 0.07 ± 56% -0.0 0.04 ± 72% perf-profile.children.cycles-pp.get_obj_cgroup_from_current 0.37 ± 30% -0.0 0.34 ± 20% perf-profile.children.cycles-pp.free_pgtables 0.32 ± 29% -0.0 0.29 ± 24% perf-profile.children.cycles-pp.__open64_nocancel 0.04 ±102% -0.0 0.01 ±223% perf-profile.children.cycles-pp.update_numa_stats 0.07 ± 31% -0.0 0.04 ± 75% perf-profile.children.cycles-pp.mas_next_entry 0.63 ± 34% -0.0 0.60 ± 11% perf-profile.children.cycles-pp.proc_pid_readdir 0.38 ± 38% -0.0 0.35 ± 15% perf-profile.children.cycles-pp.proc_fill_cache 0.13 ± 40% -0.0 0.10 ± 23% perf-profile.children.cycles-pp.alloc_thread_stack_node 0.02 ± 99% -0.0 0.00 perf-profile.children.cycles-pp.alloc_fd 0.02 ± 99% -0.0 0.00 perf-profile.children.cycles-pp.perf_swevent_event 0.02 ± 99% -0.0 0.00 perf-profile.children.cycles-pp.task_prio 0.06 ± 50% -0.0 0.03 ±100% perf-profile.children.cycles-pp._copy_to_user 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.proc_ns_get_link 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.ns_get_path 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.kallsyms_expand_symbol 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__ns_get_path 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.sysfs__read_str 0.24 ± 32% -0.0 0.21 ± 13% perf-profile.children.cycles-pp.xas_load 0.17 ± 24% -0.0 0.14 ± 14% perf-profile.children.cycles-pp.user_path_at_empty 0.10 ± 31% -0.0 0.07 ± 50% perf-profile.children.cycles-pp.move_page_tables 0.10 ± 46% -0.0 0.08 ± 48% perf-profile.children.cycles-pp.free_swap_cache 0.12 ± 27% -0.0 0.09 ± 23% perf-profile.children.cycles-pp.__d_alloc 0.40 ± 30% -0.0 0.38 ± 68% perf-profile.children.cycles-pp.update_process_times 0.26 ± 18% -0.0 0.23 ± 21% perf-profile.children.cycles-pp.vma_interval_tree_insert 0.14 ± 58% -0.0 0.11 ± 47% perf-profile.children.cycles-pp.aa_file_perm 0.03 ±108% -0.0 0.01 ±223% perf-profile.children.cycles-pp.acpi_ns_search_and_enter 0.03 ±108% -0.0 0.01 ±223% perf-profile.children.cycles-pp.acpi_ns_lookup 0.03 ±103% -0.0 0.01 ±223% perf-profile.children.cycles-pp.close_fd 0.02 ±143% -0.0 0.00 perf-profile.children.cycles-pp.release_pte_folio 0.13 ± 37% -0.0 0.11 ± 23% perf-profile.children.cycles-pp.percpu_counter_add_batch 0.14 ± 38% -0.0 0.12 ± 25% perf-profile.children.cycles-pp.vm_area_dup 0.12 ± 30% -0.0 0.09 ± 24% perf-profile.children.cycles-pp.seq_open 0.06 ± 28% -0.0 0.04 ± 72% perf-profile.children.cycles-pp.arch_irq_stat_cpu 0.05 ± 80% -0.0 0.03 ±102% perf-profile.children.cycles-pp.vmap_pages_pud_range 0.03 ±157% -0.0 0.01 ±223% perf-profile.children.cycles-pp.tlb_is_not_lazy 0.32 ± 74% -0.0 0.29 ± 64% perf-profile.children.cycles-pp.shmem_alloc_folio 0.30 ± 36% -0.0 0.28 ± 25% perf-profile.children.cycles-pp.get_user_pages_remote 0.20 ± 25% -0.0 0.17 ± 29% perf-profile.children.cycles-pp.blk_mq_in_flight 0.20 ± 25% -0.0 0.17 ± 29% perf-profile.children.cycles-pp.blk_mq_queue_tag_busy_iter 0.38 ± 40% -0.0 0.36 ± 18% perf-profile.children.cycles-pp.dup_mmap 0.19 ± 28% -0.0 0.16 ± 20% perf-profile.children.cycles-pp.restore_fpregs_from_fpstate 0.18 ± 30% -0.0 0.16 ± 22% perf-profile.children.cycles-pp.pid_nr_ns 0.17 ± 54% -0.0 0.15 ± 55% perf-profile.children.cycles-pp.perf_mmap_fault 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__dsos__find 0.04 ±107% -0.0 0.02 ±142% perf-profile.children.cycles-pp.remove_vm_area 0.15 ± 36% -0.0 0.13 ± 23% perf-profile.children.cycles-pp.unlink_file_vma 0.02 ±146% -0.0 0.00 perf-profile.children.cycles-pp.find_unlink_vmap_area 0.02 ±142% -0.0 0.00 perf-profile.children.cycles-pp.acpi_ds_exec_end_op 0.02 ±146% -0.0 0.00 perf-profile.children.cycles-pp.acpi_ps_get_next_namepath 0.05 ± 80% -0.0 0.03 ±101% perf-profile.children.cycles-pp.__vmap_pages_range_noflush 0.02 ±142% -0.0 0.00 perf-profile.children.cycles-pp.mntput_no_expire 0.02 ±142% -0.0 0.00 perf-profile.children.cycles-pp.folio_putback_lru 0.02 ±142% -0.0 0.00 perf-profile.children.cycles-pp.mem_cgroup_from_task 0.50 ± 36% -0.0 0.48 ± 29% perf-profile.children.cycles-pp.khugepaged 0.37 ± 58% -0.0 0.35 ± 28% perf-profile.children.cycles-pp.rmqueue 0.13 ± 40% -0.0 0.11 ± 10% perf-profile.children.cycles-pp.__do_set_cpus_allowed 0.13 ± 26% -0.0 0.11 ± 26% perf-profile.children.cycles-pp.unlink_anon_vmas 0.12 ± 37% -0.0 0.10 ± 45% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.update_iter_mod 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.module_get_kallsym 0.05 ± 77% -0.0 0.03 ±100% perf-profile.children.cycles-pp.handle_pte_fault 0.02 ±144% -0.0 0.00 perf-profile.children.cycles-pp.readdir64 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__sysvec_call_function_single 0.02 ±144% -0.0 0.00 perf-profile.children.cycles-pp.follow_page_pte 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.map__put 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.perf_trace_destroy 0.02 ±144% -0.0 0.00 perf-profile.children.cycles-pp.do_pipe2 0.02 ±144% -0.0 0.00 perf-profile.children.cycles-pp.__x64_sys_pipe 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__x2apic_send_IPI_mask 0.03 ±101% -0.0 0.01 ±223% perf-profile.children.cycles-pp.__vm_enough_memory 0.04 ±223% -0.0 0.02 ±223% perf-profile.children.cycles-pp.delay_halt_tpause 0.12 ± 43% -0.0 0.10 ± 14% perf-profile.children.cycles-pp.__x64_sys_close 0.11 ± 21% -0.0 0.09 ± 27% perf-profile.children.cycles-pp.drm_gem_fb_vmap 0.33 ± 28% -0.0 0.31 ± 20% perf-profile.children.cycles-pp.diskstats_show 0.29 ± 36% -0.0 0.27 ± 25% perf-profile.children.cycles-pp.__get_user_pages 0.46 ± 31% -0.0 0.44 ± 78% perf-profile.children.cycles-pp.tick_sched_timer 0.11 ± 34% -0.0 0.09 ± 26% perf-profile.children.cycles-pp.kmem_cache_alloc_lru 0.19 ± 31% -0.0 0.18 ± 30% perf-profile.children.cycles-pp.copy_strings 0.09 ± 81% -0.0 0.07 ± 81% perf-profile.children.cycles-pp.folio_mark_accessed 0.08 ± 35% -0.0 0.07 ± 14% perf-profile.children.cycles-pp.__switch_to_asm 0.08 ± 33% -0.0 0.06 ± 45% perf-profile.children.cycles-pp.unmap_single_vma 0.03 ±147% -0.0 0.01 ±223% perf-profile.children.cycles-pp.cap_vm_enough_memory 0.02 ±142% -0.0 0.00 perf-profile.children.cycles-pp.hrtimer_active 0.06 ± 52% -0.0 0.04 ± 76% perf-profile.children.cycles-pp.do_notify_parent 0.03 ±100% -0.0 0.01 ±223% perf-profile.children.cycles-pp.mutex_unlock 0.03 ±102% -0.0 0.01 ±223% perf-profile.children.cycles-pp.mas_store 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.write_cpu_topology 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.perf_trace_event_unreg 0.02 ±142% -0.0 0.00 perf-profile.children.cycles-pp.mntput 0.02 ±142% -0.0 0.00 perf-profile.children.cycles-pp.__tlb_remove_page_size 0.10 ± 27% -0.0 0.08 ± 19% perf-profile.children.cycles-pp.select_task_rq 0.42 ± 31% -0.0 0.40 ± 72% perf-profile.children.cycles-pp.tick_sched_handle 0.47 ± 36% -0.0 0.45 ± 29% perf-profile.children.cycles-pp.collapse_huge_page 0.16 ± 38% -0.0 0.14 ± 15% perf-profile.children.cycles-pp.__set_cpus_allowed_ptr_locked 0.18 ± 37% -0.0 0.16 ± 17% perf-profile.children.cycles-pp.__cond_resched 0.02 ±152% -0.0 0.01 ±223% perf-profile.children.cycles-pp.events_stats__inc 0.02 ±141% -0.0 0.01 ±223% perf-profile.children.cycles-pp.security_file_free 0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.__get_vma_policy 0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.folio_add_lru_vma 0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.proc_exe_link 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.trace_event_reg 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.tracepoint_probe_unregister 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.text_poke_bp_batch 0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack 0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp._copy_from_user 0.12 ± 39% -0.0 0.11 ± 18% perf-profile.children.cycles-pp.perf_event_read 0.11 ± 21% -0.0 0.09 ± 26% perf-profile.children.cycles-pp.drm_atomic_helper_prepare_planes 0.06 ± 94% -0.0 0.04 ±105% perf-profile.children.cycles-pp.workingset_activation 0.08 ± 42% -0.0 0.07 ± 13% perf-profile.children.cycles-pp.filp_close 1.58 ± 57% -0.0 1.56 ± 59% perf-profile.children.cycles-pp.fault_in_readable 1.58 ± 57% -0.0 1.57 ± 59% perf-profile.children.cycles-pp.fault_in_iov_iter_readable 0.48 ± 36% -0.0 0.47 ± 29% perf-profile.children.cycles-pp.hpage_collapse_scan_pmd 0.48 ± 36% -0.0 0.47 ± 29% perf-profile.children.cycles-pp.khugepaged_scan_mm_slot 0.29 ± 44% -0.0 0.28 ± 30% perf-profile.children.cycles-pp.__do_fault 0.23 ± 29% -0.0 0.21 ± 17% perf-profile.children.cycles-pp.single_open 0.22 ± 36% -0.0 0.20 ± 21% perf-profile.children.cycles-pp.perf_event_mmap 0.05 ±109% -0.0 0.03 ±101% perf-profile.children.cycles-pp.__do_sys_brk 0.07 ± 85% -0.0 0.05 ± 74% perf-profile.children.cycles-pp.__mod_lruvec_state 0.02 ±152% -0.0 0.01 ±223% perf-profile.children.cycles-pp.alloc_inode 0.04 ±165% -0.0 0.02 ±145% perf-profile.children.cycles-pp.smp_call_function 0.02 ±143% -0.0 0.01 ±223% perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.new_inode_pseudo 0.05 ±111% -0.0 0.03 ±100% perf-profile.children.cycles-pp.cfree 0.07 ± 16% -0.0 0.05 ± 51% perf-profile.children.cycles-pp.___slab_alloc 0.07 ± 33% -0.0 0.05 ± 51% perf-profile.children.cycles-pp.file_ra_state_init 0.03 ±101% -0.0 0.02 ±223% perf-profile.children.cycles-pp.__list_add_valid 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.sched_ttwu_pending 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.kernfs_fop_release 0.02 ±223% -0.0 0.00 perf-profile.children.cycles-pp.proc_single_open 0.47 ± 43% -0.0 0.46 ± 17% perf-profile.children.cycles-pp.pick_link 0.02 ±152% -0.0 0.01 ±223% perf-profile.children.cycles-pp.sched_clock_cpu 0.18 ± 40% -0.0 0.16 ± 20% perf-profile.children.cycles-pp.proc_pid_get_link 0.12 ± 45% -0.0 0.11 ± 18% perf-profile.children.cycles-pp.copy_p4d_range 0.04 ±116% -0.0 0.03 ±103% perf-profile.children.cycles-pp.brk 0.04 ±104% -0.0 0.02 ±146% perf-profile.children.cycles-pp.string 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.symbol__is_idle 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__dsos__findnew_link_by_longname_id 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.rb_insert_color 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.s_start 0.03 ±100% -0.0 0.01 ±223% perf-profile.children.cycles-pp.native_sched_clock 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.machines__for_each_thread 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.rb_first 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.kernfs_iop_permission 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.mmap__munmap 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.perf_mmap__munmap 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.kfree 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.do_huge_pmd_wp_page 0.10 ± 34% -0.0 0.08 ± 18% perf-profile.children.cycles-pp.xas_find 0.21 ± 35% -0.0 0.20 ± 21% perf-profile.children.cycles-pp.perf_event_mmap_event 0.10 ± 61% -0.0 0.08 ± 82% perf-profile.children.cycles-pp.__perf_mmap__read_init 0.13 ± 43% -0.0 0.12 ± 21% perf-profile.children.cycles-pp.copy_page_range 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.path_put 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.maps__clone 0.02 ±144% -0.0 0.01 ±223% perf-profile.children.cycles-pp.format_decode 0.02 ±144% -0.0 0.01 ±223% perf-profile.children.cycles-pp.slab_pre_alloc_hook 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.dsos__findnew_id 0.02 ±144% -0.0 0.01 ±223% perf-profile.children.cycles-pp.strlen 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.get_kthread_comm 0.02 ±146% -0.0 0.01 ±223% perf-profile.children.cycles-pp.meminfo_proc_show 0.02 ±144% -0.0 0.01 ±223% perf-profile.children.cycles-pp.rcu_all_qs 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.ZSTD_freeDCtx 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.comm__free 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.access_error 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.mpol_shared_policy_lookup 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.fclose 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.lockref_put_or_lock 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.sched_move_task 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.xyarray__entry 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.get_user_cpu_mask 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.perf_event_ctx_lock_nested 0.04 ±111% -0.0 0.03 ±101% perf-profile.children.cycles-pp.__perf_session__process_decomp_events 0.05 ± 91% -0.0 0.04 ± 72% perf-profile.children.cycles-pp.anon_vma_clone 0.15 ± 70% -0.0 0.14 ± 80% perf-profile.children.cycles-pp.find_vma 0.12 ± 26% -0.0 0.10 ± 22% perf-profile.children.cycles-pp.__access_remote_vm 0.56 ± 55% -0.0 0.56 ± 22% perf-profile.children.cycles-pp.get_page_from_freelist 0.30 ± 28% -0.0 0.29 ± 17% perf-profile.children.cycles-pp.__kmem_cache_alloc_node 0.15 ± 28% -0.0 0.14 ± 28% perf-profile.children.cycles-pp.mas_walk 0.20 ± 41% -0.0 0.20 ± 33% perf-profile.children.cycles-pp.get_arg_page 0.12 ± 40% -0.0 0.11 ± 27% perf-profile.children.cycles-pp.cgroup_rstat_updated 0.07 ± 48% -0.0 0.06 ± 73% perf-profile.children.cycles-pp.evsel__compute_deltas 0.05 ± 47% -0.0 0.04 ± 71% perf-profile.children.cycles-pp.native_flush_tlb_one_user 0.07 ± 60% -0.0 0.06 ± 55% perf-profile.children.cycles-pp.page_add_file_rmap 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__rblist__findnew 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.number 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.save_fpregs_to_fpstate 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.try_to_unmap_flush 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.inode_init_always 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.arch_tlbbatch_flush 0.02 ±142% -0.0 0.01 ±223% perf-profile.children.cycles-pp.xas_store 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__mutex_lock 0.02 ±144% -0.0 0.01 ±223% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.apparmor_file_free_security 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.uncharge_batch 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.perf_env__arch 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.cpu_topology__new 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.check_move_unevictable_folios 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.getopt_long 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__irqentry_text_end 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__check_heap_object 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.check_preempt_wakeup 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__sysconf 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__vmalloc_area_node 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.create_pipe_files 0.10 ± 39% -0.0 0.09 ± 15% perf-profile.children.cycles-pp._copy_to_iter 0.16 ± 45% -0.0 0.15 ± 74% perf-profile.children.cycles-pp.perf_mmap__read_init 0.10 ± 34% -0.0 0.09 ± 22% perf-profile.children.cycles-pp.pid_task 0.06 ± 54% -0.0 0.06 ± 50% perf-profile.children.cycles-pp.exit_notify 0.03 ±102% -0.0 0.02 ±141% perf-profile.children.cycles-pp.alloc_new_pud 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__comm_str__findnew 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.access 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__static_call_update 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.static_key_disable 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.arch_static_call_transform 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.static_key_disable_cpuslocked 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__static_call_transform 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.arch_jump_label_transform_apply 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.text_poke_bp 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.text_poke_finish 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.perf_trace_sched_migrate_task 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.can_change_pte_writable 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__get_task_ioprio 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.maps__find 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.set_cpus_allowed_common 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.do_write 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__update_load_avg_se 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.perf_tp_event 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.acpi_ex_system_memory_space_handler 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.task_h_load 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.update_cfs_group 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.fdarray__filter 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.apparmor_file_alloc_security 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.page_get_link 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__alloc_pages_bulk 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.generic_file_mmap 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.acpi_ex_field_datum_io 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.acpi_ex_access_region 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.acpi_ev_address_space_dispatch 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.simple_lookup 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.stop_one_cpu 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.migrate_task_rq_fair 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp._IO_file_close_it 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.cpu__max_cpu 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.balance_dirty_pages_ratelimited_flags 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.security_mmap_file 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.__put_anon_vma 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.pcpu_memcg_post_alloc_hook 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.perf_event_comm 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.acpi_ds_evaluate_name_path 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.acpi_ex_resolve_node_to_value 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.acpi_ex_read_data_from_field 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.copy_creds 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.devkmsg_write 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.devkmsg_emit 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.lru_add_drain_per_cpu 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.remove_migration_pte 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.vm_memory_committed 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.wake_affine 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.memmove 0.01 ±223% -0.0 0.00 perf-profile.children.cycles-pp.malloc@plt 0.10 ± 30% -0.0 0.09 ± 25% perf-profile.children.cycles-pp.page_remove_rmap 0.10 ± 44% -0.0 0.09 ± 14% perf-profile.children.cycles-pp.seq_puts 0.04 ±112% -0.0 0.03 ±100% perf-profile.children.cycles-pp.__folio_throttle_swaprate 0.05 ±119% -0.0 0.04 ± 73% perf-profile.children.cycles-pp.anon_vma_fork 0.24 ± 33% -0.0 0.24 ± 25% perf-profile.children.cycles-pp.enqueue_task_fair 0.23 ± 36% -0.0 0.22 ± 17% perf-profile.children.cycles-pp.d_hash_and_lookup 0.20 ± 28% -0.0 0.20 ± 18% perf-profile.children.cycles-pp.vfs_getattr_nosec 0.02 ± 99% -0.0 0.02 ±142% perf-profile.children.cycles-pp.rcu_report_qs_rdp 0.09 ± 34% -0.0 0.09 ± 21% perf-profile.children.cycles-pp.aa_get_task_label 0.09 ± 32% -0.0 0.08 ± 14% perf-profile.children.cycles-pp.__legitimize_path 0.11 ± 41% -0.0 0.11 ± 15% perf-profile.children.cycles-pp.errseq_sample 0.02 ±141% -0.0 0.01 ±223% perf-profile.children.cycles-pp.__evlist__disable 0.02 ±141% -0.0 0.01 ±223% perf-profile.children.cycles-pp.kthread_data 0.02 ±141% -0.0 0.01 ±223% perf-profile.children.cycles-pp.kmem_cache_free_bulk 0.03 ±161% -0.0 0.02 ±149% perf-profile.children.cycles-pp.zero_user_segments 0.36 ± 30% -0.0 0.36 ± 16% perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook 0.20 ± 30% -0.0 0.19 ± 16% perf-profile.children.cycles-pp.generic_fillattr 0.10 ± 39% -0.0 0.09 ± 11% perf-profile.children.cycles-pp.try_to_unlazy 0.06 ± 70% -0.0 0.06 ± 88% perf-profile.children.cycles-pp.perf_mmap_to_page 0.10 ± 51% -0.0 0.09 ± 58% perf-profile.children.cycles-pp.perf_stat_process_counter 0.07 ± 54% -0.0 0.06 ± 53% perf-profile.children.cycles-pp.gather_stats 0.10 ± 37% -0.0 0.10 ± 30% perf-profile.children.cycles-pp.create_elf_tables 0.07 ± 33% -0.0 0.07 ± 47% perf-profile.children.cycles-pp.switch_mm_irqs_off 0.02 ± 99% -0.0 0.02 ±144% perf-profile.children.cycles-pp.prepend_path 0.03 ±103% -0.0 0.03 ±100% perf-profile.children.cycles-pp.mas_push_data 0.03 ±108% -0.0 0.03 ±147% perf-profile.children.cycles-pp.expand_downwards 0.02 ±142% -0.0 0.02 ±141% perf-profile.children.cycles-pp.m_next 0.10 ± 14% -0.0 0.09 ± 17% perf-profile.children.cycles-pp.__kmem_cache_free 0.21 ± 32% -0.0 0.20 ± 18% perf-profile.children.cycles-pp.d_alloc_parallel 0.03 ±103% -0.0 0.03 ±100% perf-profile.children.cycles-pp.remove_vma 0.02 ±144% -0.0 0.02 ±141% perf-profile.children.cycles-pp._find_next_bit 0.05 ± 71% -0.0 0.04 ± 71% perf-profile.children.cycles-pp.fsnotify_perm 0.01 ±223% -0.0 0.01 ±223% perf-profile.children.cycles-pp.perf_event_mmap_output 0.01 ±223% -0.0 0.01 ±223% perf-profile.children.cycles-pp.proc_invalidate_siblings_dcache 0.01 ±223% -0.0 0.01 ±223% perf-profile.children.cycles-pp.ptep_clear_flush 0.03 ±152% -0.0 0.03 ±100% perf-profile.children.cycles-pp.blk_cgroup_congested 0.16 ± 32% -0.0 0.16 ± 19% perf-profile.children.cycles-pp.apparmor_ptrace_access_check 0.14 ± 36% -0.0 0.14 ± 25% perf-profile.children.cycles-pp.__wake_up_common 0.20 ± 32% -0.0 0.20 ± 20% perf-profile.children.cycles-pp.d_alloc 0.20 ± 26% -0.0 0.20 ± 18% perf-profile.children.cycles-pp.dequeue_task_fair 0.21 ± 36% -0.0 0.21 ± 21% perf-profile.children.cycles-pp.may_open 0.10 ± 69% -0.0 0.10 ± 30% perf-profile.children.cycles-pp.rmqueue_bulk 0.13 ± 56% -0.0 0.12 ± 28% perf-profile.children.cycles-pp.pte_alloc_one 0.06 ± 75% -0.0 0.06 ±139% perf-profile.children.cycles-pp.clockevents_program_event 0.01 ±223% -0.0 0.01 ±223% perf-profile.children.cycles-pp.alloc_cpumask_var_node 0.02 ±144% -0.0 0.02 ±142% perf-profile.children.cycles-pp.inode_to_bdi 0.02 ±144% -0.0 0.02 ±142% perf-profile.children.cycles-pp.free_percpu 0.01 ±223% -0.0 0.01 ±223% perf-profile.children.cycles-pp.put_prev_task_stop 0.12 ± 43% -0.0 0.12 ± 27% perf-profile.children.cycles-pp.copy_string_kernel 0.10 ± 52% -0.0 0.10 ± 59% perf-profile.children.cycles-pp.process_counters 0.07 ± 58% -0.0 0.07 ± 17% perf-profile.children.cycles-pp.__anon_vma_prepare 0.03 ±111% -0.0 0.03 ±103% perf-profile.children.cycles-pp.obj_cgroup_uncharge_pages 0.04 ±149% -0.0 0.04 ±106% perf-profile.children.cycles-pp.workingset_age_nonresident 0.10 ± 34% +0.0 0.10 ± 23% perf-profile.children.cycles-pp.__entry_text_start 0.08 ± 49% +0.0 0.08 ± 16% perf-profile.children.cycles-pp.smp_call_function_single 0.08 ± 47% +0.0 0.08 ± 12% perf-profile.children.cycles-pp.generic_exec_single 0.07 ± 67% +0.0 0.07 ± 50% perf-profile.children.cycles-pp.__pte_alloc 0.06 ± 92% +0.0 0.06 ± 17% perf-profile.children.cycles-pp.copy_pte_range 0.04 ± 73% +0.0 0.04 ± 72% perf-profile.children.cycles-pp.__pud_alloc 0.05 ± 74% +0.0 0.05 ± 71% perf-profile.children.cycles-pp.d_path 0.02 ±141% +0.0 0.02 ±141% perf-profile.children.cycles-pp.dup_fd 0.02 ±223% +0.0 0.02 ±141% perf-profile.children.cycles-pp.do_wp_page 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.mas_next_nentry 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.__anon_vma_interval_tree_remove 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.syscall_enter_from_user_mode 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.mas_mab_cp 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.perf_event_task 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.copy_page_to_iter 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.mas_store_b_node 0.04 ±106% +0.0 0.04 ± 73% perf-profile.children.cycles-pp.do_brk_flags 0.26 ± 51% +0.0 0.26 ± 24% perf-profile.children.cycles-pp.perf_read 0.07 ± 47% +0.0 0.07 ± 49% perf-profile.children.cycles-pp.find_extend_vma 0.04 ±107% +0.0 0.04 ± 72% perf-profile.children.cycles-pp.__legitimize_mnt 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.mas_wr_store_entry 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.mas_wmb_replace 0.03 ±101% +0.0 0.03 ±102% perf-profile.children.cycles-pp._IO_setvbuf 0.02 ±145% +0.0 0.03 ±100% perf-profile.children.cycles-pp.lookup_open 0.04 ±104% +0.0 0.04 ± 71% perf-profile.children.cycles-pp.xas_start 0.24 ± 30% +0.0 0.25 ± 15% perf-profile.children.cycles-pp.next_tgid 0.21 ± 34% +0.0 0.21 ± 19% perf-profile.children.cycles-pp.wq_worker_comm 0.13 ± 32% +0.0 0.13 ± 17% perf-profile.children.cycles-pp.part_stat_read_all 0.56 ± 28% +0.0 0.56 ± 39% perf-profile.children.cycles-pp.alloc_empty_file 0.21 ± 57% +0.0 0.22 ± 29% perf-profile.children.cycles-pp.apparmor_file_permission 0.14 ± 30% +0.0 0.14 ± 29% perf-profile.children.cycles-pp.strnlen_user 0.13 ± 62% +0.0 0.13 ± 28% perf-profile.children.cycles-pp.vm_area_alloc 0.04 ± 76% +0.0 0.05 ± 73% perf-profile.children.cycles-pp.__p4d_alloc 0.06 ± 60% +0.0 0.06 ± 50% perf-profile.children.cycles-pp.current_time 0.02 ±141% +0.0 0.02 ±149% perf-profile.children.cycles-pp.ui_progress__update 0.01 ±223% +0.0 0.01 ±223% perf-profile.children.cycles-pp.idle_cpu 0.12 ± 27% +0.0 0.12 ± 26% perf-profile.children.cycles-pp.padzero 0.09 ± 31% +0.0 0.09 ± 20% perf-profile.children.cycles-pp.clock_gettime 0.06 ± 76% +0.0 0.06 ± 49% perf-profile.children.cycles-pp.folio_add_new_anon_rmap 0.10 ± 29% +0.0 0.11 ± 20% perf-profile.children.cycles-pp.num_to_str 0.05 ± 78% +0.0 0.05 ±105% perf-profile.children.cycles-pp.atomic_read 0.05 ± 77% +0.0 0.05 ±105% perf-profile.children.cycles-pp.refcount_read 0.06 ± 78% +0.0 0.06 ± 63% perf-profile.children.cycles-pp._raw_spin_trylock 0.55 ± 28% +0.0 0.55 ± 40% perf-profile.children.cycles-pp.__alloc_file 0.24 ± 55% +0.0 0.24 ± 25% perf-profile.children.cycles-pp.security_file_permission 0.14 ± 32% +0.0 0.14 ± 13% perf-profile.children.cycles-pp.dequeue_entity 0.09 ± 19% +0.0 0.09 ± 19% perf-profile.children.cycles-pp.getenv 0.04 ±104% +0.0 0.04 ± 72% perf-profile.children.cycles-pp.check_preempt_curr 0.25 ± 34% +0.0 0.26 ± 18% perf-profile.children.cycles-pp.security_ptrace_access_check 0.14 ± 38% +0.0 0.14 ± 19% perf-profile.children.cycles-pp.ttwu_do_activate 0.01 ±223% +0.0 0.02 ±223% perf-profile.children.cycles-pp._find_next_and_bit 0.02 ±146% +0.0 0.03 ±102% perf-profile.children.cycles-pp.printk_get_next_message 0.03 ±108% +0.0 0.04 ± 72% perf-profile.children.cycles-pp.prepare_creds 0.02 ±144% +0.0 0.03 ±100% perf-profile.children.cycles-pp.cap_ptrace_access_check 0.17 ± 73% +0.0 0.18 ± 55% perf-profile.children.cycles-pp.mt_find 0.53 ± 45% +0.0 0.54 ± 22% perf-profile.children.cycles-pp.readn 0.48 ± 54% +0.0 0.49 ± 28% perf-profile.children.cycles-pp.__folio_alloc 0.41 ± 25% +0.0 0.42 ± 21% perf-profile.children.cycles-pp.kmem_cache_free 0.10 ± 36% +0.0 0.10 ± 21% perf-profile.children.cycles-pp.get_zeroed_page 0.02 ±152% +0.0 0.03 ±103% perf-profile.children.cycles-pp.mas_destroy 0.07 ± 53% +0.0 0.08 ± 25% perf-profile.children.cycles-pp.vmf_insert_pfn_prot 0.03 ±150% +0.0 0.04 ± 71% perf-profile.children.cycles-pp.__get_obj_cgroup_from_memcg 0.02 ±146% +0.0 0.03 ±101% perf-profile.children.cycles-pp.vma_expand 0.01 ±223% +0.0 0.02 ±141% perf-profile.children.cycles-pp.__fsnotify_parent 0.01 ±223% +0.0 0.02 ±141% perf-profile.children.cycles-pp.mas_find 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.strtoll 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.sched_mm_cid_remote_clear 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.hrtimer_update_next_event 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.timerqueue_add 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.__bitmap_and 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.rcu_gp_cleanup 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.flush_smp_call_function_queue 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.pm_qos_read_value 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.pgd_ctor 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.strnlen 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp._IO_file_open 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.make_child 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.free_pgd_range 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.lockref_get_not_dead 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.free_p4d_range 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.free_pud_range 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.swake_up_one 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.xas_create 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.generic_write_checks 0.06 ± 61% +0.0 0.06 ± 47% perf-profile.children.cycles-pp._exit 0.06 ± 55% +0.0 0.07 ± 21% perf-profile.children.cycles-pp.security_file_alloc 0.19 ± 17% +0.0 0.20 ± 21% perf-profile.children.cycles-pp.update_curr 0.28 ± 40% +0.0 0.29 ± 22% perf-profile.children.cycles-pp.do_mprotect_pkey 0.38 ± 46% +0.0 0.39 ± 27% perf-profile.children.cycles-pp.__collapse_huge_page_copy 0.13 ± 25% +0.0 0.14 ± 13% perf-profile.children.cycles-pp.__task_pid_nr_ns 0.06 ± 58% +0.0 0.07 ± 26% perf-profile.children.cycles-pp.track_pfn_insert 0.02 ±149% +0.0 0.03 ±106% perf-profile.children.cycles-pp.file_update_time 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.snprintf 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.get_next_timer_interrupt 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.vfree 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.schedule_hrtimeout_range_clock 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.enqueue_hrtimer 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.update_rt_rq_load_avg 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.timerqueue_del 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.uptime_proc_show 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.strscpy_pad 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.irqentry_enter_from_user_mode 0.06 ± 63% +0.0 0.07 ± 29% perf-profile.children.cycles-pp.thread_group_cputime 0.06 ± 58% +0.0 0.07 ± 23% perf-profile.children.cycles-pp.lookup_memtype 0.06 ± 58% +0.0 0.07 ± 23% perf-profile.children.cycles-pp.pat_pagerange_is_ram 0.06 ± 58% +0.0 0.07 ± 23% perf-profile.children.cycles-pp.walk_system_ram_range 0.06 ± 58% +0.0 0.07 ± 23% perf-profile.children.cycles-pp.find_next_iomem_res 0.05 ± 49% +0.0 0.06 ± 54% perf-profile.children.cycles-pp.__pmd_alloc 0.28 ± 40% +0.0 0.29 ± 21% perf-profile.children.cycles-pp.__x64_sys_mprotect 0.22 ± 33% +0.0 0.23 ± 18% perf-profile.children.cycles-pp.__lookup_slow 0.18 ± 38% +0.0 0.19 ± 32% perf-profile.children.cycles-pp.pipe_write 0.04 ±105% +0.0 0.05 ± 74% perf-profile.children.cycles-pp.security_file_open 0.02 ±141% +0.0 0.03 ±100% perf-profile.children.cycles-pp.__send_signal_locked 0.01 ±223% +0.0 0.02 ±223% perf-profile.children.cycles-pp.update_irq_load_avg 0.01 ±223% +0.0 0.02 ±141% perf-profile.children.cycles-pp.__vm_area_free 0.01 ±223% +0.0 0.02 ±141% perf-profile.children.cycles-pp.set_next_entity 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.irqtime_account_process_tick 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.tcp_orphan_count_sum 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.tcp_orphan_update 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.irqentry_enter 0.02 ±223% +0.0 0.03 ±100% perf-profile.children.cycles-pp.proc_pid_permission 0.10 ± 36% +0.0 0.12 ± 24% perf-profile.children.cycles-pp.mas_split 0.05 ± 52% +0.0 0.06 ± 23% perf-profile.children.cycles-pp.aa_may_ptrace 0.16 ± 22% +0.0 0.18 ± 16% perf-profile.children.cycles-pp.perf_iterate_sb 0.12 ± 34% +0.0 0.13 ± 22% perf-profile.children.cycles-pp.seq_put_decimal_ull_width 0.03 ±143% +0.0 0.04 ± 72% perf-profile.children.cycles-pp.vma_link 0.01 ±223% +0.0 0.03 ±100% perf-profile.children.cycles-pp.do_faccessat 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.lapic_next_deadline 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.read_tsc 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.tick_irq_enter 0.00 +0.0 0.01 ±223% perf-profile.children.cycles-pp.arch_scale_freq_tick 0.10 ± 48% +0.0 0.12 ± 33% perf-profile.children.cycles-pp.autoremove_wake_function 0.01 ±223% +0.0 0.02 ± 99% perf-profile.children.cycles-pp.acct_collect 0.08 ± 52% +0.0 0.10 ± 17% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template 0.12 ± 22% +0.0 0.13 ± 24% perf-profile.children.cycles-pp.__kmalloc_node 0.36 ± 15% +0.0 0.38 ± 13% perf-profile.children.cycles-pp.try_to_wake_up 0.15 ± 33% +0.0 0.17 ± 24% perf-profile.children.cycles-pp.__fdget_pos 0.14 ± 40% +0.0 0.16 ± 26% perf-profile.children.cycles-pp.enqueue_entity 0.04 ±101% +0.0 0.05 ± 46% perf-profile.children.cycles-pp.arch_dup_task_struct 0.06 ± 74% +0.0 0.08 ± 15% perf-profile.children.cycles-pp.__perf_event_read 0.06 ± 60% +0.0 0.08 ± 30% perf-profile.children.cycles-pp.thread_group_cputime_adjusted 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.arch_cpu_idle_enter 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.tsc_verify_tsc_adjust 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.run_posix_cpu_timers 0.70 ± 37% +0.0 0.71 ± 21% perf-profile.children.cycles-pp.sched_exec 0.23 ± 40% +0.0 0.24 ± 19% perf-profile.children.cycles-pp.mprotect_fixup 0.06 ± 86% +0.0 0.08 ± 22% perf-profile.children.cycles-pp.__get_free_pages 0.15 ± 39% +0.0 0.17 ± 25% perf-profile.children.cycles-pp.atime_needs_update 0.04 ±106% +0.0 0.06 ± 47% perf-profile.children.cycles-pp.insert_vm_struct 0.00 +0.0 0.02 ±141% perf-profile.children.cycles-pp.get_sigframe 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.get_cpu_device 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.irq_enter_rcu 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.tick_sched_do_timer 0.01 ±223% +0.0 0.03 ±102% perf-profile.children.cycles-pp.show_map_vma 0.01 ±223% +0.0 0.03 ±102% perf-profile.children.cycles-pp.show_map 0.21 ± 42% +0.0 0.23 ± 21% perf-profile.children.cycles-pp.do_cow_fault 0.76 ± 37% +0.0 0.78 ± 21% perf-profile.children.cycles-pp.__x64_sys_vfork 0.52 ± 52% +0.0 0.54 ± 28% perf-profile.children.cycles-pp.vma_alloc_folio 1.12 ± 31% +0.0 1.14 ± 10% perf-profile.children.cycles-pp.do_open 0.88 ± 31% +0.0 0.90 ± 9% perf-profile.children.cycles-pp.do_dentry_open 0.02 ±144% +0.0 0.04 ± 71% perf-profile.children.cycles-pp.__pipe 0.04 ±112% +0.0 0.06 ± 14% perf-profile.children.cycles-pp.__install_special_mapping 0.00 +0.0 0.02 ±142% perf-profile.children.cycles-pp.vm_unmapped_area 0.00 +0.0 0.02 ±142% perf-profile.children.cycles-pp.x64_setup_rt_frame 0.00 +0.0 0.02 ±142% perf-profile.children.cycles-pp._IO_setb 0.00 +0.0 0.02 ±142% perf-profile.children.cycles-pp.__exit_signal 0.00 +0.0 0.02 ±142% perf-profile.children.cycles-pp.down_read_trylock 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.timekeeping_max_deferment 0.00 +0.0 0.02 ±142% perf-profile.children.cycles-pp.delayed_vfree_work 0.00 +0.0 0.02 ±142% perf-profile.children.cycles-pp.strscpy 0.02 ±141% +0.0 0.04 ± 71% perf-profile.children.cycles-pp._IO_link_in 0.03 ±158% +0.0 0.05 ± 75% perf-profile.children.cycles-pp.ktime_get_coarse_real_ts64 0.15 ± 37% +0.0 0.17 ± 19% perf-profile.children.cycles-pp.load_elf_interp 0.10 ± 45% +0.0 0.12 ± 29% perf-profile.children.cycles-pp.clear_page_erms 0.13 ± 34% +0.0 0.15 ± 24% perf-profile.children.cycles-pp.__wake_up_common_lock 0.04 ±114% +0.0 0.06 ± 11% perf-profile.children.cycles-pp.map_vdso 0.01 ±223% +0.0 0.03 ±100% perf-profile.children.cycles-pp.handle_signal 0.00 +0.0 0.02 ±144% perf-profile.children.cycles-pp._IO_getc 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.__run_timers 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.rcu_pending 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.call_timer_fn 0.02 ±223% +0.0 0.04 ±109% perf-profile.children.cycles-pp.__kmalloc 1.31 ± 35% +0.0 1.33 ± 22% perf-profile.children.cycles-pp.find_idlest_group 0.28 ± 28% +0.0 0.30 ± 39% perf-profile.children.cycles-pp.show_stat 0.79 ± 49% +0.0 0.81 ± 21% perf-profile.children.cycles-pp.__alloc_pages 0.32 ± 44% +0.0 0.34 ± 22% perf-profile.children.cycles-pp.elf_map 0.18 ± 33% +0.0 0.20 ± 5% perf-profile.children.cycles-pp.malloc 0.05 ±119% +0.0 0.08 ± 20% perf-profile.children.cycles-pp.__libc_calloc 0.04 ±112% +0.0 0.06 ± 79% perf-profile.children.cycles-pp.kcpustat_cpu_fetch 0.03 ±100% +0.0 0.05 ± 48% perf-profile.children.cycles-pp.filemap_read 0.01 ±223% +0.0 0.04 ±101% perf-profile.children.cycles-pp.rcu_nocb_try_bypass 0.01 ±223% +0.0 0.03 ±103% perf-profile.children.cycles-pp._IO_file_doallocate 0.01 ±223% +0.0 0.03 ±101% perf-profile.children.cycles-pp.mas_store_gfp 0.03 ±143% +0.0 0.05 ± 49% perf-profile.children.cycles-pp.devkmsg_read 0.06 ± 49% +0.0 0.08 ± 28% perf-profile.children.cycles-pp.memcg_account_kmem 0.09 ± 33% +0.0 0.12 ± 9% perf-profile.children.cycles-pp._raw_spin_lock_irq 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.tick_nohz_irq_exit 0.00 +0.0 0.02 ±223% perf-profile.children.cycles-pp.force_qs_rnp 0.01 ±223% +0.0 0.04 ± 71% perf-profile.children.cycles-pp.cpuacct_charge 0.01 ±223% +0.0 0.04 ± 71% perf-profile.children.cycles-pp.get_pid_task 0.10 ± 34% +0.0 0.13 ± 22% perf-profile.children.cycles-pp.__ptrace_may_access 0.39 ± 32% +0.0 0.41 ± 11% perf-profile.children.cycles-pp.ptrace_may_access 0.02 ±141% +0.0 0.04 ±107% perf-profile.children.cycles-pp.__perf_read_group_add 0.02 ±149% +0.0 0.05 ± 80% perf-profile.children.cycles-pp.pgd_alloc 0.14 ± 34% +0.0 0.16 ± 24% perf-profile.children.cycles-pp.__memcg_kmem_charge_page 0.58 ± 34% +0.0 0.60 ± 22% perf-profile.children.cycles-pp.wp_page_copy 0.13 ± 34% +0.0 0.15 ± 33% perf-profile.children.cycles-pp.__call_rcu_common 0.04 ± 71% +0.0 0.06 ± 19% perf-profile.children.cycles-pp.put_dec_trunc8 0.39 ± 61% +0.0 0.42 ± 46% perf-profile.children.cycles-pp.shmem_write_end 0.15 ± 37% +0.0 0.18 ± 27% perf-profile.children.cycles-pp.mas_wr_bnode 1.28 ± 36% +0.0 1.31 ± 22% perf-profile.children.cycles-pp.update_sg_wakeup_stats 0.05 ± 84% +0.0 0.08 ± 22% perf-profile.children.cycles-pp.release_task 0.03 ±102% +0.0 0.06 ± 49% perf-profile.children.cycles-pp.touch_atime 0.00 +0.0 0.03 ±223% perf-profile.children.cycles-pp.rcu_sched_clock_irq 0.22 ± 36% +0.0 0.25 ± 57% perf-profile.children.cycles-pp.dup_task_struct 0.08 ± 47% +0.0 0.12 ± 25% perf-profile.children.cycles-pp.do_open_execat 0.03 ±100% +0.0 0.06 ± 50% perf-profile.children.cycles-pp.path_init 0.06 ± 66% +0.0 0.09 ± 19% perf-profile.children.cycles-pp.wait_task_zombie 0.00 +0.0 0.03 ±103% perf-profile.children.cycles-pp.open_exec 0.06 ± 50% +0.0 0.10 ±143% perf-profile.children.cycles-pp.__intel_pmu_enable_all 0.10 ± 31% +0.0 0.13 ± 17% perf-profile.children.cycles-pp.down_write 0.38 ± 34% +0.0 0.41 ± 20% perf-profile.children.cycles-pp.__split_vma 0.00 +0.0 0.03 ±100% perf-profile.children.cycles-pp.__collapse_huge_page_copy_succeeded 0.00 +0.0 0.03 ±223% perf-profile.children.cycles-pp.cpuidle_governor_latency_req 0.26 ± 37% +0.0 0.30 ± 29% perf-profile.children.cycles-pp.write 0.14 ± 57% +0.0 0.17 ± 24% perf-profile.children.cycles-pp.perf_evlist__poll 0.00 +0.0 0.04 ± 71% perf-profile.children.cycles-pp.memset@plt 0.31 ± 35% +0.0 0.35 ± 19% perf-profile.children.cycles-pp.proc_task_name 0.02 ±144% +0.0 0.06 ± 47% perf-profile.children.cycles-pp.get_unmapped_area 0.08 ± 81% +0.0 0.12 ±172% perf-profile.children.cycles-pp.ktime_get 0.10 ± 35% +0.0 0.14 ± 28% perf-profile.children.cycles-pp.file_free_rcu 0.28 ± 24% +0.0 0.32 ± 25% perf-profile.children.cycles-pp.__mmap 0.08 ± 48% +0.0 0.12 ±122% perf-profile.children.cycles-pp.perf_rotate_context 0.16 ± 62% +0.0 0.20 ± 28% perf-profile.children.cycles-pp.start_thread 0.16 ± 62% +0.0 0.20 ± 28% perf-profile.children.cycles-pp.perf_evlist__poll_thread 0.01 ±223% +0.0 0.05 ± 49% perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown 0.23 ± 30% +0.0 0.27 ± 17% perf-profile.children.cycles-pp.getname_flags 0.02 ±142% +0.0 0.06 ±223% perf-profile.children.cycles-pp.kmem_cache_alloc_node 0.20 ± 32% +0.0 0.25 ± 35% perf-profile.children.cycles-pp.charge_memcg 0.08 ± 17% +0.0 0.12 ± 20% perf-profile.children.cycles-pp.mas_preallocate 0.13 ± 17% +0.0 0.18 ± 16% perf-profile.children.cycles-pp.strncpy_from_user 0.00 +0.0 0.04 ± 73% perf-profile.children.cycles-pp.sscanf 0.00 +0.0 0.04 ± 73% perf-profile.children.cycles-pp.ktime_get_update_offsets_now 0.03 ± 70% +0.0 0.08 ± 48% perf-profile.children.cycles-pp.get_idle_time 0.00 +0.0 0.05 ±159% perf-profile.children.cycles-pp.wait_for_completion_state 0.00 +0.0 0.05 ±223% perf-profile.children.cycles-pp.__update_blocked_fair 0.78 ± 39% +0.0 0.83 ± 20% perf-profile.children.cycles-pp.__vfork 0.08 ± 16% +0.0 0.13 ± 21% perf-profile.children.cycles-pp.mas_alloc_nodes 0.01 ±223% +0.0 0.06 ±112% perf-profile.children.cycles-pp.wait_for_partner 0.10 ± 45% +0.1 0.14 ± 23% perf-profile.children.cycles-pp.vma_complete 0.09 ± 56% +0.1 0.14 ± 67% perf-profile.children.cycles-pp.__wait_for_common 0.24 ± 33% +0.1 0.29 ± 17% perf-profile.children.cycles-pp.__slab_free 0.04 ±104% +0.1 0.09 ±160% perf-profile.children.cycles-pp.balance_fair 0.73 ± 31% +0.1 0.78 ± 21% perf-profile.children.cycles-pp.ksys_mmap_pgoff 0.07 ± 37% +0.1 0.12 ±109% perf-profile.children.cycles-pp.obj_cgroup_charge 0.00 +0.1 0.06 ± 78% perf-profile.children.cycles-pp.drm_gem_check_release_pagevec 0.00 +0.1 0.06 ± 65% perf-profile.children.cycles-pp.get_cpu_sleep_time_us 0.01 ±223% +0.1 0.07 ± 90% perf-profile.children.cycles-pp.fifo_open 0.13 ± 23% +0.1 0.19 ± 59% perf-profile.children.cycles-pp.try_charge_memcg 7.93 ± 55% +0.1 7.99 ± 51% perf-profile.children.cycles-pp.__libc_start_main 7.92 ± 55% +0.1 7.98 ± 51% perf-profile.children.cycles-pp.main 7.92 ± 55% +0.1 7.98 ± 51% perf-profile.children.cycles-pp.run_builtin 0.00 +0.1 0.06 ±223% perf-profile.children.cycles-pp.tick_nohz_next_event 0.12 ± 33% +0.1 0.19 ±123% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler 0.14 ± 41% +0.1 0.20 ± 28% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm 0.18 ± 27% +0.1 0.25 ± 8% perf-profile.children.cycles-pp.rcu_cblist_dequeue 0.03 ±102% +0.1 0.10 ± 24% perf-profile.children.cycles-pp.kmem_cache_alloc_bulk 0.01 ±223% +0.1 0.08 ±141% perf-profile.children.cycles-pp.schedule_tail 0.00 +0.1 0.07 ± 53% perf-profile.children.cycles-pp.__snprintf_chk 5.97 ± 27% +0.1 6.04 ± 15% perf-profile.children.cycles-pp.read 0.01 ±223% +0.1 0.09 ± 22% perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk 0.07 ±134% +0.1 0.15 ± 90% perf-profile.children.cycles-pp.schedule_idle 0.00 +0.1 0.08 ±168% perf-profile.children.cycles-pp.run_rebalance_domains 0.00 +0.1 0.08 ± 17% perf-profile.children.cycles-pp.copyout 0.65 ± 30% +0.1 0.74 ± 86% perf-profile.children.cycles-pp.__hrtimer_run_queues 0.00 +0.1 0.08 ± 20% perf-profile.children.cycles-pp._IO_fread 0.00 +0.1 0.08 ±193% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length 1.87 ± 37% +0.1 1.96 ± 24% perf-profile.children.cycles-pp.process_interval 0.14 ±112% +0.1 0.23 ±128% perf-profile.children.cycles-pp.__pthread_disable_asynccancel 0.00 +0.1 0.09 ±165% perf-profile.children.cycles-pp.update_blocked_averages 1.87 ± 37% +0.1 1.96 ± 24% perf-profile.children.cycles-pp.dispatch_events 2.63 ± 26% +0.1 2.72 ± 15% perf-profile.children.cycles-pp.proc_single_show 2.62 ± 26% +0.1 2.71 ± 15% perf-profile.children.cycles-pp.do_task_stat 0.75 ± 42% +0.1 0.85 ± 14% perf-profile.children.cycles-pp.rcu_do_batch 1.75 ± 37% +0.1 1.85 ± 24% perf-profile.children.cycles-pp.read_counters 0.00 +0.1 0.10 ± 32% perf-profile.children.cycles-pp.cgroup_rstat_flush 0.04 ±168% +0.1 0.14 ±169% perf-profile.children.cycles-pp.find_busiest_queue 0.91 ± 33% +0.1 1.01 ± 20% perf-profile.children.cycles-pp.mmap_region 0.00 +0.1 0.10 ± 35% perf-profile.children.cycles-pp.get_signal 0.82 ± 38% +0.1 0.93 ± 14% perf-profile.children.cycles-pp.rcu_core 0.35 ± 34% +0.1 0.46 ± 9% perf-profile.children.cycles-pp.__mem_cgroup_charge 1.00 ± 34% +0.1 1.11 ± 19% perf-profile.children.cycles-pp.vm_mmap_pgoff 0.63 ± 39% +0.1 0.74 ± 18% perf-profile.children.cycles-pp.dup_mm 1.12 ± 36% +0.1 1.23 ± 19% perf-profile.children.cycles-pp.__do_sys_clone 0.35 ± 65% +0.1 0.46 ± 84% perf-profile.children.cycles-pp.perf_mmap__read_head 0.35 ± 66% +0.1 0.46 ± 83% perf-profile.children.cycles-pp.ring_buffer_read_head 0.01 ±223% +0.1 0.12 ±126% perf-profile.children.cycles-pp.do_task_dead 0.04 ± 89% +0.1 0.16 ± 16% perf-profile.children.cycles-pp.mas_wr_modify 0.97 ± 34% +0.1 1.09 ± 19% perf-profile.children.cycles-pp.do_mmap 0.79 ± 28% +0.1 0.91 ± 83% perf-profile.children.cycles-pp.hrtimer_interrupt 0.02 ±223% +0.1 0.14 ± 19% perf-profile.children.cycles-pp.mas_wr_node_store 0.02 ±146% +0.1 0.15 ± 33% perf-profile.children.cycles-pp.arch_do_signal_or_restart 1.87 ± 36% +0.1 2.00 ± 18% perf-profile.children.cycles-pp.kernel_clone 0.81 ± 28% +0.1 0.94 ± 82% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 0.12 ± 51% +0.1 0.25 ± 63% perf-profile.children.cycles-pp.evlist__id2evsel 0.22 ± 46% +0.1 0.36 ± 21% perf-profile.children.cycles-pp.mas_store_prealloc 0.47 ± 32% +0.1 0.61 ± 19% perf-profile.children.cycles-pp.do_anonymous_page 1.06 ± 33% +0.1 1.20 ± 19% perf-profile.children.cycles-pp.copy_process 0.00 +0.1 0.14 ± 36% perf-profile.children.cycles-pp.update_rq_clock_task 3.77 ± 23% +0.1 3.91 ± 35% perf-profile.children.cycles-pp.kthread 0.16 ± 43% +0.2 0.32 ± 49% perf-profile.children.cycles-pp.pipe_read 0.80 ± 44% +0.2 0.96 ± 24% perf-profile.children.cycles-pp.syscall_exit_to_user_mode 0.00 +0.2 0.16 ± 18% perf-profile.children.cycles-pp.check_heap_object 0.66 ± 46% +0.2 0.82 ± 35% perf-profile.children.cycles-pp.finish_task_switch 0.65 ± 39% +0.2 0.82 ± 34% perf-profile.children.cycles-pp.__percpu_counter_sum 0.71 ± 38% +0.2 0.89 ± 32% perf-profile.children.cycles-pp.__mmdrop 0.00 +0.2 0.18 ±185% perf-profile.children.cycles-pp.menu_select 0.00 +0.2 0.19 ± 29% perf-profile.children.cycles-pp.mm_account_fault 1.77 ± 27% +0.2 1.96 ± 24% perf-profile.children.cycles-pp.cmd_stat 1.18 ± 36% +0.2 1.37 ± 19% perf-profile.children.cycles-pp.__libc_fork 3.79 ± 23% +0.2 4.00 ± 37% perf-profile.children.cycles-pp.ret_from_fork 0.75 ± 38% +0.2 0.97 ± 24% perf-profile.children.cycles-pp.alloc_bprm 0.00 +0.2 0.22 ± 39% perf-profile.children.cycles-pp.mm_init_cid 0.00 +0.2 0.24 ± 28% perf-profile.children.cycles-pp.__rmqueue_pcplist 0.23 ± 39% +0.2 0.48 ± 60% perf-profile.children.cycles-pp.kernel_wait4 0.01 ±223% +0.2 0.26 ±164% perf-profile.children.cycles-pp.arch_call_rest_init 0.01 ±223% +0.2 0.26 ±164% perf-profile.children.cycles-pp.start_kernel 0.01 ±223% +0.2 0.26 ±164% perf-profile.children.cycles-pp.rest_init 0.66 ± 28% +0.2 0.90 ± 25% perf-profile.children.cycles-pp.exec_mmap 0.23 ± 40% +0.2 0.48 ± 60% perf-profile.children.cycles-pp.do_wait 0.00 +0.3 0.26 ±164% perf-profile.children.cycles-pp.x86_64_start_kernel 0.00 +0.3 0.26 ±164% perf-profile.children.cycles-pp.x86_64_start_reservations 0.00 +0.3 0.26 ± 30% perf-profile.children.cycles-pp.mtree_range_walk 0.87 ± 38% +0.3 1.15 ± 21% perf-profile.children.cycles-pp.pcpu_alloc 0.74 ± 30% +0.3 1.01 ± 24% perf-profile.children.cycles-pp.begin_new_exec 0.00 +0.3 0.28 ± 19% perf-profile.children.cycles-pp.drm_fbdev_generic_damage_blit_real 0.00 +0.3 0.29 ± 21% perf-profile.children.cycles-pp.activate_task 1.66 ± 34% +0.3 1.96 ± 22% perf-profile.children.cycles-pp.exec_binprm 1.65 ± 34% +0.3 1.95 ± 22% perf-profile.children.cycles-pp.search_binary_handler 0.29 ± 57% +0.3 0.60 ± 61% perf-profile.children.cycles-pp.evsel__read_counter 0.25 ± 71% +0.3 0.56 ± 72% perf-profile.children.cycles-pp.rebalance_domains 1.64 ± 33% +0.3 1.94 ± 22% perf-profile.children.cycles-pp.load_elf_binary 0.00 +0.3 0.31 ± 29% perf-profile.children.cycles-pp.drm_gem_shmem_get_pages 0.00 +0.3 0.32 ± 22% perf-profile.children.cycles-pp.__wp_page_copy_user 0.33 ± 39% +0.3 0.66 ± 23% perf-profile.children.cycles-pp.copy_mc_enhanced_fast_string 2.48 ± 35% +0.3 2.82 ± 21% perf-profile.children.cycles-pp.bprm_execve 0.28 ± 41% +0.3 0.62 ± 50% perf-profile.children.cycles-pp.wait4 0.94 ± 39% +0.4 1.28 ± 22% perf-profile.children.cycles-pp.mm_init 0.00 +0.4 0.36 ± 28% perf-profile.children.cycles-pp.drm_gem_shmem_vmap_locked 0.00 +0.4 0.37 ± 29% perf-profile.children.cycles-pp.drm_gem_shmem_object_vmap 0.03 ±223% +0.4 0.40 ±198% perf-profile.children.cycles-pp.intel_idle 0.00 +0.5 0.48 ± 11% perf-profile.children.cycles-pp.__memcpy 0.92 ± 38% +0.5 1.41 ± 22% perf-profile.children.cycles-pp.irqentry_exit_to_user_mode 1.14 ± 34% +0.5 1.66 ± 34% perf-profile.children.cycles-pp.__do_softirq 3.57 ± 36% +0.5 4.11 ± 22% perf-profile.children.cycles-pp.execve 3.57 ± 36% +0.5 4.12 ± 22% perf-profile.children.cycles-pp.do_execveat_common 3.58 ± 36% +0.6 4.13 ± 22% perf-profile.children.cycles-pp.__x64_sys_execve 0.01 ±223% +0.6 0.60 ±175% perf-profile.children.cycles-pp.rcu_gp_fqs_loop 0.72 ± 36% +0.6 1.32 ± 50% perf-profile.children.cycles-pp.__irq_exit_rcu 0.01 ±223% +0.6 0.62 ±169% perf-profile.children.cycles-pp.schedule_timeout 0.29 ± 42% +0.6 0.92 ± 20% perf-profile.children.cycles-pp.__memset 0.02 ±223% +0.6 0.66 ±201% perf-profile.children.cycles-pp.intel_idle_xstate 0.02 ±146% +0.6 0.67 ±172% perf-profile.children.cycles-pp.rcu_gp_kthread 1.68 ± 32% +0.7 2.35 ± 21% perf-profile.children.cycles-pp.exit_to_user_mode_prepare 1.43 ± 33% +0.7 2.14 ± 22% perf-profile.children.cycles-pp.exit_to_user_mode_loop 0.89 ± 43% +0.8 1.69 ± 20% perf-profile.children.cycles-pp.task_work_run 1.54 ± 26% +0.8 2.35 ± 63% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt 1.50 ± 40% +0.9 2.43 ± 67% perf-profile.children.cycles-pp.schedule 0.42 ± 69% +0.9 1.35 ±143% perf-profile.children.cycles-pp.pick_next_task_fair 0.39 ± 72% +1.0 1.35 ±154% perf-profile.children.cycles-pp.newidle_balance 0.52 ± 72% +1.0 1.52 ±126% perf-profile.children.cycles-pp.update_sg_lb_stats 0.54 ± 72% +1.1 1.59 ±126% perf-profile.children.cycles-pp.update_sd_lb_stats 0.54 ± 73% +1.1 1.60 ±126% perf-profile.children.cycles-pp.find_busiest_group 0.00 +1.1 1.08 ± 15% perf-profile.children.cycles-pp.drm_fbdev_generic_helper_fb_dirty 1.68 ± 41% +1.1 2.79 ± 68% perf-profile.children.cycles-pp.__schedule 4.06 ± 25% +1.1 5.18 ± 27% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt 0.00 +1.1 1.13 ± 20% perf-profile.children.cycles-pp.task_mm_cid_work 0.63 ± 72% +1.2 1.84 ±126% perf-profile.children.cycles-pp.load_balance 0.21 ± 94% +2.0 2.26 ±192% perf-profile.children.cycles-pp.cpuidle_enter_state 0.00 +2.0 2.05 ± 58% perf-profile.children.cycles-pp.copyin 0.21 ± 94% +2.1 2.26 ±192% perf-profile.children.cycles-pp.cpuidle_enter 0.31 ± 89% +2.1 2.42 ±187% perf-profile.children.cycles-pp.start_secondary 0.23 ± 91% +2.3 2.48 ±191% perf-profile.children.cycles-pp.cpuidle_idle_call 0.33 ± 87% +2.3 2.67 ±184% perf-profile.children.cycles-pp.secondary_startup_64_no_verify 0.33 ± 87% +2.3 2.67 ±184% perf-profile.children.cycles-pp.cpu_startup_entry 0.33 ± 87% +2.3 2.67 ±184% perf-profile.children.cycles-pp.do_idle 53.13 ± 31% +5.9 59.07 ± 11% perf-profile.children.cycles-pp.__cmd_record 45.08 ± 45% +6.9 51.98 ± 8% perf-profile.children.cycles-pp.record__finish_output 45.07 ± 45% +6.9 51.98 ± 8% perf-profile.children.cycles-pp.perf_session__process_events 44.88 ± 45% +7.1 51.95 ± 8% perf-profile.children.cycles-pp.reader__read_event 39.40 ± 47% +11.2 50.62 ± 10% perf-profile.children.cycles-pp.process_simple 38.23 ± 48% +11.6 49.86 ± 11% perf-profile.children.cycles-pp.ordered_events__queue 38.12 ± 48% +11.9 50.04 ± 10% perf-profile.children.cycles-pp.queue_event 8.95 ± 66% -7.8 1.16 ±147% perf-profile.self.cycles-pp.copy_page 2.34 ± 51% -2.3 0.00 perf-profile.self.cycles-pp.copy_user_enhanced_fast_string 1.88 ±111% -1.4 0.42 ±223% perf-profile.self.cycles-pp.evlist__parse_sample 0.92 ±109% -0.8 0.12 ±223% perf-profile.self.cycles-pp.evsel__parse_sample 0.67 ±105% -0.5 0.17 ±223% perf-profile.self.cycles-pp.__ordered_events__flush 0.49 ± 31% -0.5 0.00 perf-profile.self.cycles-pp.memcpy_erms 0.44 ± 43% -0.4 0.00 perf-profile.self.cycles-pp.memset_erms 0.44 ±128% -0.3 0.12 ±223% perf-profile.self.cycles-pp.machine__findnew_thread 0.60 ±124% -0.3 0.31 ± 59% perf-profile.self.cycles-pp.reader__read_event 0.70 ±120% -0.3 0.44 ± 67% perf-profile.self.cycles-pp.evlist__parse_sample_timestamp 1.52 ± 22% -0.3 1.26 ± 17% perf-profile.self.cycles-pp.asm_sysvec_apic_timer_interrupt 0.30 ± 47% -0.2 0.06 ±118% perf-profile.self.cycles-pp.smp_call_function_many_cond 0.21 ±144% -0.2 0.00 perf-profile.self.cycles-pp.rb_erase 0.62 ± 27% -0.2 0.43 ± 12% perf-profile.self.cycles-pp.io_serial_in 0.17 ± 73% -0.2 0.01 ±223% perf-profile.self.cycles-pp.mt_find 0.36 ± 33% -0.1 0.21 ± 36% perf-profile.self.cycles-pp.perf_poll 0.17 ± 90% -0.1 0.03 ±223% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath 0.65 ± 23% -0.1 0.52 ± 23% perf-profile.self.cycles-pp.next_uptodate_page 0.15 ± 28% -0.1 0.01 ±223% perf-profile.self.cycles-pp.mas_walk 0.64 ± 27% -0.1 0.51 ± 22% perf-profile.self.cycles-pp.zap_pte_range 0.12 ± 45% -0.1 0.00 perf-profile.self.cycles-pp.common_perm_cond 0.29 ±109% -0.1 0.16 ± 59% perf-profile.self.cycles-pp.evsel__parse_sample_timestamp 0.38 ± 27% -0.1 0.26 ± 12% perf-profile.self.cycles-pp.__fget_light 0.13 ± 40% -0.1 0.01 ±223% perf-profile.self.cycles-pp.llist_add_batch 0.10 ±195% -0.1 0.00 perf-profile.self.cycles-pp.rb_next 0.14 ± 58% -0.1 0.03 ±111% perf-profile.self.cycles-pp.free_unref_page_prepare 0.38 ± 42% -0.1 0.27 ± 19% perf-profile.self.cycles-pp.__list_del_entry_valid 0.23 ± 61% -0.1 0.13 ± 18% perf-profile.self.cycles-pp.mutex_lock 0.45 ± 41% -0.1 0.36 ± 28% perf-profile.self.cycles-pp.native_irq_return_iret 0.50 ± 28% -0.1 0.41 ± 21% perf-profile.self.cycles-pp.fold_vm_numa_events 0.08 ± 22% -0.1 0.00 perf-profile.self.cycles-pp.update_rq_clock 0.09 ± 17% -0.1 0.01 ±223% perf-profile.self.cycles-pp.handle_mm_fault 0.34 ± 23% -0.1 0.26 ± 22% perf-profile.self.cycles-pp._dl_addr 0.40 ± 24% -0.1 0.33 ± 23% perf-profile.self.cycles-pp._raw_spin_lock_irqsave 0.07 ± 82% -0.1 0.00 perf-profile.self.cycles-pp.task_numa_compare 0.06 ± 65% -0.1 0.00 perf-profile.self.cycles-pp.__flush_smp_call_function_queue 0.06 ± 47% -0.1 0.00 perf-profile.self.cycles-pp.__lock_task_sighand 0.06 ±133% -0.1 0.00 perf-profile.self.cycles-pp.find_idlest_cpu 0.31 ± 42% -0.1 0.25 ± 26% perf-profile.self.cycles-pp.filemap_get_entry 0.06 ± 94% -0.1 0.00 perf-profile.self.cycles-pp.llist_reverse_order 0.06 ±223% -0.1 0.00 perf-profile.self.cycles-pp.__free_pages 0.47 ± 34% -0.1 0.41 ± 14% perf-profile.self.cycles-pp._raw_spin_lock 0.22 ± 27% -0.1 0.17 ± 15% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state 0.06 ± 76% -0.1 0.00 perf-profile.self.cycles-pp.map_id_up 0.08 ± 43% -0.1 0.02 ±142% perf-profile.self.cycles-pp.folio_unlock 0.18 ± 29% -0.1 0.13 ± 33% perf-profile.self.cycles-pp.__count_memcg_events 0.73 ± 30% -0.1 0.67 ± 18% perf-profile.self.cycles-pp.__d_lookup_rcu 0.07 ± 39% -0.1 0.02 ±142% perf-profile.self.cycles-pp.alloc_vmap_area 0.08 ± 60% -0.1 0.03 ±101% perf-profile.self.cycles-pp.vsnprintf 0.06 ±145% -0.0 0.01 ±223% perf-profile.self.cycles-pp.thread__find_map 0.06 ±107% -0.0 0.01 ±223% perf-profile.self.cycles-pp.flush_tlb_func 0.21 ± 36% -0.0 0.16 ± 31% perf-profile.self.cycles-pp.___perf_sw_event 0.33 ± 38% -0.0 0.29 ± 22% perf-profile.self.cycles-pp.all_vm_events 0.04 ±151% -0.0 0.00 perf-profile.self.cycles-pp.thread__get 0.11 ± 60% -0.0 0.07 ± 50% perf-profile.self.cycles-pp.__mod_node_page_state 0.06 ± 14% -0.0 0.02 ±142% perf-profile.self.cycles-pp.memchr_inv 0.05 ± 84% -0.0 0.01 ±223% perf-profile.self.cycles-pp.change_pte_range 0.08 ± 29% -0.0 0.04 ± 75% perf-profile.self.cycles-pp.syscall_return_via_sysret 0.26 ± 33% -0.0 0.22 ± 17% perf-profile.self.cycles-pp.__alloc_file 0.24 ± 30% -0.0 0.20 ± 12% perf-profile.self.cycles-pp.__d_lookup 0.29 ± 24% -0.0 0.25 ± 19% perf-profile.self.cycles-pp.memcpy_toio 0.06 ± 84% -0.0 0.02 ±142% perf-profile.self.cycles-pp.vma_alloc_folio 0.06 ± 47% -0.0 0.02 ±146% perf-profile.self.cycles-pp.fput 0.17 ± 32% -0.0 0.13 ± 25% perf-profile.self.cycles-pp.update_load_avg 0.18 ± 32% -0.0 0.15 ± 25% perf-profile.self.cycles-pp.vma_interval_tree_remove 0.06 ± 49% -0.0 0.02 ±141% perf-profile.self.cycles-pp.lock_vma_under_rcu 0.04 ± 71% -0.0 0.00 perf-profile.self.cycles-pp.__mod_memcg_state 0.05 ± 75% -0.0 0.01 ±223% perf-profile.self.cycles-pp.__perf_event_read 0.06 ± 51% -0.0 0.03 ±102% perf-profile.self.cycles-pp.__check_object_size 0.20 ± 14% -0.0 0.16 ± 19% perf-profile.self.cycles-pp.sync_regs 0.16 ± 44% -0.0 0.12 ± 19% perf-profile.self.cycles-pp.lookup_fast 0.14 ± 29% -0.0 0.10 ± 15% perf-profile.self.cycles-pp.io_serial_out 0.04 ±101% -0.0 0.00 perf-profile.self.cycles-pp.ttwu_do_activate 0.07 ± 31% -0.0 0.04 ±107% perf-profile.self.cycles-pp.kstat_irqs_usr 0.35 ± 32% -0.0 0.31 ± 15% perf-profile.self.cycles-pp.gather_pte_stats 0.12 ± 29% -0.0 0.08 ± 29% perf-profile.self.cycles-pp.mem_cgroup_css_rstat_flush 0.03 ±223% -0.0 0.00 perf-profile.self.cycles-pp.io__get_char 0.03 ±100% -0.0 0.00 perf-profile.self.cycles-pp.prepare_task_switch 0.03 ±108% -0.0 0.00 perf-profile.self.cycles-pp._compound_head 0.03 ±108% -0.0 0.00 perf-profile.self.cycles-pp.step_into 0.09 ± 67% -0.0 0.06 ± 47% perf-profile.self.cycles-pp.perf_exclude_event 0.05 ±121% -0.0 0.02 ±141% perf-profile.self.cycles-pp.__free_one_page 0.10 ± 23% -0.0 0.07 ± 15% perf-profile.self.cycles-pp.__switch_to 0.03 ±223% -0.0 0.00 perf-profile.self.cycles-pp.map__process_kallsym_symbol 0.03 ±146% -0.0 0.00 perf-profile.self.cycles-pp.generic_permission 0.35 ± 35% -0.0 0.32 ± 18% perf-profile.self.cycles-pp.kmem_cache_alloc 0.19 ± 26% -0.0 0.16 ± 16% perf-profile.self.cycles-pp.xas_load 0.03 ±144% -0.0 0.00 perf-profile.self.cycles-pp.perf_session__process_events 0.03 ±100% -0.0 0.00 perf-profile.self.cycles-pp.irqtime_account_irq 0.26 ± 18% -0.0 0.23 ± 20% perf-profile.self.cycles-pp.vma_interval_tree_insert 0.04 ±105% -0.0 0.01 ±223% perf-profile.self.cycles-pp.handle_pte_fault 0.03 ±100% -0.0 0.00 perf-profile.self.cycles-pp.read 0.03 ±100% -0.0 0.00 perf-profile.self.cycles-pp.proc_pid_readlink 0.03 ±100% -0.0 0.00 perf-profile.self.cycles-pp.get_obj_cgroup_from_current 0.08 ± 56% -0.0 0.06 ± 48% perf-profile.self.cycles-pp.vfs_read 0.02 ± 99% -0.0 0.00 perf-profile.self.cycles-pp.perf_swevent_event 0.02 ± 99% -0.0 0.00 perf-profile.self.cycles-pp.get_cpu_idle_time_us 0.02 ±152% -0.0 0.00 perf-profile.self.cycles-pp.events_stats__inc 0.02 ±223% -0.0 0.00 perf-profile.self.cycles-pp.kallsyms_expand_symbol 0.02 ±145% -0.0 0.00 perf-profile.self.cycles-pp.exit_to_user_mode_prepare 0.10 ± 47% -0.0 0.08 ± 48% perf-profile.self.cycles-pp.free_swap_cache 0.04 ±104% -0.0 0.01 ±223% perf-profile.self.cycles-pp.update_numa_stats 0.03 ±108% -0.0 0.01 ±223% perf-profile.self.cycles-pp.acpi_ns_search_and_enter 0.02 ±149% -0.0 0.00 perf-profile.self.cycles-pp.lru_add_drain_cpu 0.02 ±223% -0.0 0.00 perf-profile.self.cycles-pp.kernfs_dop_revalidate 0.14 ± 37% -0.0 0.12 ± 24% perf-profile.self.cycles-pp.mod_objcg_state 0.06 ± 28% -0.0 0.04 ± 72% perf-profile.self.cycles-pp.arch_irq_stat_cpu 0.05 ± 80% -0.0 0.03 ±102% perf-profile.self.cycles-pp.vmap_pages_pud_range 0.03 ±157% -0.0 0.01 ±223% perf-profile.self.cycles-pp.tlb_is_not_lazy 0.27 ± 58% -0.0 0.25 ± 83% perf-profile.self.cycles-pp.record__mmap_read_evlist 0.10 ± 68% -0.0 0.07 ± 76% perf-profile.self.cycles-pp.perf_mmap_fault 0.19 ± 28% -0.0 0.16 ± 20% perf-profile.self.cycles-pp.restore_fpregs_from_fpstate 0.02 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__dsos__find 0.10 ± 58% -0.0 0.08 ± 49% perf-profile.self.cycles-pp.get_page_from_freelist 0.02 ±142% -0.0 0.00 perf-profile.self.cycles-pp.security_vm_enough_memory_mm 0.02 ±142% -0.0 0.00 perf-profile.self.cycles-pp.mntput_no_expire 0.02 ±142% -0.0 0.00 perf-profile.self.cycles-pp.mem_cgroup_from_task 0.07 ± 78% -0.0 0.05 ±114% perf-profile.self.cycles-pp.generic_perform_write 0.02 ±144% -0.0 0.00 perf-profile.self.cycles-pp.readdir64 0.13 ± 57% -0.0 0.11 ± 47% perf-profile.self.cycles-pp.aa_file_perm 0.18 ± 29% -0.0 0.16 ± 25% perf-profile.self.cycles-pp.__kmem_cache_alloc_node 0.12 ± 43% -0.0 0.10 ± 21% perf-profile.self.cycles-pp.percpu_counter_add_batch 0.08 ± 35% -0.0 0.07 ± 14% perf-profile.self.cycles-pp.__switch_to_asm 0.03 ±147% -0.0 0.01 ±223% perf-profile.self.cycles-pp.cap_vm_enough_memory 0.02 ±142% -0.0 0.00 perf-profile.self.cycles-pp.hrtimer_active 0.02 ±223% -0.0 0.00 perf-profile.self.cycles-pp.map__put 0.02 ±142% -0.0 0.00 perf-profile.self.cycles-pp.mntput 0.17 ± 30% -0.0 0.15 ± 20% perf-profile.self.cycles-pp.pid_nr_ns 0.12 ± 39% -0.0 0.10 ± 48% perf-profile.self.cycles-pp.do_user_addr_fault 0.03 ±155% -0.0 0.01 ±223% perf-profile.self.cycles-pp.blk_cgroup_congested 0.02 ±223% -0.0 0.00 perf-profile.self.cycles-pp.____machine__findnew_thread 0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.__get_vma_policy 0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.task_prio 0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.schedule 0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.task_tick_fair 0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack 0.13 ± 50% -0.0 0.11 ± 17% perf-profile.self.cycles-pp.proc_fill_cache 0.07 ± 69% -0.0 0.05 ±102% perf-profile.self.cycles-pp.__perf_mmap__read_init 0.18 ± 25% -0.0 0.17 ± 29% perf-profile.self.cycles-pp.blk_mq_queue_tag_busy_iter 0.02 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__x2apic_send_IPI_mask 0.02 ±223% -0.0 0.00 perf-profile.self.cycles-pp.proc_single_open 0.03 ±100% -0.0 0.01 ±223% perf-profile.self.cycles-pp.__list_add_valid 0.12 ± 21% -0.0 0.10 ± 24% perf-profile.self.cycles-pp.try_charge_memcg 0.06 ± 48% -0.0 0.05 ± 74% perf-profile.self.cycles-pp.evsel__compute_deltas 0.08 ± 35% -0.0 0.07 ± 50% perf-profile.self.cycles-pp.__mod_lruvec_page_state 0.05 ± 87% -0.0 0.03 ± 70% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe 0.03 ±108% -0.0 0.02 ±142% perf-profile.self.cycles-pp.enqueue_task_fair 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.kallsyms__parse 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.module_get_kallsym 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__dsos__findnew_link_by_longname_id 0.03 ±100% -0.0 0.01 ±223% perf-profile.self.cycles-pp.native_sched_clock 0.02 ±142% -0.0 0.01 ±223% perf-profile.self.cycles-pp.exit_to_user_mode_loop 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.perf_event_read 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.kfree 0.04 ± 75% -0.0 0.03 ±100% perf-profile.self.cycles-pp.charge_memcg 0.03 ±103% -0.0 0.02 ±144% perf-profile.self.cycles-pp.string 0.04 ± 72% -0.0 0.02 ± 99% perf-profile.self.cycles-pp.unlink_anon_vmas 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.rb_insert_color 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.format_decode 0.02 ±144% -0.0 0.01 ±223% perf-profile.self.cycles-pp.strlen 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.get_kthread_comm 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.rb_first 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.perf_evsel__read 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__perf_sw_event 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__cond_resched 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.anon_vma_clone 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.rw_verify_area 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.access_error 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.mpol_shared_policy_lookup 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.fclose 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.xyarray__entry 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.perf_event_ctx_lock_nested 0.04 ±111% -0.0 0.03 ±101% perf-profile.self.cycles-pp.__perf_session__process_decomp_events 0.11 ± 40% -0.0 0.10 ± 14% perf-profile.self.cycles-pp.release_pages 0.11 ± 28% -0.0 0.10 ± 16% perf-profile.self.cycles-pp.update_curr 0.06 ± 58% -0.0 0.06 ± 45% perf-profile.self.cycles-pp.unmap_single_vma 0.05 ± 47% -0.0 0.04 ± 71% perf-profile.self.cycles-pp.native_flush_tlb_one_user 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.io__get_hex 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.map__new 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.number 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.save_fpregs_to_fpstate 0.02 ±144% -0.0 0.01 ±223% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.apparmor_file_free_security 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.perf_env__arch 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.getname_flags 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.shmem_alloc_and_acct_folio 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.filp_close 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__irqentry_text_end 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.put_prev_task_stop 0.12 ± 29% -0.0 0.12 ± 16% perf-profile.self.cycles-pp.task_dump_owner 0.11 ± 57% -0.0 0.10 ± 33% perf-profile.self.cycles-pp.process_simple 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.update_iter 0.02 ±141% -0.0 0.01 ±223% perf-profile.self.cycles-pp.kmem_cache_alloc_lru 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__vm_enough_memory 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__comm_str__findnew 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.may_open 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.obj_cgroup_uncharge_pages 0.02 ±141% -0.0 0.01 ±223% perf-profile.self.cycles-pp.mutex_unlock 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.perf_trace_sched_migrate_task 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.m_next 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__update_load_avg_se 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.asm_exc_page_fault 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.ksys_read 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.do_mmap 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.finish_task_switch 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.asm_sysvec_call_function 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.try_to_wake_up 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.perf_event_mmap_event 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.do_set_pte 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.kmalloc_trace 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.add_wait_queue 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__do_set_cpus_allowed 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.select_task_rq 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.vfs_readlink 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.file_ra_state_init 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.lockref_put_or_lock 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.rcu_all_qs 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.__check_heap_object 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.check_preempt_wakeup 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.acpi_ex_system_memory_space_handler 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.task_h_load 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.update_cfs_group 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.fdarray__filter 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.apparmor_file_alloc_security 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.migrate_task_rq_fair 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp._IO_file_close_it 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.cpu__max_cpu 0.01 ±223% -0.0 0.00 perf-profile.self.cycles-pp.malloc@plt 0.09 ± 35% -0.0 0.08 ± 22% perf-profile.self.cycles-pp.pid_task 0.06 ± 82% -0.0 0.05 ± 46% perf-profile.self.cycles-pp.wq_worker_comm 0.11 ± 43% -0.0 0.10 ± 15% perf-profile.self.cycles-pp.errseq_sample 0.02 ± 99% -0.0 0.02 ±142% perf-profile.self.cycles-pp.seq_read_iter 0.02 ±149% -0.0 0.02 ±141% perf-profile.self.cycles-pp.folio_batch_move_lru 0.02 ±141% -0.0 0.01 ±223% perf-profile.self.cycles-pp.kthread_data 0.02 ±223% -0.0 0.01 ±223% perf-profile.self.cycles-pp.shmem_get_folio_gfp 0.04 ±105% -0.0 0.03 ±103% perf-profile.self.cycles-pp.page_remove_rmap 0.06 ± 70% -0.0 0.06 ± 88% perf-profile.self.cycles-pp.perf_mmap_to_page 0.10 ± 39% -0.0 0.10 ± 26% perf-profile.self.cycles-pp.cgroup_rstat_updated 0.09 ± 36% -0.0 0.09 ± 21% perf-profile.self.cycles-pp.aa_get_task_label 0.07 ± 53% -0.0 0.06 ± 55% perf-profile.self.cycles-pp.gather_stats 0.11 ± 25% -0.0 0.10 ± 18% perf-profile.self.cycles-pp.__virt_addr_valid 0.04 ± 79% -0.0 0.04 ± 75% perf-profile.self.cycles-pp.thread_group_cputime 0.12 ± 38% -0.0 0.11 ± 25% perf-profile.self.cycles-pp.do_syscall_64 0.07 ± 35% -0.0 0.07 ± 47% perf-profile.self.cycles-pp.switch_mm_irqs_off 0.02 ±142% -0.0 0.02 ±142% perf-profile.self.cycles-pp.vfs_write 0.01 ±223% -0.0 0.01 ±223% perf-profile.self.cycles-pp.num_to_str 0.01 ±223% -0.0 0.01 ±223% perf-profile.self.cycles-pp.___slab_alloc 0.05 ± 79% -0.0 0.05 ± 86% perf-profile.self.cycles-pp._raw_spin_trylock 0.03 ±108% -0.0 0.03 ±100% perf-profile.self.cycles-pp.security_ptrace_access_check 0.02 ±144% -0.0 0.02 ±142% perf-profile.self.cycles-pp.inode_to_bdi 0.01 ±223% -0.0 0.01 ±223% perf-profile.self.cycles-pp.rcu_nocb_try_bypass 0.01 ±223% -0.0 0.01 ±223% perf-profile.self.cycles-pp._find_next_bit 0.22 ± 54% -0.0 0.21 ± 35% perf-profile.self.cycles-pp.read_counters 0.04 ±149% -0.0 0.04 ±106% perf-profile.self.cycles-pp.workingset_age_nonresident 0.04 ± 73% -0.0 0.04 ±102% perf-profile.self.cycles-pp.check_move_unevictable_pages 0.13 ± 32% +0.0 0.13 ± 15% perf-profile.self.cycles-pp.part_stat_read_all 0.10 ± 36% +0.0 0.10 ± 16% perf-profile.self.cycles-pp.filemap_map_pages 0.04 ±105% +0.0 0.04 ± 71% perf-profile.self.cycles-pp.vm_area_dup 0.02 ±142% +0.0 0.02 ±142% perf-profile.self.cycles-pp.cfree 0.01 ±223% +0.0 0.01 ±223% perf-profile.self.cycles-pp.mas_next_nentry 0.01 ±223% +0.0 0.01 ±223% perf-profile.self.cycles-pp.shmem_write_begin 0.01 ±223% +0.0 0.01 ±223% perf-profile.self.cycles-pp.switch_fpu_return 0.01 ±223% +0.0 0.01 ±223% perf-profile.self.cycles-pp.__anon_vma_interval_tree_remove 0.01 ±223% +0.0 0.01 ±223% perf-profile.self.cycles-pp.kmem_cache_free_bulk 0.01 ±223% +0.0 0.01 ±223% perf-profile.self.cycles-pp.syscall_enter_from_user_mode 0.14 ± 30% +0.0 0.14 ± 31% perf-profile.self.cycles-pp.strnlen_user 0.01 ±223% +0.0 0.01 ±223% perf-profile.self.cycles-pp.mmap_region 0.04 ±109% +0.0 0.04 ± 73% perf-profile.self.cycles-pp.__legitimize_mnt 0.03 ±106% +0.0 0.04 ± 71% perf-profile.self.cycles-pp.xas_start 0.18 ± 27% +0.0 0.19 ± 15% perf-profile.self.cycles-pp.generic_fillattr 0.02 ±141% +0.0 0.03 ±100% perf-profile.self.cycles-pp.enqueue_entity 0.02 ±141% +0.0 0.02 ±149% perf-profile.self.cycles-pp.ui_progress__update 0.01 ±223% +0.0 0.01 ±223% perf-profile.self.cycles-pp.idle_cpu 0.05 ± 77% +0.0 0.05 ±105% perf-profile.self.cycles-pp.atomic_read 0.05 ± 85% +0.0 0.05 ± 13% perf-profile.self.cycles-pp.perf_read 0.01 ±223% +0.0 0.02 ±141% perf-profile.self.cycles-pp.cap_ptrace_access_check 0.33 ± 33% +0.0 0.33 ± 19% perf-profile.self.cycles-pp.inode_permission 0.20 ± 52% +0.0 0.21 ± 37% perf-profile.self.cycles-pp.__handle_mm_fault 0.10 ± 27% +0.0 0.11 ± 4% perf-profile.self.cycles-pp.malloc 0.08 ± 59% +0.0 0.09 ± 47% perf-profile.self.cycles-pp.__alloc_pages 0.01 ±223% +0.0 0.02 ±223% perf-profile.self.cycles-pp._find_next_and_bit 0.05 ±104% +0.0 0.05 ± 71% perf-profile.self.cycles-pp.rmqueue_bulk 0.03 ±111% +0.0 0.04 ± 70% perf-profile.self.cycles-pp.pick_link 0.06 ± 58% +0.0 0.07 ± 19% perf-profile.self.cycles-pp.find_next_iomem_res 0.03 ±148% +0.0 0.04 ± 72% perf-profile.self.cycles-pp.__get_obj_cgroup_from_memcg 0.01 ±223% +0.0 0.02 ±141% perf-profile.self.cycles-pp.__fsnotify_parent 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.free_percpu 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp._IO_setb 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.strtoll 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.sched_mm_cid_remote_clear 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.set_task_cpu 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.newidle_balance 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.sysvec_apic_timer_interrupt 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.__do_softirq 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.update_process_times 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.__bitmap_and 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.do_anonymous_page 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.down_read_trylock 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.pgd_ctor 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.strnlen 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp._IO_file_open 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.lockref_get_not_dead 0.18 ± 19% +0.0 0.19 ± 11% perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook 0.13 ± 25% +0.0 0.14 ± 13% perf-profile.self.cycles-pp.__task_pid_nr_ns 0.01 ±223% +0.0 0.02 ±142% perf-profile.self.cycles-pp.strncpy_from_user 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.open64 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.cpuidle_idle_call 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.__irq_exit_rcu 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.rcu_sched_clock_irq 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.update_rt_rq_load_avg 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.irqentry_enter_from_user_mode 0.10 ± 42% +0.0 0.11 ± 28% perf-profile.self.cycles-pp.atime_needs_update 0.03 ±101% +0.0 0.04 ± 71% perf-profile.self.cycles-pp.fsnotify_perm 0.01 ±223% +0.0 0.02 ±223% perf-profile.self.cycles-pp.update_irq_load_avg 0.01 ±223% +0.0 0.02 ±141% perf-profile.self.cycles-pp.path_init 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.__hrtimer_run_queues 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.tsc_verify_tsc_adjust 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.irqtime_account_process_tick 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.tcp_orphan_count_sum 0.09 ± 86% +0.0 0.10 ± 58% perf-profile.self.cycles-pp.evlist_cpu_iterator__next 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.lapic_next_deadline 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.read_tsc 0.00 +0.0 0.01 ±223% perf-profile.self.cycles-pp.arch_scale_freq_tick 0.02 ±142% +0.0 0.03 ±103% perf-profile.self.cycles-pp.smpboot_thread_fn 0.03 ±103% +0.0 0.05 ± 74% perf-profile.self.cycles-pp.link_path_walk 0.08 ± 50% +0.0 0.09 ± 31% perf-profile.self.cycles-pp.__call_rcu_common 0.10 ± 17% +0.0 0.11 ± 19% perf-profile.self.cycles-pp.perf_iterate_sb 0.00 +0.0 0.02 ±223% perf-profile.self.cycles-pp.tick_sched_do_timer 0.00 +0.0 0.02 ±223% perf-profile.self.cycles-pp.run_posix_cpu_timers 0.01 ±223% +0.0 0.03 ±100% perf-profile.self.cycles-pp.get_pid_task 0.09 ± 30% +0.0 0.10 ± 16% perf-profile.self.cycles-pp.next_tgid 0.00 +0.0 0.02 ±141% perf-profile.self.cycles-pp._IO_link_in 0.00 +0.0 0.02 ±223% perf-profile.self.cycles-pp.rcu_pending 0.00 +0.0 0.02 ±223% perf-profile.self.cycles-pp.get_cpu_device 0.05 ±117% +0.0 0.07 ± 19% perf-profile.self.cycles-pp.__libc_calloc 0.12 ± 48% +0.0 0.14 ± 71% perf-profile.self.cycles-pp.record__pushfn 0.10 ± 45% +0.0 0.12 ± 30% perf-profile.self.cycles-pp.clear_page_erms 0.04 ±100% +0.0 0.06 ± 50% perf-profile.self.cycles-pp.__entry_text_start 0.03 ±150% +0.0 0.05 ± 49% perf-profile.self.cycles-pp.__fput 0.01 ±223% +0.0 0.03 ±102% perf-profile.self.cycles-pp.security_file_open 0.00 +0.0 0.02 ±142% perf-profile.self.cycles-pp.vprintk_emit 0.00 +0.0 0.02 ±223% perf-profile.self.cycles-pp.timekeeping_max_deferment 0.00 +0.0 0.02 ±142% perf-profile.self.cycles-pp.strscpy 0.02 ±141% +0.0 0.04 ± 71% perf-profile.self.cycles-pp.perf_trace_sched_wakeup_template 0.03 ±158% +0.0 0.05 ± 75% perf-profile.self.cycles-pp.ktime_get_coarse_real_ts64 0.04 ±112% +0.0 0.06 ± 77% perf-profile.self.cycles-pp.kcpustat_cpu_fetch 0.00 +0.0 0.02 ±223% perf-profile.self.cycles-pp.rebalance_domains 0.00 +0.0 0.02 ±223% perf-profile.self.cycles-pp.scheduler_tick 0.04 ± 77% +0.0 0.06 ± 23% perf-profile.self.cycles-pp.aa_may_ptrace 0.10 ± 34% +0.0 0.12 ± 21% perf-profile.self.cycles-pp.__ptrace_may_access 0.01 ±223% +0.0 0.03 ±105% perf-profile.self.cycles-pp._IO_file_doallocate 0.06 ± 46% +0.0 0.08 ± 23% perf-profile.self.cycles-pp.getenv 0.06 ± 60% +0.0 0.08 ± 27% perf-profile.self.cycles-pp.__memcg_kmem_charge_page 0.03 ±101% +0.0 0.05 ± 47% perf-profile.self.cycles-pp.__kmem_cache_free 0.35 ± 34% +0.0 0.37 ± 23% perf-profile.self.cycles-pp.pid_revalidate 0.01 ±223% +0.0 0.04 ± 71% perf-profile.self.cycles-pp.cpuacct_charge 0.07 ±104% +0.0 0.09 ± 32% perf-profile.self.cycles-pp.ordered_events__queue 0.09 ± 34% +0.0 0.12 ± 9% perf-profile.self.cycles-pp._raw_spin_lock_irq 1.28 ± 36% +0.0 1.30 ± 22% perf-profile.self.cycles-pp.update_sg_wakeup_stats 0.04 ±110% +0.0 0.07 ± 57% perf-profile.self.cycles-pp.rmqueue 0.09 ± 31% +0.0 0.12 ± 17% perf-profile.self.cycles-pp.down_write 0.07 ± 97% +0.0 0.10 ± 30% perf-profile.self.cycles-pp.apparmor_file_permission 0.04 ± 71% +0.0 0.06 ± 19% perf-profile.self.cycles-pp.put_dec_trunc8 0.00 +0.0 0.03 ±223% perf-profile.self.cycles-pp.tick_nohz_next_event 0.30 ± 30% +0.0 0.33 ± 19% perf-profile.self.cycles-pp.kmem_cache_free 0.06 ± 50% +0.0 0.10 ±143% perf-profile.self.cycles-pp.__intel_pmu_enable_all 0.07 ± 83% +0.0 0.10 ±170% perf-profile.self.cycles-pp.ktime_get 0.00 +0.0 0.04 ± 71% perf-profile.self.cycles-pp.memset@plt 0.00 +0.0 0.04 ± 71% perf-profile.self.cycles-pp.check_heap_object 0.00 +0.0 0.04 ± 72% perf-profile.self.cycles-pp.ktime_get_update_offsets_now 0.10 ± 35% +0.0 0.14 ± 28% perf-profile.self.cycles-pp.file_free_rcu 0.29 ± 60% +0.0 0.33 ± 43% perf-profile.self.cycles-pp.shmem_write_end 0.01 ±223% +0.0 0.05 ± 72% perf-profile.self.cycles-pp.exc_page_fault 0.00 +0.0 0.04 ±223% perf-profile.self.cycles-pp.__update_blocked_fair 0.15 ± 29% +0.0 0.19 ± 20% perf-profile.self.cycles-pp.__schedule 0.00 +0.0 0.04 ±105% perf-profile.self.cycles-pp.__snprintf_chk 0.00 +0.0 0.04 ± 73% perf-profile.self.cycles-pp.mm_account_fault 0.00 +0.0 0.04 ±128% perf-profile.self.cycles-pp.load_balance 0.00 +0.0 0.04 ± 74% perf-profile.self.cycles-pp.memcg_account_kmem 0.00 +0.0 0.04 ±171% perf-profile.self.cycles-pp.update_sd_lb_stats 0.05 ± 78% +0.0 0.10 ± 83% perf-profile.self.cycles-pp.evsel__read_counter 0.23 ± 34% +0.1 0.28 ± 17% perf-profile.self.cycles-pp.__slab_free 0.26 ± 29% +0.1 0.32 ± 17% perf-profile.self.cycles-pp.do_dentry_open 0.00 +0.1 0.05 ±223% perf-profile.self.cycles-pp.menu_select 0.00 +0.1 0.05 ± 8% perf-profile.self.cycles-pp.copyout 0.00 +0.1 0.05 ± 84% perf-profile.self.cycles-pp.mm_init 0.00 +0.1 0.06 ± 61% perf-profile.self.cycles-pp.get_cpu_sleep_time_us 0.18 ± 27% +0.1 0.24 ± 6% perf-profile.self.cycles-pp.rcu_cblist_dequeue 0.13 ± 44% +0.1 0.20 ± 28% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm 1.38 ± 27% +0.1 1.45 ± 16% perf-profile.self.cycles-pp.do_task_stat 0.00 +0.1 0.07 ± 23% perf-profile.self.cycles-pp._IO_fread 0.56 ± 56% +0.1 0.63 ± 67% perf-profile.self.cycles-pp.fault_in_readable 0.01 ±223% +0.1 0.09 ± 20% perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk 0.01 ±223% +0.1 0.09 ± 13% perf-profile.self.cycles-pp.mas_wr_node_store 0.64 ± 39% +0.1 0.72 ± 19% perf-profile.self.cycles-pp.__percpu_counter_sum 0.14 ±110% +0.1 0.23 ±129% perf-profile.self.cycles-pp.__pthread_disable_asynccancel 0.00 +0.1 0.09 ± 27% perf-profile.self.cycles-pp.update_rq_clock_task 0.04 ±168% +0.1 0.13 ±169% perf-profile.self.cycles-pp.find_busiest_queue 0.45 ± 41% +0.1 0.55 ± 22% perf-profile.self.cycles-pp.pcpu_alloc 0.33 ± 70% +0.1 0.45 ± 85% perf-profile.self.cycles-pp.ring_buffer_read_head 0.12 ± 53% +0.1 0.24 ± 63% perf-profile.self.cycles-pp.evlist__id2evsel 0.00 +0.1 0.14 ±223% perf-profile.self.cycles-pp.cpuidle_enter_state 0.00 +0.2 0.22 ± 39% perf-profile.self.cycles-pp.mm_init_cid 0.00 +0.3 0.25 ± 29% perf-profile.self.cycles-pp.mtree_range_walk 0.33 ± 39% +0.3 0.65 ± 23% perf-profile.self.cycles-pp.copy_mc_enhanced_fast_string 0.03 ±223% +0.4 0.40 ±198% perf-profile.self.cycles-pp.intel_idle 0.00 +0.4 0.45 ± 13% perf-profile.self.cycles-pp.__memcpy 0.00 +0.6 0.60 ± 23% perf-profile.self.cycles-pp.__memset 0.02 ±223% +0.6 0.66 ±200% perf-profile.self.cycles-pp.intel_idle_xstate 0.52 ± 72% +1.0 1.51 ±125% perf-profile.self.cycles-pp.update_sg_lb_stats 0.00 +1.1 1.11 ± 21% perf-profile.self.cycles-pp.task_mm_cid_work 0.00 +2.0 2.04 ± 58% perf-profile.self.cycles-pp.copyin 37.64 ± 48% +11.6 49.20 ± 10% perf-profile.self.cycles-pp.queue_event > > Thanks > - Raghu > > >
On 7/16/2023 7:47 PM, Oliver Sang wrote: > hi, Raghavendra K T, > > On Wed, Jul 05, 2023 at 11:18:37AM +0530, Raghavendra K T wrote: >> On 5/31/2023 9:55 AM, Raghavendra K T wrote: >>> With the numa scan enhancements [1], only the threads which had previously >>> accessed vma are allowed to scan. >>> >>> While this had improved significant system time overhead, there were corner >>> cases, which genuinely need some relaxation. For e.g., >>> >>> 1) Concern raised by PeterZ, where if there are N partition sets of vmas >>> belonging to tasks, then unfairness in allowing these threads to scan could >>> potentially amplify the side effect of some of the vmas being left >>> unscanned. >>> >>> 2) Below reports of LKP numa01 benchmark regression. >>> >>> Currently this is handled by allowing first two scanning unconditional >>> as indicated by mm->numa_scan_seq. This is imprecise since for some >>> benchmark vma scanning might itself start at numa_scan_seq > 2. >>> >>> Solution: >>> Allow unconditional scanning of vmas of tasks depending on vma size. This >>> is achieved by maintaining a per vma scan counter, where >>> >>> f(allowed_to_scan) = f(scan_counter < vma_size / scan_size) >>> >>> Result: >>> numa01_THREAD_ALLOC result on 6.4.0-rc2 (that has numascan enhancement) >>> base-numascan base base+fix >>> real 1m1.507s 1m23.259s 1m2.632s >>> user 213m51.336s 251m46.363s 220m35.528s >>> sys 3m3.397s 0m12.492s 2m41.393s >>> >>> numa_hit 5615517 4560123 4963875 >>> numa_local 5615505 4560024 4963700 >>> numa_other 12 99 175 >>> numa_pte_updates 1822797 493 1559111 >>> numa_hint_faults 1307113 523 1469031 >>> numa_hint_faults_local 612617 488 884829 >>> numa_pages_migrated 694370 35 584202 >>> >>> Summary: Regression in base is recovered by allowing scanning as required. >>> >>> [1] https://lore.kernel.org/lkml/cover.1677672277.git.raghavendra.kt@amd.com/T/#t >>> >>> Fixes: fc137c0ddab2 ("sched/numa: enhance vma scanning logic") >>> regression. >>> Reported-by: Aithal Srikanth <sraithal@amd.com> >>> Reported-by: kernel test robot <oliver.sang@intel.com> >>> Closes: https://lore.kernel.org/lkml/db995c11-08ba-9abf-812f-01407f70a5d4@amd.com/T/ >>> Signed-off-by: Raghavendra K T <raghavendra.kt@amd.com> >> >> Hello kernel test robot, >> >> Gentle ping to check if the patch has helped your regression report. > > sorry for late. > > previously we found a 118.9% regression of autonuma-benchmark.numa01.seconds > on a Cascade Lake, which happened to be converted for other test purposes, so > we cannot test your patch on it again. > > however, we also found a 39.3% regression on a Sapphire Rapids test machine: > > ========================================================================================= > compiler/cpufreq_governor/iterations/kconfig/rootfs/tbox_group/test/testcase: > gcc-11/performance/4x/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-spr-r02/numa02_SMT/autonuma-benchmark > > ef6a22b70f6d9044 fc137c0ddab29b591db6a091dc6 > ---------------- --------------------------- > %stddev %change %stddev > \ | \ > 193.14 +39.2% 268.84 autonuma-benchmark.numa01.seconds > 8.14 -0.7% 8.09 autonuma-benchmark.numa02.seconds > > > now we tested your v3 patch on it, found regression mostly recovered > (55fd15913b18d6a790c17d947df is just [RFC PATCH V3 1/1] sched/numa: Fix disjoint set vma scan regression) > > ========================================================================================= > compiler/cpufreq_governor/iterations/kconfig/rootfs/tbox_group/test/testcase: > gcc-11/performance/4x/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-spr-r02/numa01_THREAD_ALLOC/autonuma-benchmark > > ef6a22b70f6d9044 55fd15913b18d6a790c17d947df > ---------------- --------------------------- > %stddev %change %stddev > \ | \ > 193.14 +5.8% 204.37 ± 3% autonuma-benchmark.numa01.seconds > 8.14 -0.9% 8.06 autonuma-benchmark.numa02.seconds > > detail comparison as below: > Thank you for the confirmation. So looks like we got back most of the regression with the patch for numa01_THREAD_ALLOC case. Andrew, could you please help by picking up this patch unless Mel , PeterZ do not have any concern about the patch / direction. (as we note it brings back little bit system time overhead by allowing some scanning..) Thanks and Regards - Raghu
On Wed, May 31, 2023 at 09:55:26AM +0530, Raghavendra K T wrote: > With the numa scan enhancements [1], only the threads which had previously > accessed vma are allowed to scan. > I know it's very late to be reading this properly for the first time. It's a brief review as it's late but I'm not 100% convinced by the patch as-is. To start off -- reference the specific commit e.g. Since commit fc137c0ddab2 ("sched/numa: enhance vma scanning logic")... > While this had improved significant system time overhead, there were corner > cases, which genuinely need some relaxation. For e.g., > > 1) Concern raised by PeterZ, where if there are N partition sets of vmas > belonging to tasks, then unfairness in allowing these threads to scan could > potentially amplify the side effect of some of the vmas being left > unscanned. > This needs a much better description as there is too much guesswork required to figure out what problem is being solved. You may be referring to an issue where a VMA not accessed for a long time may be skipped for scanning indefinitely. You could also be referring to an issue where a highly active thread always traps the NUMA hinting fault first and other tasks never pass the vma_is_accessed test. It also could be the case that due to a malloc implementation behaviour that the thread using a particular VMA changes and it's not detected. It's simply not clear how the VMAs are partitioned or what exactly is unfair. I certainly can guess a number of reasons why a patch like this is necessary but in a years time, it'll be hard to be certain about the original intent. > 2) Below reports of LKP numa01 benchmark regression. > > Currently this is handled by allowing first two scanning unconditional > as indicated by mm->numa_scan_seq. This is imprecise since for some > benchmark vma scanning might itself start at numa_scan_seq > 2. > Well, it's also not useful in many cases. There is nothing special about the first two scans other than they happen early in the lifetime of the process. A major change in phase behaviour that occurs after 2 scans could miss the new access pattern. > Solution: > Allow unconditional scanning of vmas of tasks depending on vma size. This > is achieved by maintaining a per vma scan counter, where > > f(allowed_to_scan) = f(scan_counter < vma_size / scan_size) > This is a vague description as well and does not mention that the scan_counter resets and the code does not appear to match the equation. It's also a bit arbitrary that unconditional scanning occurs for every task after the scan_count passes a threshold. The timing of when this occurs may vary and all threads sharing the address space may conduct the scanning which may be overkill. It should only be necessary for at least one task to unconditionally mark the VMA for hinting, no? I'm generally a bit wary that the time it takes to detect a VMA needs unconditonal scanning depends on the size as due to VMA-merging, it may be non-deterministic when a VMA gets scanned and there may be run-to-run variance to the timing of threads allocating address space. > Result: > numa01_THREAD_ALLOC result on 6.4.0-rc2 (that has numascan enhancement) > base-numascan base base+fix > real 1m1.507s 1m23.259s 1m2.632s > user 213m51.336s 251m46.363s 220m35.528s > sys 3m3.397s 0m12.492s 2m41.393s > > numa_hit 5615517 4560123 4963875 > numa_local 5615505 4560024 4963700 > numa_other 12 99 175 > numa_pte_updates 1822797 493 1559111 > numa_hint_faults 1307113 523 1469031 > numa_hint_faults_local 612617 488 884829 > numa_pages_migrated 694370 35 584202 > It's not clear what these kernels are. Is base-numascan a 6.4-rc2 kernel with at least commit fc137c0ddab2 ("sched/numa: enhance vma scanning logic") reverted? The numa_[hint|local|other] stats are not helpful. They are updated on the allocation side and not related to actual behaviour. As I'm unsure what the kernels are, I'm not sure how interesting it is that the pte update stats and number of faults trapped are very different for base to base+fix. Sure enough, the overall runtime is lower so it's *probably* good but not sure as it may be also the case that disabling NUMA balancing would be faster again given the nature of this particular test :P The testing is a little thin and I'd worry slightly that this patch is very specific to this particular workload. > Summary: Regression in base is recovered by allowing scanning as required. > > [1] https://lore.kernel.org/lkml/cover.1677672277.git.raghavendra.kt@amd.com/T/#t > > Fixes: fc137c0ddab2 ("sched/numa: enhance vma scanning logic") > regression. Fixes may be a bit overkill as this patch is more of an enhancement than something that justifies a backport to -stable but I don't feel strongly about it. > Reported-by: Aithal Srikanth <sraithal@amd.com> > Reported-by: kernel test robot <oliver.sang@intel.com> > Closes: https://lore.kernel.org/lkml/db995c11-08ba-9abf-812f-01407f70a5d4@amd.com/T/ > Signed-off-by: Raghavendra K T <raghavendra.kt@amd.com> > --- > include/linux/mm_types.h | 1 + > kernel/sched/fair.c | 31 ++++++++++++++++++++++++------- > 2 files changed, 25 insertions(+), 7 deletions(-) > > diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h > index 306a3d1a0fa6..992e460a713e 100644 > --- a/include/linux/mm_types.h > +++ b/include/linux/mm_types.h > @@ -479,6 +479,7 @@ struct vma_numab_state { > unsigned long next_scan; > unsigned long next_pid_reset; > unsigned long access_pids[2]; > + unsigned int scan_counter; > }; Vague name as it's not counting and gets reset. It might have been clearer to name is something like scan_selective with an initial value related to the VMA size that decrements. When it hits 0, the scan is forced *once* for the unlucky task. The suggested name is also bad, I'm terrible at naming but "counter" gives no hints either. > > /* > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > index 373ff5f55884..4e71fb58085b 100644 > --- a/kernel/sched/fair.c > +++ b/kernel/sched/fair.c > @@ -2931,17 +2931,30 @@ static void reset_ptenuma_scan(struct task_struct *p) > static bool vma_is_accessed(struct vm_area_struct *vma) > { > unsigned long pids; > + unsigned int vma_size; > + unsigned int scan_threshold; > + unsigned int scan_size; > + > + pids = vma->numab_state->access_pids[0] | vma->numab_state->access_pids[1]; > + > + if (test_bit(hash_32(current->pid, ilog2(BITS_PER_LONG)), &pids)) > + return true; > + > + scan_size = READ_ONCE(sysctl_numa_balancing_scan_size); > + /* vma size in MB */ > + vma_size = (vma->vm_end - vma->vm_start) >> 20; > + > + /* Total scans needed to cover VMA */ > + scan_threshold = vma_size / scan_size; > + > /* > - * Allow unconditional access first two times, so that all the (pages) > - * of VMAs get prot_none fault introduced irrespective of accesses. > + * Allow the scanning of half of disjoint set's VMA to induce > + * prot_none fault irrespective of accesses. > * This is also done to avoid any side effect of task scanning > * amplifying the unfairness of disjoint set of VMAs' access. > */ Much better description is required. It's not stated how the VMAs are disjoint. For example, they all belong to the same address space so from that perspective they are not disjoint. Clarify that it's related to accesses at the very least or be clear on how a set of VMAs can be disjoint. Also, it's not half the total VMAs or some other critieria either, it's related to scanning events. There is a lot of guesswork involved here. > - if (READ_ONCE(current->mm->numa_scan_seq) < 2) > - return true; > - > - pids = vma->numab_state->access_pids[0] | vma->numab_state->access_pids[1]; > - return test_bit(hash_32(current->pid, ilog2(BITS_PER_LONG)), &pids); > + scan_threshold = 1 + (scan_threshold >> 1); > + return (vma->numab_state->scan_counter < scan_threshold); > } > > #define VMA_PID_RESET_PERIOD (4 * sysctl_numa_balancing_scan_delay) > @@ -3058,6 +3071,8 @@ static void task_numa_work(struct callback_head *work) > /* Reset happens after 4 times scan delay of scan start */ > vma->numab_state->next_pid_reset = vma->numab_state->next_scan + > msecs_to_jiffies(VMA_PID_RESET_PERIOD); > + > + vma->numab_state->scan_counter = 0; > } > > /* > @@ -3084,6 +3099,8 @@ static void task_numa_work(struct callback_head *work) > vma->numab_state->access_pids[1] = 0; > } > > + vma->numab_state->scan_counter++; > + > do { > start = max(start, vma->vm_start); > end = ALIGN(start + (pages << PAGE_SHIFT), HPAGE_SIZE); This has a potential timing issue as only tasks that accessed the VMA update the counter. A task could create a VMA, access is heavily in an init phase and then go completely idle waiting on other tasks to complete (e.g. parallelised load that uses a single thread to setup the data and then creates worker threads to process that data). As there is no guarantee other tasks would trap a fault, the access bits would never be set, the counter never increments and the VMA is ignored forever. It's late on a Friday so maybe I'm wrong but it looks like this could have weird corner cases. The counter should at least increase if the VMA could have been scanned except the access bits were not set with the caveat that it may incur excessive cache misses.
On 7/21/2023 8:48 PM, Mel Gorman wrote: > On Wed, May 31, 2023 at 09:55:26AM +0530, Raghavendra K T wrote: >> With the numa scan enhancements [1], only the threads which had previously >> accessed vma are allowed to scan. >> > > I know it's very late to be reading this properly for the first time. It's a > brief review as it's late but I'm not 100% convinced by the patch as-is. > Thank you Mel for the review. Expert opinion / intuition from you, PeterZ helped, helps alot to improve to solve the problem in better way finally. > To start off -- reference the specific commit e.g. Since commit fc137c0ddab2 > ("sched/numa: enhance vma scanning logic")... > Okay. >> While this had improved significant system time overhead, there were corner >> cases, which genuinely need some relaxation. For e.g., >> >> 1) Concern raised by PeterZ, where if there are N partition sets of vmas >> belonging to tasks, then unfairness in allowing these threads to scan could >> potentially amplify the side effect of some of the vmas being left >> unscanned. >> > > This needs a much better description as there is too much guesswork > required to figure out what problem is being solved. Agree I needed to be more descriptive. You may be referring > to an issue where a VMA not accessed for a long time may be skipped for > scanning indefinitely. I feel this is mostly a favorable case because we don't incur overhead of scanning unneeded VMA. so not this, but below ones You could also be referring to an issue where a > highly active thread always traps the NUMA hinting fault first and other > tasks never pass the vma_is_accessed test. It also could be the case that > due to a malloc implementation behaviour that the thread using a particular > VMA changes and it's not detected. It's simply not clear how the VMAs are > partitioned or what exactly is unfair. I certainly can guess a number of > reasons why a patch like this is necessary but in a years time, it'll be > hard to be certain about the original intent. Agree, to summarize here there are multiple layers of issues 1) PeterZ pointed this (Thread1/Set1)---->access(vma1) <== gets 65% opportunity (Thread2/Set2)---->access(vma2) <== gets 35% opprtunity Suppose Thread1 (or threads from Set1) are actively getting more scanning opportunity, we may miss or scan less of vma2. I did some experiment and found there was indeed a bit of unfairness in the way set1/set2 gets opportunity to scan. Link: https://lore.kernel.org/lkml/c730dee0-a711-8a8e-3eb1-1bfdd21e6add@amd.com/ 2) Suppose for the above case if we have vma1, vma2 represents sum of vmas accessed by set1 and set2 respectively in disjoint way. and vma1 size is far less than vma2 size, effect of unfairness in scanning gets amplified. e.g., (Thread1/Set1)---->access(vma1 size = 1MB) <== actively scanning (Thread2/Set2)---->access(vma2 size = 10MB) 3) malloc behavior you pointed. I was trying to minimize the side effect of unfairness in scanning overall by introducing a bit of unconditional scans. (But the unfairness amongst the threads, which thread (or set of thread) gets opportunity to scan remains) I may be missing some more cases here > >> 2) Below reports of LKP numa01 benchmark regression. >> >> Currently this is handled by allowing first two scanning unconditional >> as indicated by mm->numa_scan_seq. This is imprecise since for some >> benchmark vma scanning might itself start at numa_scan_seq > 2. >> > > Well, it's also not useful in many cases. There is nothing special about > the first two scans other than they happen early in the lifetime of the > process. A major change in phase behaviour that occurs after 2 scans > could miss the new access pattern. Fully agree. That is why that condition is removed in the patch now. What I also saw was some times when new vma gets created mm->numa_scan_seq could be already more than 2. (To handle that correctly I would have to have per vma start_numa_seq kind of things etc which was messy and more space consuming) Note: Somewhat related, In my experiments what I saw is normal short running benchmarks used to finish by numa_scan_seq < 16 whereas some benchmark like NAS used to go till 100-150. (to set context for any long running benchmark optimization in future) > >> Solution: >> Allow unconditional scanning of vmas of tasks depending on vma size. This >> is achieved by maintaining a per vma scan counter, where >> >> f(allowed_to_scan) = f(scan_counter < vma_size / scan_size) >> > > This is a vague description as well and does not mention that the > scan_counter resets and the code does not appear to match the equation. > It's also a bit arbitrary that unconditional scanning occurs for every task > after the scan_count passes a threshold. The timing of when this occurs may > vary and all threads sharing the address space may conduct the scanning > which may be overkill. It should only be necessary for at least one task > to unconditionally mark the VMA for hinting, no? I'm generally a bit wary > that the time it takes to detect a VMA needs unconditonal scanning depends > on the size as due to VMA-merging, I do agree that changing of vma size due to merge/split can affect scanning. But I was assuming that it is Okay because this calculation is done every time we hit task_numa_work which may be wrong (and imprecise). it may be non-deterministic when a VMA > gets scanned and there may be run-to-run variance to the timing of threads > allocating address space > Agree >> Result: >> numa01_THREAD_ALLOC result on 6.4.0-rc2 (that has numascan enhancement) >> base-numascan base base+fix >> real 1m1.507s 1m23.259s 1m2.632s >> user 213m51.336s 251m46.363s 220m35.528s >> sys 3m3.397s 0m12.492s 2m41.393s >> >> numa_hit 5615517 4560123 4963875 >> numa_local 5615505 4560024 4963700 >> numa_other 12 99 175 >> numa_pte_updates 1822797 493 1559111 >> numa_hint_faults 1307113 523 1469031 >> numa_hint_faults_local 612617 488 884829 >> numa_pages_migrated 694370 35 584202 >> > > It's not clear what these kernels are. Is base-numascan a 6.4-rc2 kernel > with at least commit fc137c0ddab2 ("sched/numa: enhance vma scanning > logic") reverted? This is correct. To compare the system time overhead before commit fc137c0ddab2 ("sched/numa: enhance vma scanning logic") > The numa_[hint|local|other] stats are not helpful. They are updated on > the allocation side and not related to actual behaviour. will trim this next time. > > As I'm unsure what the kernels are, I'm not sure how interesting it is > that the pte update stats and number of faults trapped are very different > for base to base+fix. Number of faults trapped are excessively refined for base (6.4.0-rc2 vanilla) for this particular case, because of malloc behaviour you pointed above. Sure enough, the overall runtime is lower so it's > *probably* good but not sure as it may be also the case that disabling NUMA > balancing would be faster again given the nature of this particular test :P > > The testing is a little thin and I'd worry slightly that this patch is > very specific to this particular workload. I do agree that it is mostly trying to address LKP numa01_THREAD_ALLOC case. But regarding the testing though I reported only above result, I do run 1) kernbench, autonuma from mmtest, 2) some microbenchmarks and also NAS/hashjoin etc long running benchmarks to ensure there are no surprises. > >> Summary: Regression in base is recovered by allowing scanning as required. >> >> [1] https://lore.kernel.org/lkml/cover.1677672277.git.raghavendra.kt@amd.com/T/#t >> >> Fixes: fc137c0ddab2 ("sched/numa: enhance vma scanning logic") >> regression. > > Fixes may be a bit overkill as this patch is more of an enhancement than > something that justifies a backport to -stable but I don't feel strongly > about it. Sure. > >> Reported-by: Aithal Srikanth <sraithal@amd.com> >> Reported-by: kernel test robot <oliver.sang@intel.com> >> Closes: https://lore.kernel.org/lkml/db995c11-08ba-9abf-812f-01407f70a5d4@amd.com/T/ >> Signed-off-by: Raghavendra K T <raghavendra.kt@amd.com> >> --- >> include/linux/mm_types.h | 1 + >> kernel/sched/fair.c | 31 ++++++++++++++++++++++++------- >> 2 files changed, 25 insertions(+), 7 deletions(-) >> >> diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h >> index 306a3d1a0fa6..992e460a713e 100644 >> --- a/include/linux/mm_types.h >> +++ b/include/linux/mm_types.h >> @@ -479,6 +479,7 @@ struct vma_numab_state { >> unsigned long next_scan; >> unsigned long next_pid_reset; >> unsigned long access_pids[2]; >> + unsigned int scan_counter; >> }; > > Vague name as it's not counting and gets reset. It might have been clearer > to name is something like scan_selective with an initial value related to > the VMA size that decrements. When it hits 0, the scan is forced *once* > for the unlucky task. The suggested name is also bad, I'm terrible at > naming but "counter" gives no hints either. > Agree that naming is bad, will use scan_selective / or try if I get more idea on naming. So you are suggesting to try other-way where we force scan once we hit zero. Looks fair to me. will try the idea and see how it goes. >> >> /* >> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c >> index 373ff5f55884..4e71fb58085b 100644 >> --- a/kernel/sched/fair.c >> +++ b/kernel/sched/fair.c >> @@ -2931,17 +2931,30 @@ static void reset_ptenuma_scan(struct task_struct *p) >> static bool vma_is_accessed(struct vm_area_struct *vma) >> { >> unsigned long pids; >> + unsigned int vma_size; >> + unsigned int scan_threshold; >> + unsigned int scan_size; >> + >> + pids = vma->numab_state->access_pids[0] | vma->numab_state->access_pids[1]; >> + >> + if (test_bit(hash_32(current->pid, ilog2(BITS_PER_LONG)), &pids)) >> + return true; >> + >> + scan_size = READ_ONCE(sysctl_numa_balancing_scan_size); >> + /* vma size in MB */ >> + vma_size = (vma->vm_end - vma->vm_start) >> 20; >> + >> + /* Total scans needed to cover VMA */ >> + scan_threshold = vma_size / scan_size; >> + >> /* >> - * Allow unconditional access first two times, so that all the (pages) >> - * of VMAs get prot_none fault introduced irrespective of accesses. >> + * Allow the scanning of half of disjoint set's VMA to induce >> + * prot_none fault irrespective of accesses. >> * This is also done to avoid any side effect of task scanning >> * amplifying the unfairness of disjoint set of VMAs' access. >> */ > > Much better description is required. It's not stated how the VMAs are > disjoint. For example, they all belong to the same address space so from that > perspective they are not disjoint. Clarify that it's related to accesses > at the very least or be clear on how a set of VMAs can be disjoint. Also, > it's not half the total VMAs or some other critieria either, it's related > to scanning events. There is a lot of guesswork involved here. Sure. will add more details. ALso I was not clear here where does that half came from. Idea was allowing unconditional scanning = vma_size for a unrelated task that has not accessed vma is too much.. since we are letting unconditional scans let it be reduced to atleast half of that. I will experiment a bit here to get optimal f(vma_size/scan_size) > >> - if (READ_ONCE(current->mm->numa_scan_seq) < 2) >> - return true; >> - >> - pids = vma->numab_state->access_pids[0] | vma->numab_state->access_pids[1]; >> - return test_bit(hash_32(current->pid, ilog2(BITS_PER_LONG)), &pids); >> + scan_threshold = 1 + (scan_threshold >> 1); >> + return (vma->numab_state->scan_counter < scan_threshold); >> } >> >> #define VMA_PID_RESET_PERIOD (4 * sysctl_numa_balancing_scan_delay) >> @@ -3058,6 +3071,8 @@ static void task_numa_work(struct callback_head *work) >> /* Reset happens after 4 times scan delay of scan start */ >> vma->numab_state->next_pid_reset = vma->numab_state->next_scan + >> msecs_to_jiffies(VMA_PID_RESET_PERIOD); >> + >> + vma->numab_state->scan_counter = 0; >> } >> >> /* >> @@ -3084,6 +3099,8 @@ static void task_numa_work(struct callback_head *work) >> vma->numab_state->access_pids[1] = 0; >> } >> >> + vma->numab_state->scan_counter++; >> + >> do { >> start = max(start, vma->vm_start); >> end = ALIGN(start + (pages << PAGE_SHIFT), HPAGE_SIZE); > > This has a potential timing issue as only tasks that accessed the VMA > update the counter. I think there is a confusion because of reverted logic. we do allow unconditionally to let the scan happen initially till scan_counter reaches scan_threshold. So even though task did not access vma, we get to scan initially. So idea is update scan_counter only if we are able to successfully come till here because we are sure that there is not condition that returns back after this without scanning. A task could create a VMA, access is heavily in an > init phase and then go completely idle waiting on other tasks to complete > (e.g. parallelised load that uses a single thread to setup the data and then > creates worker threads to process that data). As there is no guarantee other > tasks would trap a fault, the access bits would never be set, the counter > never increments and the VMA is ignored forever. It's late on a Friday so > maybe I'm wrong but it looks like this could have weird corner cases. > > The counter should at least increase if the VMA could have been scanned > except the access bits were not set with the caveat that it may incur > excessive cache misses. > Explained above Thanks a lot, Raghu
On 7/24/2023 1:11 PM, Raghavendra K T wrote: > On 7/21/2023 8:48 PM, Mel Gorman wrote: >> On Wed, May 31, 2023 at 09:55:26AM +0530, Raghavendra K T wrote: >>> With the numa scan enhancements [1], only the threads which had >>> previously >>> accessed vma are allowed to scan. >>> >> >> I know it's very late to be reading this properly for the first time. >> It's a >> brief review as it's late but I'm not 100% convinced by the patch as-is. >> > > Thank you Mel for the review. Expert opinion / intuition from you, > PeterZ helped, helps alot to improve to solve the problem in better way > finally. [...] >>> --- >>> include/linux/mm_types.h | 1 + >>> kernel/sched/fair.c | 31 ++++++++++++++++++++++++------- >>> 2 files changed, 25 insertions(+), 7 deletions(-) >>> >>> diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h >>> index 306a3d1a0fa6..992e460a713e 100644 >>> --- a/include/linux/mm_types.h >>> +++ b/include/linux/mm_types.h >>> @@ -479,6 +479,7 @@ struct vma_numab_state { >>> unsigned long next_scan; >>> unsigned long next_pid_reset; >>> unsigned long access_pids[2]; >>> + unsigned int scan_counter; >>> }; >> >> Vague name as it's not counting and gets reset. It might have been >> clearer >> to name is something like scan_selective with an initial value related to >> the VMA size that decrements. When it hits 0, the scan is forced *once* >> for the unlucky task. The suggested name is also bad, I'm terrible at >> naming but "counter" gives no hints either. >> > > Agree that naming is bad, will use scan_selective / or try if I get more > idea on naming. > So you are suggesting to try other-way where we force scan once we hit > zero. Looks fair to me. will try the idea and see how it goes. > Hello Mel, Sorry for going slow from my side. I was experimenting further with combination patches, I will be posting set of patches that I mention further down here. Just want us to be on same page, posting full description of the problem, details of solution, and how it looks like (code) etc Also sorry for long description. Problem statement (Disjoint VMA set): ==================== Since commit fc137c0ddab2 ("sched/numa: enhance vma scanning logic") vma scanning is allowed if 1) The task had accessed vma 2) Early phase of the scan where mm->numa_scan_seq is less than 2 while this works for most of the times to reduce scanning overhead there are concerning corner case(s) associated with it. Let's look at these corner caseswith below example of tasks and their access pattern. Set 1 tasks accessing vma_x (group of vmas) Set 2 tasks accessing vma_y (group of vmas) Set1 Set2 ------------------- -------------------- | task_1..task_n/2 | | task_n/2+1..task_n | ------------------- -------------------- | | V V ------------------- -------------------- | vma_x | | vma_y | ------------------- -------------------- Figure. (1) Tasks accessing disjoint vma Corner cases: (a) Out of N tasks, not all of them gets fair opportunity to scan. (PeterZ). suppose Set1 tasks gets more opportunity to scan in the above example, then vma_x gets scanned more number of times than vma_y. some experiment is also done here: Link: https://lore.kernel.org/lkml/c730dee0-a711-8a8e-3eb1-1bfdd21e6add@amd.com/ (b) Sizes of vmas can differ. suppose size of vma_y is far greater than size of vma_x, then bigger portion of vma_y can potentially be left unscanned since scanning is bounded by scan_size (c) highly active threads trap a few vmas frequently, and some of the vmas not accessed for long time can potentially get starved of scanning indefinitely (Mel) (because task access history gets reset periodically. Not sure if this is is a bigger issue if vma is not important but on the other hand we will be left with no hints for migration later if necessary) (d) allocation of memory in some specific manner (Mel). (e) vmas that are created after two full scans of mm (mm->numa_scan_seq > 2), will never get scanned. (I have not seen that occurring in some rare cases) on top of this Combination of some of the above (e.g., (a) and (b)) can potentially amplify the side effect of numa scanning overall. High level ideas to address the issue: =================================== Idea 1) depending on vma_size populate per vma scan_select value, decrement it and when it hits zero do force scan (Mel) this is how vma phases looks like after implementation: |<--p1--->|<-----p2----->|<-----p2----->|... p1: new vma, initial phase do not scan till scan_delay p2: allow scanning if task has accessed vma or vma_scan_selectiv hit zero repeat p2 pros/cons: + : Ratelimiting is inbuilt to the approach +/-: scanning continues forever - : changes in vma size is taken care after force scan Idea 2) Have a per vma_scan_seq. allow the unconditional scan till vma_scan_seq reaches a value proportional (or equal) to vma_size/scan_size this is how vma phases looks like after implementation: |<--p1--->|<-----p2----->|<-----p3----->|<-----p4----->...||<-----p2----->|<-----p3----->|<-----p4-----> ...|| RESET RESET p1: new vma, initial phase do not scan till scan_delay p2: allow scanning if task has accessed vma or vma_scan_seq has reached till f1(vma_size)/scan_size) for e.g., number of scan to cover full or half of vmas p3: allow scanning if task has accessed vma or vma_scan_seq has reached till f2(vma_size)/scan_size in rate limited manner (Optional) p4: allow scanning iff task has accessed vma Reset after p4 ( f(scan_delay) ??) and repeat p2, p3 p4 + : Ratelimiting need to be taken care separately if needed +/-: scanning continues only if RESET of vma_scan_seq is implemented + : changes in vma size is taken care in every scan - : Additional optimizations (unconditionally allow if vma is hot i.e., either accessed very recently or number of tasks accessing vma is very high) Idea 3) Take bitmask_weight of pid_access history of vma. (suggested by Bharata) (vma->numab_state->pid_access[1] | vma->numab_state->pid_access[0]) If number of tasks accessing vma is > THRESHOLD (say 3), unconditionally allow scanning Idea 4) Take bitmask_weight of latest pid_access value (vma->numab_state->pid_access[1]) If number of tasks accessing vma is >= 1, i.e. recently accessed unconditionally allow scanning. Other ideas still to explore: 5) Can we maintain more pid access history (vs last two as maintained now) The whole idea with code has been already given by PeterZ. Some of the observations from my experiments: ----------------------------------- - It is very critical to allow scanning after initial scan_delay (as mentioned by Mel) so as to capture early patterns. - It is also important to limit (or rate limit) unconditional scans to avoid excess scanning overhead. Example code for idea1: ======================= static inline unsigned int vma_scan_ratelimit(struct vm_area_struct *vma) { unsigned int vma_size, ratelimit; /* vma size */ vma_size = (vma->vm_end - vma->vm_start); if (vma_size) ratelimit = (1 << 15) / vma_size; /* for 4KB vma rate limit is 1 in 9 */ /* and for > 64K it is 1 in 2 */ return 1 + ratelimit; } static bool task_disjoint_vma_select(struct vm_area_struct *vma) { if (vma->numab_state->vma_scan_seq > 0) { vma->numab_state->vma_scan_selective--; return false; } else vma->numab_state->vma_scan_selective = vma_scan_ratelimit(vma); } Example code for idea 2: ======================= static inline unsigned int disjoint_vma_scan_threshold(struct vm_area_struct *vma) { unsigned int vma_size, scan_size, threshold; scan_size = READ_ONCE(sysctl_numa_balancing_scan_size); /* vma size in MB */ vma_size = (vma->vm_end - vma->vm_start) >> 20; /* Total scans needed to cover VMA. Add minimum of one */ threshold = 1 + (vma_size / scan_size); return threshold; } static bool task_disjoint_vma_select(struct vm_area_struct *vma) { return (vma->numab_state->vma_scan_seq <= disjoint_vma_scan_threshold(vma)); } Example code for idea 3,4 combined: ================================ #define HOT_VMA_NR_TASK_THRESH 3 static bool vma_is_accessed(struct vm_area_struct *vma) { unsigned long pids, recent_hist; unsigned int hashval; recent_hist = READ_ONCE(vma->numab_state->access_pids[1]); pids = vma->numab_state->access_pids[0] | recent_hist ; /* * Check if vma was recently accessed OR * vma is being accessed by more than 2 tasks */ if (recent_hist || bitmap_wight(&pids, BITS_PER_LONG)) return true; hashval = hash_32(current->pid, ilog2(BITS_PER_LONG)); return test_bit(hashval, &pids); } Additionally, to Implement vma_scan_seq reset have used something like #define VMA_SCANSEQ_RESET_PERIOD (4 * VMA_PID_RESET_PERIOD) /* this is to recent per vma scan_seq */ I will be posting combination of idea1 and idea3, idea4 by next week mostly as a new enhancement series. (OR idea2 and idea3, idea4 in case you have any positive thought about idea2) initial results are looking good for both the ideas, I am taking more time to experiment further and fine tune. Thanks - Raghu
diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h index 306a3d1a0fa6..992e460a713e 100644 --- a/include/linux/mm_types.h +++ b/include/linux/mm_types.h @@ -479,6 +479,7 @@ struct vma_numab_state { unsigned long next_scan; unsigned long next_pid_reset; unsigned long access_pids[2]; + unsigned int scan_counter; }; /* diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 373ff5f55884..4e71fb58085b 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -2931,17 +2931,30 @@ static void reset_ptenuma_scan(struct task_struct *p) static bool vma_is_accessed(struct vm_area_struct *vma) { unsigned long pids; + unsigned int vma_size; + unsigned int scan_threshold; + unsigned int scan_size; + + pids = vma->numab_state->access_pids[0] | vma->numab_state->access_pids[1]; + + if (test_bit(hash_32(current->pid, ilog2(BITS_PER_LONG)), &pids)) + return true; + + scan_size = READ_ONCE(sysctl_numa_balancing_scan_size); + /* vma size in MB */ + vma_size = (vma->vm_end - vma->vm_start) >> 20; + + /* Total scans needed to cover VMA */ + scan_threshold = vma_size / scan_size; + /* - * Allow unconditional access first two times, so that all the (pages) - * of VMAs get prot_none fault introduced irrespective of accesses. + * Allow the scanning of half of disjoint set's VMA to induce + * prot_none fault irrespective of accesses. * This is also done to avoid any side effect of task scanning * amplifying the unfairness of disjoint set of VMAs' access. */ - if (READ_ONCE(current->mm->numa_scan_seq) < 2) - return true; - - pids = vma->numab_state->access_pids[0] | vma->numab_state->access_pids[1]; - return test_bit(hash_32(current->pid, ilog2(BITS_PER_LONG)), &pids); + scan_threshold = 1 + (scan_threshold >> 1); + return (vma->numab_state->scan_counter < scan_threshold); } #define VMA_PID_RESET_PERIOD (4 * sysctl_numa_balancing_scan_delay) @@ -3058,6 +3071,8 @@ static void task_numa_work(struct callback_head *work) /* Reset happens after 4 times scan delay of scan start */ vma->numab_state->next_pid_reset = vma->numab_state->next_scan + msecs_to_jiffies(VMA_PID_RESET_PERIOD); + + vma->numab_state->scan_counter = 0; } /* @@ -3084,6 +3099,8 @@ static void task_numa_work(struct callback_head *work) vma->numab_state->access_pids[1] = 0; } + vma->numab_state->scan_counter++; + do { start = max(start, vma->vm_start); end = ALIGN(start + (pages << PAGE_SHIFT), HPAGE_SIZE);
With the numa scan enhancements [1], only the threads which had previously accessed vma are allowed to scan. While this had improved significant system time overhead, there were corner cases, which genuinely need some relaxation. For e.g., 1) Concern raised by PeterZ, where if there are N partition sets of vmas belonging to tasks, then unfairness in allowing these threads to scan could potentially amplify the side effect of some of the vmas being left unscanned. 2) Below reports of LKP numa01 benchmark regression. Currently this is handled by allowing first two scanning unconditional as indicated by mm->numa_scan_seq. This is imprecise since for some benchmark vma scanning might itself start at numa_scan_seq > 2. Solution: Allow unconditional scanning of vmas of tasks depending on vma size. This is achieved by maintaining a per vma scan counter, where f(allowed_to_scan) = f(scan_counter < vma_size / scan_size) Result: numa01_THREAD_ALLOC result on 6.4.0-rc2 (that has numascan enhancement) base-numascan base base+fix real 1m1.507s 1m23.259s 1m2.632s user 213m51.336s 251m46.363s 220m35.528s sys 3m3.397s 0m12.492s 2m41.393s numa_hit 5615517 4560123 4963875 numa_local 5615505 4560024 4963700 numa_other 12 99 175 numa_pte_updates 1822797 493 1559111 numa_hint_faults 1307113 523 1469031 numa_hint_faults_local 612617 488 884829 numa_pages_migrated 694370 35 584202 Summary: Regression in base is recovered by allowing scanning as required. [1] https://lore.kernel.org/lkml/cover.1677672277.git.raghavendra.kt@amd.com/T/#t Fixes: fc137c0ddab2 ("sched/numa: enhance vma scanning logic") regression. Reported-by: Aithal Srikanth <sraithal@amd.com> Reported-by: kernel test robot <oliver.sang@intel.com> Closes: https://lore.kernel.org/lkml/db995c11-08ba-9abf-812f-01407f70a5d4@amd.com/T/ Signed-off-by: Raghavendra K T <raghavendra.kt@amd.com> --- include/linux/mm_types.h | 1 + kernel/sched/fair.c | 31 ++++++++++++++++++++++++------- 2 files changed, 25 insertions(+), 7 deletions(-)