Hello, we have no idea how this change could impact performance, however, we saw several performance improvements as below. for stress-ng.nanosleep.ops_per_sec, we did some cross platform check, still observed similar performance improvement. so we still make out below report FYI what we observed in our tests. kernel test robot noticed a 299.9% improvement of stress-ng.io-uring.ops_per_sec on: commit: 4a9f46a472591e7ffd0e9c5d8cdce26751852554 ("[PATCH] acpi: convert strncpy into strscpy") url: https://github.com/intel-lab-lkp/linux/commits/goralbaris/acpi-convert-strncpy-into-strscpy/20250507-232259 base: v6.15-rc5 patch link: https://lore.kernel.org/all/20250507152045.12266-1-goralbaris@xxxxxxxxx/ patch subject: [PATCH] acpi: convert strncpy into strscpy testcase: stress-ng config: x86_64-rhel-9.4 compiler: gcc-12 test machine: 192 threads 2 sockets Intel(R) Xeon(R) 6740E CPU @ 2.4GHz (Sierra Forest) with 256G memory parameters: nr_threads: 100% testtime: 60s test: io-uring cpufreq_governor: performance In addition to that, the commit also has significant impact on the following tests: +------------------+----------------------------------------------------------------------------------------------+ | testcase: change | stress-ng: stress-ng.nanosleep.ops_per_sec 23.8% improvement | | test machine | 256 threads 2 sockets Intel(R) Xeon(R) 6767P CPU @ 2.4GHz (Granite Rapids) with 256G memory | | test parameters | cpufreq_governor=performance | | | nr_threads=100% | | | test=nanosleep | | | testtime=60s | +------------------+----------------------------------------------------------------------------------------------+ | testcase: change | stress-ng: stress-ng.nanosleep.ops_per_sec 115.3% improvement | | test machine | 192 threads 2 sockets Intel(R) Xeon(R) 6740E CPU @ 2.4GHz (Sierra Forest) with 256G memory | | test parameters | cpufreq_governor=performance | | | nr_threads=100% | | | test=nanosleep | | | testtime=60s | +------------------+----------------------------------------------------------------------------------------------+ Details are as below: --------------------------------------------------------------------------------------------------> The kernel config and materials to reproduce are available at: https://download.01.org/0day-ci/archive/20250519/202505191014.718ca1d6-lkp@xxxxxxxxx ========================================================================================= compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime: gcc-12/performance/x86_64-rhel-9.4/100%/debian-12-x86_64-20240206.cgz/lkp-srf-2sp2/io-uring/stress-ng/60s commit: v6.15-rc5 4a9f46a472 ("acpi: convert strncpy into strscpy") v6.15-rc5 4a9f46a472591e7ffd0e9c5d8cd ---------------- --------------------------- %stddev %change %stddev \ | \ 20.29 -4.2% 19.44 boot-time.dhcp 8.679e+09 ± 9% -100.0% 0.00 cpuidle..time 3.258e+08 -100.0% 0.00 cpuidle..usage 20990053 +305.8% 85176369 ± 2% numa-numastat.node1.local_node 21106117 +304.1% 85283811 ± 2% numa-numastat.node1.numa_hit 124217 ± 9% +25.7% 156113 ± 6% perf-c2c.HITM.local 126125 ± 9% +25.5% 158230 ± 6% perf-c2c.HITM.total 3.642e+08 +299.9% 1.456e+09 ± 25% stress-ng.io-uring.ops 6069877 +299.9% 24271031 ± 25% stress-ng.io-uring.ops_per_sec 63452087 +102.5% 1.285e+08 ± 19% stress-ng.time.involuntary_context_switches 5431 +182.5% 15343 ± 20% stress-ng.time.percent_of_cpu_this_job_got 3051 +195.8% 9025 ± 21% stress-ng.time.system_time 4.461e+08 +289.0% 1.735e+09 ± 25% stress-ng.time.voluntary_context_switches 48.09 ± 6% -34.7 13.42 ± 66% mpstat.cpu.all.idle% 19.76 ± 6% -14.5 5.29 ± 89% mpstat.cpu.all.iowait% 4.84 ± 6% -2.6 2.21 ± 77% mpstat.cpu.all.irq% 0.32 ± 6% -0.3 0.04 ± 75% mpstat.cpu.all.soft% 24.66 ± 6% +52.0 76.61 ± 19% mpstat.cpu.all.sys% 17.00 +16.7% 19.83 ± 12% mpstat.max_utilization.seconds 57.31 +57.7% 90.36 ± 10% mpstat.max_utilization_pct 49.43 ± 6% -67.3% 16.17 ± 53% vmstat.cpu.id 29.04 ± 6% +162.8% 76.30 ± 17% vmstat.cpu.sy 19.17 ± 6% -73.3% 5.12 ± 89% vmstat.cpu.wa 71.25 ± 5% -72.5% 19.58 ± 90% vmstat.procs.b 97.21 ± 7% +124.6% 218.37 ± 15% vmstat.procs.r 12688269 ± 6% +142.9% 30825156 ± 19% vmstat.system.cs 5224657 ± 6% -62.8% 1943708 ± 46% vmstat.system.in 289935 ± 3% -22.2% 225623 ± 5% numa-meminfo.node0.SUnreclaim 386160 ± 6% -21.4% 303599 ± 12% numa-meminfo.node0.Slab 11844753 ± 3% -36.4% 7538537 ± 31% numa-meminfo.node1.Active 11844753 ± 3% -36.4% 7538537 ± 31% numa-meminfo.node1.Active(anon) 58156 ± 5% -62.4% 21876 ± 6% numa-meminfo.node1.KernelStack 1029721 ± 12% +32.1% 1360593 ± 11% numa-meminfo.node1.Mapped 234556 ± 4% -25.9% 173820 ± 3% numa-meminfo.node1.SUnreclaim 10965603 ± 4% -37.8% 6824977 ± 33% numa-meminfo.node1.Shmem 305421 ± 8% -17.7% 251285 ± 12% numa-meminfo.node1.Slab 72401 ± 3% -22.1% 56378 ± 5% numa-vmstat.node0.nr_slab_unreclaimable 2963605 ± 3% -36.4% 1885576 ± 31% numa-vmstat.node1.nr_active_anon 58030 ± 5% -62.4% 21811 ± 6% numa-vmstat.node1.nr_kernel_stack 258544 ± 12% +32.1% 341648 ± 11% numa-vmstat.node1.nr_mapped 2743743 ± 4% -37.8% 1707508 ± 33% numa-vmstat.node1.nr_shmem 58559 ± 4% -25.8% 43437 ± 3% numa-vmstat.node1.nr_slab_unreclaimable 2963598 ± 3% -36.4% 1885576 ± 31% numa-vmstat.node1.nr_zone_active_anon 21105653 +304.1% 85285462 ± 2% numa-vmstat.node1.numa_hit 20989589 +305.8% 85178020 ± 2% numa-vmstat.node1.numa_local 12284550 ± 3% -33.5% 8175355 ± 28% meminfo.Active 12284550 ± 3% -33.5% 8175355 ± 28% meminfo.Active(anon) 14613224 ± 3% -28.3% 10483768 ± 22% meminfo.Cached 15092839 ± 2% -26.1% 11151503 ± 20% meminfo.Committed_AS 118013 ± 4% -47.1% 62476 ± 29% meminfo.KernelStack 1189737 ± 9% +24.5% 1481177 ± 12% meminfo.Mapped 19156418 ± 2% -23.0% 14742923 ± 15% meminfo.Memused 524570 -23.8% 399461 ± 2% meminfo.SUnreclaim 11073466 ± 4% -37.3% 6944224 ± 33% meminfo.Shmem 691643 -19.8% 554889 meminfo.Slab 399938 -13.6% 345539 ± 5% meminfo.VmallocUsed 19273868 ± 2% -23.1% 14829426 ± 15% meminfo.max_used_kB 3072505 ± 3% -33.5% 2044367 ± 28% proc-vmstat.nr_active_anon 303038 +1.5% 307548 proc-vmstat.nr_anon_pages 5974768 +1.8% 6084995 proc-vmstat.nr_dirty_background_threshold 11964146 +1.8% 12184870 proc-vmstat.nr_dirty_threshold 3655222 ± 3% -28.3% 2622497 ± 22% proc-vmstat.nr_file_pages 60117910 +1.8% 61221778 proc-vmstat.nr_free_pages 60050485 +1.8% 61151282 proc-vmstat.nr_free_pages_blocks 117206 ± 4% -46.9% 62248 ± 28% proc-vmstat.nr_kernel_stack 298065 ± 9% +24.9% 372258 ± 12% proc-vmstat.nr_mapped 2770279 ± 4% -37.3% 1737611 ± 33% proc-vmstat.nr_shmem 41774 -7.0% 38862 ± 3% proc-vmstat.nr_slab_reclaimable 130994 -23.8% 99846 ± 2% proc-vmstat.nr_slab_unreclaimable 3072505 ± 3% -33.5% 2044367 ± 28% proc-vmstat.nr_zone_active_anon 37968818 +257.4% 1.357e+08 ± 25% proc-vmstat.numa_hit 768.00 +22.1% 937.67 ± 11% proc-vmstat.numa_huge_pte_updates 37769910 +258.7% 1.355e+08 ± 25% proc-vmstat.numa_local 113917 ± 3% -58.9% 46844 ± 9% proc-vmstat.numa_pages_migrated 622038 ± 3% +16.6% 725068 ± 7% proc-vmstat.numa_pte_updates 38252525 +255.2% 1.359e+08 ± 25% proc-vmstat.pgalloc_normal 34002082 +291.1% 1.33e+08 ± 25% proc-vmstat.pgfree 113917 ± 3% -58.9% 46844 ± 9% proc-vmstat.pgmigrate_success 19955 -43.2% 11333 proc-vmstat.pgreuse 17603355 -73.2% 4719412 ± 91% proc-vmstat.unevictable_pgs_culled 0.31 ± 57% -71.0% 0.09 ± 6% perf-stat.i.MPKI 2.695e+10 ± 6% +157.7% 6.946e+10 ± 18% perf-stat.i.branch-instructions 0.93 -0.5 0.45 ± 12% perf-stat.i.branch-miss-rate% 2.182e+08 ± 6% +24.3% 2.712e+08 ± 7% perf-stat.i.branch-misses 3.06 ± 33% +5.9 8.91 ± 64% perf-stat.i.cache-miss-rate% 25459047 ± 6% -18.4% 20785561 ± 16% perf-stat.i.cache-misses 1.322e+09 ± 6% -69.1% 4.09e+08 ± 64% perf-stat.i.cache-references 13341377 ± 6% +142.6% 32367405 ± 19% perf-stat.i.context-switches 3.31 ± 4% -47.1% 1.75 ± 11% perf-stat.i.cpi 3.765e+11 ± 6% +45.5% 5.479e+11 ± 8% perf-stat.i.cpu-cycles 5523943 ± 6% -75.4% 1360465 ±100% perf-stat.i.cpu-migrations 25535 ± 4% +58.9% 40578 ± 10% perf-stat.i.cycles-between-cache-misses 1.337e+11 ± 6% +156.7% 3.432e+11 ± 19% perf-stat.i.instructions 0.35 +76.4% 0.62 ± 10% perf-stat.i.ipc 98.25 ± 6% +78.8% 175.66 ± 14% perf-stat.i.metric.K/sec 1.29 ± 70% +7.8 9.13 ± 71% perf-stat.overall.cache-miss-rate% 9811 ± 70% +173.4% 26822 ± 11% perf-stat.overall.cycles-between-cache-misses 0.24 ± 70% +162.2% 0.62 ± 10% perf-stat.overall.ipc 1.729e+10 ± 71% +295.2% 6.833e+10 ± 18% perf-stat.ps.branch-instructions 1.401e+08 ± 71% +90.5% 2.669e+08 ± 7% perf-stat.ps.branch-misses 8559237 ± 71% +272.1% 31851996 ± 19% perf-stat.ps.context-switches 2.415e+11 ± 71% +123.3% 5.391e+11 ± 8% perf-stat.ps.cpu-cycles 8.577e+10 ± 71% +293.6% 3.376e+11 ± 19% perf-stat.ps.instructions 5.678e+12 ± 70% +267.4% 2.086e+13 ± 19% perf-stat.total.instructions 1174613 +256.2% 4183841 ± 28% sched_debug.cfs_rq:/.avg_vruntime.avg 3797895 ± 12% +61.5% 6135427 ± 13% sched_debug.cfs_rq:/.avg_vruntime.max 970333 +185.9% 2773983 ± 52% sched_debug.cfs_rq:/.avg_vruntime.min 0.31 ± 6% +108.3% 0.64 ± 14% sched_debug.cfs_rq:/.h_nr_queued.avg 0.29 ± 6% +107.6% 0.59 ± 15% sched_debug.cfs_rq:/.h_nr_runnable.avg 138886 ± 16% -69.1% 42880 ±100% sched_debug.cfs_rq:/.left_deadline.avg 138860 ± 16% -69.1% 42873 ±100% sched_debug.cfs_rq:/.left_vruntime.avg 39.33 ± 48% -53.8% 18.15 ± 27% sched_debug.cfs_rq:/.load_avg.avg 1174613 +256.2% 4183841 ± 28% sched_debug.cfs_rq:/.min_vruntime.avg 3797895 ± 12% +61.5% 6135427 ± 13% sched_debug.cfs_rq:/.min_vruntime.max 970333 +185.9% 2773983 ± 52% sched_debug.cfs_rq:/.min_vruntime.min 0.21 ± 7% +114.0% 0.44 ± 16% sched_debug.cfs_rq:/.nr_queued.avg 0.37 ± 4% -31.8% 0.25 ± 40% sched_debug.cfs_rq:/.nr_queued.stddev 138860 ± 16% -69.1% 42873 ±100% sched_debug.cfs_rq:/.right_vruntime.avg 344.03 ± 4% +88.7% 649.02 ± 16% sched_debug.cfs_rq:/.runnable_avg.avg 1119 ± 10% +34.2% 1501 ± 2% sched_debug.cfs_rq:/.runnable_avg.max 297.90 ± 4% +64.8% 490.81 ± 16% sched_debug.cfs_rq:/.util_avg.avg 1064 ± 8% -10.6% 951.58 ± 3% sched_debug.cfs_rq:/.util_avg.max 72.80 ± 7% +415.2% 375.03 ± 24% sched_debug.cfs_rq:/.util_est.avg 1007 ± 12% +25.3% 1262 ± 7% sched_debug.cfs_rq:/.util_est.max 161.26 ± 7% +55.3% 250.42 ± 24% sched_debug.cfs_rq:/.util_est.stddev 585635 ± 13% +70.8% 1000253 ± 11% sched_debug.cpu.avg_idle.max 7853 ± 8% -59.3% 3197 ± 5% sched_debug.cpu.avg_idle.min 94982 ± 10% +49.7% 142227 ± 6% sched_debug.cpu.avg_idle.stddev 4964 ± 11% -45.9% 2684 ± 17% sched_debug.cpu.curr->pid.avg 36043 -65.5% 12443 ± 54% sched_debug.cpu.curr->pid.max 11755 ± 4% -74.0% 3057 ± 74% sched_debug.cpu.curr->pid.stddev 0.27 ± 7% +129.8% 0.61 ± 17% sched_debug.cpu.nr_running.avg 2.92 ± 11% -31.4% 2.00 ± 14% sched_debug.cpu.nr_running.max 2181080 +131.7% 5053532 ± 19% sched_debug.cpu.nr_switches.avg 2655743 +132.5% 6174297 ± 3% sched_debug.cpu.nr_switches.max 590551 ± 17% +325.9% 2515328 ± 53% sched_debug.cpu.nr_switches.min *************************************************************************************************** lkp-gnr-2sp3: 256 threads 2 sockets Intel(R) Xeon(R) 6767P CPU @ 2.4GHz (Granite Rapids) with 256G memory ========================================================================================= compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime: gcc-12/performance/x86_64-rhel-9.4/100%/debian-12-x86_64-20240206.cgz/lkp-gnr-2sp3/nanosleep/stress-ng/60s commit: v6.15-rc5 4a9f46a472 ("acpi: convert strncpy into strscpy") v6.15-rc5 4a9f46a472591e7ffd0e9c5d8cd ---------------- --------------------------- %stddev %change %stddev \ | \ 8.024e+09 ± 3% -100.0% 0.00 cpuidle..time 5.73e+08 ± 11% -100.0% 0.00 cpuidle..usage 5496607 ± 2% -10.7% 4906416 ± 2% numa-numastat.node1.local_node 5662628 ± 3% -10.8% 5052140 ± 2% numa-numastat.node1.numa_hit 291680 ± 3% -17.4% 240862 ± 6% numa-meminfo.node0.SUnreclaim 255763 ± 4% -16.0% 214915 ± 6% numa-meminfo.node1.SUnreclaim 13778965 ± 2% -10.7% 12305910 ± 2% numa-meminfo.node1.Shmem 36.41 ± 8% -51.2% 17.78 ± 2% vmstat.cpu.id 280.33 ± 11% +115.8% 604.80 vmstat.procs.r 12822190 ± 7% -54.9% 5782167 vmstat.system.in 370.77 ± 79% +198.2% 1105 ± 8% perf-c2c.DRAM.local 5844 ± 7% +25.8% 7349 ± 4% perf-c2c.DRAM.remote 302352 ± 3% -24.5% 228196 perf-c2c.HITM.local 306707 ± 3% -24.0% 232964 perf-c2c.HITM.total 15387789 ± 3% -9.7% 13890052 meminfo.Active 15387789 ± 3% -9.7% 13890052 meminfo.Active(anon) 547403 -16.7% 455722 meminfo.SUnreclaim 13989353 ± 3% -10.6% 12513459 meminfo.Shmem 740264 -13.2% 642211 meminfo.Slab 34.31 ± 9% -19.3 15.03 ± 2% mpstat.cpu.all.idle% 49.08 +8.1 57.22 mpstat.cpu.all.irq% 0.15 ± 18% +0.2 0.37 ± 2% mpstat.cpu.all.soft% 10.51 ± 19% +9.3 19.84 mpstat.cpu.all.sys% 82.11 +13.6% 93.31 mpstat.max_utilization_pct 72917 ± 3% -17.4% 60209 ± 6% numa-vmstat.node0.nr_slab_unreclaimable 3448051 ± 2% -10.8% 3076253 ± 2% numa-vmstat.node1.nr_shmem 63938 ± 4% -16.0% 53717 ± 6% numa-vmstat.node1.nr_slab_unreclaimable 5662986 ± 3% -10.8% 5052416 ± 2% numa-vmstat.node1.numa_hit 5496964 ± 2% -10.7% 4906693 ± 2% numa-vmstat.node1.numa_local 63477 ± 6% +60.5% 101863 stress-ng.nanosleep.nanosec_sleep_overrun 44977307 ± 3% +23.8% 55676879 stress-ng.nanosleep.ops 749576 ± 3% +23.8% 927912 stress-ng.nanosleep.ops_per_sec 12275219 ± 9% +150.4% 30741095 stress-ng.time.involuntary_context_switches 220678 +1.3% 223451 stress-ng.time.minor_page_faults 6525 +3.4% 6746 stress-ng.time.percent_of_cpu_this_job_got 2584 ± 2% +15.7% 2990 stress-ng.time.system_time 1359 ± 2% -19.8% 1089 stress-ng.time.user_time 3847978 ± 3% -9.8% 3472571 proc-vmstat.nr_active_anon 350848 -1.8% 344375 proc-vmstat.nr_anon_pages 4383290 ± 2% -8.4% 4014428 proc-vmstat.nr_file_pages 3498391 ± 3% -10.5% 3129513 proc-vmstat.nr_shmem 48220 -3.3% 46626 proc-vmstat.nr_slab_reclaimable 136860 -16.7% 113939 proc-vmstat.nr_slab_unreclaimable 3847978 ± 3% -9.8% 3472571 proc-vmstat.nr_zone_active_anon 6157573 ± 2% -10.9% 5484244 proc-vmstat.numa_hit 5893718 ± 3% -11.4% 5220452 proc-vmstat.numa_local 6454012 ± 2% -10.3% 5788637 proc-vmstat.pgalloc_normal 946345 ± 5% -17.0% 785728 ± 5% proc-vmstat.pgfault 21653 ± 4% -40.5% 12881 ± 4% proc-vmstat.pgreuse 0.25 ± 11% +62.5% 0.41 ± 2% perf-stat.i.MPKI 1.43 ± 4% +0.4 1.85 perf-stat.i.branch-miss-rate% 7.159e+08 +27.3% 9.113e+08 perf-stat.i.branch-misses 61844201 ± 6% +54.8% 95712375 ± 2% perf-stat.i.cache-misses 5.265e+09 ± 4% +34.7% 7.091e+09 perf-stat.i.cache-references 2.84 ± 5% +25.6% 3.57 perf-stat.i.cpi 7.337e+11 +21.7% 8.93e+11 perf-stat.i.cpu-cycles 15128294 ± 4% -20.6% 12012304 perf-stat.i.cpu-migrations 14582 ± 7% -32.3% 9871 ± 2% perf-stat.i.cycles-between-cache-misses 0.38 ± 5% -19.6% 0.31 ± 2% perf-stat.i.ipc 166.24 ± 5% -17.9% 136.55 perf-stat.i.metric.K/sec 10825 ± 7% -15.6% 9141 ± 7% perf-stat.i.minor-faults 10825 ± 7% -15.6% 9141 ± 7% perf-stat.i.page-faults 0.22 ± 11% +60.5% 0.35 ± 2% perf-stat.overall.MPKI 1.28 ± 4% +0.4 1.67 perf-stat.overall.branch-miss-rate% 1.18 ± 2% +0.2 1.35 ± 2% perf-stat.overall.cache-miss-rate% 2.61 ± 6% +26.5% 3.30 perf-stat.overall.cpi 11885 ± 4% -21.5% 9332 ± 2% perf-stat.overall.cycles-between-cache-misses 0.39 ± 5% -21.2% 0.30 perf-stat.overall.ipc 7.042e+08 +27.3% 8.964e+08 perf-stat.ps.branch-misses 60903975 ± 6% +54.6% 94180370 ± 2% perf-stat.ps.cache-misses 5.179e+09 ± 4% +34.7% 6.975e+09 perf-stat.ps.cache-references 7.218e+11 +21.7% 8.785e+11 perf-stat.ps.cpu-cycles 14879980 ± 4% -20.6% 11816383 perf-stat.ps.cpu-migrations 10630 ± 8% -15.4% 8990 ± 7% perf-stat.ps.minor-faults 10630 ± 8% -15.4% 8990 ± 7% perf-stat.ps.page-faults 1361488 ± 5% +28.1% 1743936 sched_debug.cfs_rq:/.avg_vruntime.avg 5955807 ± 14% +26.0% 7503578 ± 7% sched_debug.cfs_rq:/.avg_vruntime.max 1232820 ± 3% +19.5% 1472806 ± 2% sched_debug.cfs_rq:/.avg_vruntime.min 0.60 ± 10% +125.6% 1.36 ± 6% sched_debug.cfs_rq:/.h_nr_queued.avg 0.83 ± 12% +68.1% 1.40 ± 6% sched_debug.cfs_rq:/.h_nr_queued.stddev 0.53 ± 12% +146.6% 1.31 ± 6% sched_debug.cfs_rq:/.h_nr_runnable.avg 0.73 ± 13% +89.2% 1.39 ± 6% sched_debug.cfs_rq:/.h_nr_runnable.stddev 363369 ± 9% +31.7% 478423 ± 5% sched_debug.cfs_rq:/.left_deadline.avg 363289 ± 9% +31.7% 478341 ± 5% sched_debug.cfs_rq:/.left_vruntime.avg 13.19 ± 26% -52.6% 6.25 ± 8% sched_debug.cfs_rq:/.load_avg.avg 525.38 -65.1% 183.50 ± 62% sched_debug.cfs_rq:/.load_avg.max 60.90 ± 25% -75.2% 15.10 ± 41% sched_debug.cfs_rq:/.load_avg.stddev 1361488 ± 5% +28.1% 1743936 sched_debug.cfs_rq:/.min_vruntime.avg 5955807 ± 14% +26.0% 7503578 ± 7% sched_debug.cfs_rq:/.min_vruntime.max 1232820 ± 3% +19.5% 1472806 ± 2% sched_debug.cfs_rq:/.min_vruntime.min 0.31 ± 3% +33.2% 0.41 ± 3% sched_debug.cfs_rq:/.nr_queued.avg 0.33 ± 4% -18.4% 0.27 ± 5% sched_debug.cfs_rq:/.nr_queued.stddev 7.35 ± 50% -98.0% 0.15 ±346% sched_debug.cfs_rq:/.removed.load_avg.avg 511.38 -92.6% 37.88 ±346% sched_debug.cfs_rq:/.removed.load_avg.max 58.58 ± 26% -96.0% 2.36 ±346% sched_debug.cfs_rq:/.removed.load_avg.stddev 3.56 ± 53% -96.6% 0.12 ±346% sched_debug.cfs_rq:/.removed.runnable_avg.avg 28.44 ± 29% -93.1% 1.96 ±346% sched_debug.cfs_rq:/.removed.runnable_avg.stddev 3.56 ± 53% -96.6% 0.12 ±346% sched_debug.cfs_rq:/.removed.util_avg.avg 28.44 ± 29% -93.2% 1.95 ±346% sched_debug.cfs_rq:/.removed.util_avg.stddev 363289 ± 9% +31.7% 478341 ± 5% sched_debug.cfs_rq:/.right_vruntime.avg 257.62 ± 14% +322.5% 1088 ± 2% sched_debug.cfs_rq:/.runnable_avg.avg 946.38 ± 2% +160.5% 2465 ± 9% sched_debug.cfs_rq:/.runnable_avg.max 173.27 ± 10% +196.4% 513.58 sched_debug.cfs_rq:/.runnable_avg.min 114.87 ± 23% +137.8% 273.14 ± 5% sched_debug.cfs_rq:/.runnable_avg.stddev 235.05 ± 4% +22.8% 288.57 sched_debug.cfs_rq:/.util_avg.avg 104.92 ± 7% -12.0% 92.36 ± 4% sched_debug.cfs_rq:/.util_avg.stddev 8.93 ± 19% +1910.0% 179.43 ± 9% sched_debug.cfs_rq:/.util_est.avg 802.77 ± 13% +103.2% 1631 ± 20% sched_debug.cfs_rq:/.util_est.max 56.48 ± 19% +293.2% 222.08 ± 6% sched_debug.cfs_rq:/.util_est.stddev 22.54 ± 3% +26.3% 28.46 ± 3% sched_debug.cpu.clock.stddev 7188 -12.0% 6328 ± 3% sched_debug.cpu.curr->pid.max 2102 ± 3% -18.0% 1724 sched_debug.cpu.curr->pid.stddev 0.00 ± 3% +32.4% 0.00 ± 3% sched_debug.cpu.next_balance.stddev 0.56 ± 13% +131.1% 1.29 ± 6% sched_debug.cpu.nr_running.avg 0.82 ± 14% +62.1% 1.33 ± 7% sched_debug.cpu.nr_running.stddev 3679769 ± 2% -18.2% 3011353 sched_debug.cpu.nr_switches.max *************************************************************************************************** lkp-srf-2sp3: 192 threads 2 sockets Intel(R) Xeon(R) 6740E CPU @ 2.4GHz (Sierra Forest) with 256G memory ========================================================================================= compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime: gcc-12/performance/x86_64-rhel-9.4/100%/debian-12-x86_64-20240206.cgz/lkp-srf-2sp3/nanosleep/stress-ng/60s commit: v6.15-rc5 4a9f46a472 ("acpi: convert strncpy into strscpy") v6.15-rc5 4a9f46a472591e7ffd0e9c5d8cd ---------------- --------------------------- %stddev %change %stddev \ | \ 20.29 -5.2% 19.23 boot-time.dhcp 7.335e+09 -100.0% 0.00 cpuidle..time 3.421e+08 -100.0% 0.00 cpuidle..usage 1262963 -10.1% 1135588 meminfo.Mapped 475531 -20.4% 378366 meminfo.SUnreclaim 643055 -15.8% 541287 meminfo.Slab 24796 +318.7% 103826 ± 42% numa-numastat.node0.other_node 4890571 -12.8% 4263195 ± 8% numa-numastat.node1.numa_hit 173010 -45.7% 94003 ± 46% numa-numastat.node1.other_node 45.67 -42.2% 26.41 ± 4% vmstat.cpu.id 201.24 +69.8% 341.72 ± 2% vmstat.procs.r 17944140 +75.1% 31420986 ± 7% vmstat.system.cs 7977950 +20.0% 9571654 ± 8% vmstat.system.in 44.78 -19.8 24.95 ± 6% mpstat.cpu.all.idle% 34.90 +9.1 43.96 ± 2% mpstat.cpu.all.irq% 0.11 +0.0 0.15 ± 8% mpstat.cpu.all.soft% 12.22 +7.3 19.50 ± 2% mpstat.cpu.all.sys% 8.00 +3.5 11.46 ± 3% mpstat.cpu.all.usr% 66.86 +34.9% 90.21 ± 2% mpstat.max_utilization_pct 28236 +12.1% 31643 ± 6% numa-meminfo.node0.KernelStack 78589 +120.3% 173093 ± 44% numa-meminfo.node0.Mapped 252579 -16.0% 212085 ± 2% numa-meminfo.node0.SUnreclaim 5630 +2952.3% 171845 ± 63% numa-meminfo.node0.Shmem 340274 -15.3% 288246 ± 12% numa-meminfo.node0.Slab 30790 -10.2% 27655 ± 7% numa-meminfo.node1.KernelStack 1187307 -18.8% 964213 ± 6% numa-meminfo.node1.Mapped 222896 -25.4% 166281 ± 3% numa-meminfo.node1.SUnreclaim 15.55 +9.1e+06% 1415677 ±116% numa-meminfo.node1.Unevictable 317041 -10.2% 284725 proc-vmstat.nr_mapped 8595 +3.3% 8882 proc-vmstat.nr_page_table_pages 41884 -2.7% 40737 proc-vmstat.nr_slab_reclaimable 118869 -20.4% 94588 proc-vmstat.nr_slab_unreclaimable 216343 -22.6% 167393 ± 21% proc-vmstat.numa_hint_faults 212807 -22.9% 164083 ± 22% proc-vmstat.numa_hint_faults_local 747646 -23.0% 575328 ± 6% proc-vmstat.pgfault 690042 -17.9% 566229 ± 6% proc-vmstat.pgfree 18242 -41.5% 10679 proc-vmstat.pgreuse 66136 -17.8% 54373 ± 6% stress-ng.nanosleep.nanosec_sleep_overrun 32942741 +115.3% 70918788 ± 4% stress-ng.nanosleep.ops 549017 +115.3% 1181944 ± 4% stress-ng.nanosleep.ops_per_sec 7441350 +210.9% 23136625 ± 12% stress-ng.time.involuntary_context_switches 87380 +1.3% 88520 stress-ng.time.minor_page_faults 3996 +85.1% 7396 ± 7% stress-ng.time.percent_of_cpu_this_job_got 1488 +93.1% 2875 ± 6% stress-ng.time.system_time 917.88 +72.3% 1581 ± 7% stress-ng.time.user_time 7.582e+08 +77.7% 1.347e+09 ± 4% stress-ng.time.voluntary_context_switches 28240 +11.7% 31548 ± 6% numa-vmstat.node0.nr_kernel_stack 19908 +116.2% 43037 ± 43% numa-vmstat.node0.nr_mapped 1409 +2921.1% 42588 ± 63% numa-vmstat.node0.nr_shmem 63145 -16.0% 53015 ± 2% numa-vmstat.node0.nr_slab_unreclaimable 24796 +318.7% 103826 ± 42% numa-vmstat.node0.numa_other 30790 -10.4% 27580 ± 7% numa-vmstat.node1.nr_kernel_stack 298168 -18.8% 242086 ± 6% numa-vmstat.node1.nr_mapped 55724 -25.4% 41568 ± 3% numa-vmstat.node1.nr_slab_unreclaimable 3.89 +9.1e+06% 353919 ±116% numa-vmstat.node1.nr_unevictable 3.89 +9.1e+06% 353919 ±116% numa-vmstat.node1.nr_zone_unevictable 4890208 -12.8% 4262653 ± 8% numa-vmstat.node1.numa_hit 173010 -45.7% 94003 ± 46% numa-vmstat.node1.numa_other 0.29 -2.9% 0.29 perf-stat.i.MPKI 3.94e+10 +54.8% 6.101e+10 ± 4% perf-stat.i.branch-instructions 1.47 -0.1 1.36 ± 4% perf-stat.i.branch-miss-rate% 4.897e+08 +50.5% 7.371e+08 perf-stat.i.branch-misses 2.66 +0.2 2.85 ± 2% perf-stat.i.cache-miss-rate% 54650523 +44.1% 78748562 ± 6% perf-stat.i.cache-misses 2.506e+09 +32.6% 3.323e+09 perf-stat.i.cache-references 18553998 +75.0% 32467273 ± 7% perf-stat.i.context-switches 2.42 -17.7% 1.99 ± 2% perf-stat.i.cpi 4.091e+11 +34.7% 5.511e+11 ± 2% perf-stat.i.cpu-cycles 10138720 +79.7% 18216803 ± 9% perf-stat.i.cpu-migrations 9070 -16.0% 7620 ± 5% perf-stat.i.cycles-between-cache-misses 1.948e+11 +55.3% 3.025e+11 ± 4% perf-stat.i.instructions 0.47 +17.2% 0.55 perf-stat.i.ipc 149.44 +76.7% 264.00 ± 8% perf-stat.i.metric.K/sec 9609 -18.9% 7792 ± 7% perf-stat.i.minor-faults 9609 -18.9% 7792 ± 7% perf-stat.i.page-faults 0.28 -7.3% 0.26 ± 2% perf-stat.overall.MPKI 2.18 +0.2 2.37 ± 6% perf-stat.overall.cache-miss-rate% 2.10 -13.2% 1.82 ± 2% perf-stat.overall.cpi 7483 -6.3% 7014 ± 4% perf-stat.overall.cycles-between-cache-misses 0.48 +15.3% 0.55 ± 2% perf-stat.overall.ipc 3.878e+10 +54.9% 6.007e+10 ± 4% perf-stat.ps.branch-instructions 4.817e+08 +50.5% 7.25e+08 perf-stat.ps.branch-misses 53781450 +44.1% 77524937 ± 6% perf-stat.ps.cache-misses 2.466e+09 +32.6% 3.269e+09 perf-stat.ps.cache-references 18252323 +75.0% 31933413 ± 7% perf-stat.ps.context-switches 4.025e+11 +34.7% 5.422e+11 ± 2% perf-stat.ps.cpu-cycles 9974053 +79.6% 17916841 ± 9% perf-stat.ps.cpu-migrations 1.917e+11 +55.4% 2.978e+11 ± 4% perf-stat.ps.instructions 9443 -18.8% 7663 ± 7% perf-stat.ps.minor-faults 9443 -18.8% 7663 ± 7% perf-stat.ps.page-faults 1.18e+13 +55.3% 1.833e+13 ± 4% perf-stat.total.instructions 662859 +178.2% 1844033 ± 8% sched_debug.cfs_rq:/.avg_vruntime.avg 3155866 +94.4% 6133730 ± 2% sched_debug.cfs_rq:/.avg_vruntime.max 509569 +218.9% 1624879 ± 5% sched_debug.cfs_rq:/.avg_vruntime.min 219206 +76.6% 387056 ± 3% sched_debug.cfs_rq:/.avg_vruntime.stddev 0.64 +59.8% 1.02 ± 13% sched_debug.cfs_rq:/.h_nr_queued.avg 8.00 -40.0% 4.80 ± 14% sched_debug.cfs_rq:/.h_nr_queued.max 0.57 +50.0% 0.86 ± 16% sched_debug.cfs_rq:/.h_nr_runnable.avg 8.00 -41.2% 4.70 ± 10% sched_debug.cfs_rq:/.h_nr_runnable.max 119579 +285.9% 461420 ± 6% sched_debug.cfs_rq:/.left_deadline.avg 701402 +556.6% 4605499 ± 42% sched_debug.cfs_rq:/.left_deadline.max 245694 +241.9% 839983 ± 7% sched_debug.cfs_rq:/.left_deadline.stddev 119546 +285.9% 461364 ± 6% sched_debug.cfs_rq:/.left_vruntime.avg 701163 +556.8% 4605228 ± 42% sched_debug.cfs_rq:/.left_vruntime.max 245628 +241.9% 839885 ± 7% sched_debug.cfs_rq:/.left_vruntime.stddev 57782 -46.4% 30982 ± 12% sched_debug.cfs_rq:/.load.max 7422 -32.2% 5029 ± 10% sched_debug.cfs_rq:/.load.stddev 43.69 -59.6% 17.67 ± 20% sched_debug.cfs_rq:/.load_avg.avg 2.00 -25.0% 1.50 sched_debug.cfs_rq:/.load_avg.min 131.02 -51.1% 64.07 ± 26% sched_debug.cfs_rq:/.load_avg.stddev 662859 +178.2% 1844033 ± 8% sched_debug.cfs_rq:/.min_vruntime.avg 3155866 +94.4% 6133730 ± 2% sched_debug.cfs_rq:/.min_vruntime.max 509569 +218.9% 1624879 ± 5% sched_debug.cfs_rq:/.min_vruntime.min 219206 +76.6% 387056 ± 3% sched_debug.cfs_rq:/.min_vruntime.stddev 0.28 +41.3% 0.39 ± 3% sched_debug.cfs_rq:/.nr_queued.avg 34.26 -79.9% 6.90 ± 53% sched_debug.cfs_rq:/.removed.load_avg.avg 127.21 -60.3% 50.55 ± 51% sched_debug.cfs_rq:/.removed.load_avg.stddev 14.75 -79.8% 2.98 ± 53% sched_debug.cfs_rq:/.removed.runnable_avg.avg 56.08 -59.9% 22.49 ± 51% sched_debug.cfs_rq:/.removed.runnable_avg.stddev 14.75 -79.8% 2.98 ± 53% sched_debug.cfs_rq:/.removed.util_avg.avg 56.08 -59.9% 22.49 ± 51% sched_debug.cfs_rq:/.removed.util_avg.stddev 119546 +285.9% 461364 ± 6% sched_debug.cfs_rq:/.right_vruntime.avg 701163 +556.8% 4605228 ± 42% sched_debug.cfs_rq:/.right_vruntime.max 245628 +241.9% 839885 ± 7% sched_debug.cfs_rq:/.right_vruntime.stddev 255.64 +76.6% 451.41 ± 12% sched_debug.cfs_rq:/.runnable_avg.avg 127.50 +98.1% 252.60 ± 16% sched_debug.cfs_rq:/.runnable_avg.min 228.34 +54.8% 353.49 ± 5% sched_debug.cfs_rq:/.util_avg.avg 117.50 +104.7% 240.50 ± 17% sched_debug.cfs_rq:/.util_avg.min 11.37 +29.9% 14.77 ± 15% sched_debug.cfs_rq:/.util_est.avg 6599 -51.0% 3236 ± 6% sched_debug.cpu.avg_idle.min 1003 +34.0% 1344 ± 6% sched_debug.cpu.curr->pid.avg 5836 -9.8% 5267 sched_debug.cpu.curr->pid.max 1661 -10.8% 1482 ± 2% sched_debug.cpu.curr->pid.stddev 0.00 +14.5% 0.00 ± 5% sched_debug.cpu.next_balance.stddev 0.54 +71.4% 0.93 ± 11% sched_debug.cpu.nr_running.avg 8.50 -44.7% 4.70 ± 15% sched_debug.cpu.nr_running.max 2891369 +74.8% 5054237 ± 7% sched_debug.cpu.nr_switches.avg 3315998 +69.9% 5635000 sched_debug.cpu.nr_switches.max 1451647 +82.7% 2651528 ± 5% sched_debug.cpu.nr_switches.min 0.01 -53.3% 0.00 ± 57% sched_debug.cpu.nr_uninterruptible.avg 206.00 -43.6% 116.20 ± 45% sched_debug.cpu.nr_uninterruptible.max 24.50 -36.5% 15.54 ± 35% sched_debug.cpu.nr_uninterruptible.stddev Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. -- 0-DAY CI Kernel Test Service https://github.com/intel/lkp-tests/wiki