Hello, kernel test robot noticed a 53.9% improvement of fsmark.files_per_sec on: commit: 2850eee23dbc4ff9878d88625b1f84965eefcce6 ("[PATCH 13/13] writeback: set the num of writeback contexts to number of online cpus") url: https://github.com/intel-lab-lkp/linux/commits/Kundan-Kumar/writeback-add-infra-for-parallel-writeback/20250529-193523 base: https://git.kernel.org/cgit/linux/kernel/git/vfs/vfs.git vfs.all patch link: https://lore.kernel.org/all/20250529111504.89912-14-kundan.kumar@xxxxxxxxxxx/ patch subject: [PATCH 13/13] writeback: set the num of writeback contexts to number of online cpus testcase: fsmark config: x86_64-rhel-9.4 compiler: gcc-12 test machine: 192 threads 4 sockets Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz (Cascade Lake) with 176G memory parameters: iterations: 1x nr_threads: 32t disk: 1SSD fs: ext4 filesize: 16MB test_size: 60G sync_method: NoSync nr_directories: 16d nr_files_per_directory: 256fpd cpufreq_governor: performance In addition to that, the commit also has significant impact on the following tests: +------------------+------------------------------------------------------------------------------------------------+ | testcase: change | filebench: filebench.sum_operations/s 4.3% improvement | | test machine | 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory | | test parameters | cpufreq_governor=performance | | | disk=1HDD | | | fs=xfs | | | test=fivestreamwrite.f | +------------------+------------------------------------------------------------------------------------------------+ Details are as below: --------------------------------------------------------------------------------------------------> The kernel config and materials to reproduce are available at: https://download.01.org/0day-ci/archive/20250603/202506032246.89ddc1a2-lkp@xxxxxxxxx ========================================================================================= compiler/cpufreq_governor/disk/filesize/fs/iterations/kconfig/nr_directories/nr_files_per_directory/nr_threads/rootfs/sync_method/tbox_group/test_size/testcase: gcc-12/performance/1SSD/16MB/ext4/1x/x86_64-rhel-9.4/16d/256fpd/32t/debian-12-x86_64-20240206.cgz/NoSync/lkp-csl-2sp10/60G/fsmark commit: a2dadb7ea8 ("nfs: add support in nfs to handle multiple writeback contexts") 2850eee23d ("writeback: set the num of writeback contexts to number of online cpus") a2dadb7ea862d5c1 2850eee23dbc4ff9878d88625b1 ---------------- --------------------------- %stddev %change %stddev \ | \ 1641480 +13.3% 1860148 ± 2% cpuidle..usage 302.00 ± 8% +14.1% 344.67 ± 7% perf-c2c.HITM.remote 24963 ± 4% -13.3% 21647 ± 6% uptime.idle 91.64 -22.2% 71.26 ± 7% iostat.cpu.idle 7.34 ± 4% +275.7% 27.59 ± 19% iostat.cpu.iowait 0.46 ±141% +2.2 2.63 ± 66% perf-profile.calltrace.cycles-pp.setlocale 0.46 ±141% +2.2 2.63 ± 66% perf-profile.children.cycles-pp.setlocale 194019 -7.5% 179552 fsmark.app_overhead 108.10 ± 8% +53.9% 166.40 ± 10% fsmark.files_per_sec 43295 ± 7% +35.8% 58787 ± 5% fsmark.time.voluntary_context_switches 19970922 -10.5% 17867270 ± 2% meminfo.Dirty 493817 +13.1% 558422 meminfo.SUnreclaim 141708 +1439.0% 2180863 ± 14% meminfo.Writeback 4993428 -10.3% 4480219 ± 2% proc-vmstat.nr_dirty 34285 +5.8% 36262 proc-vmstat.nr_kernel_stack 123504 +13.1% 139636 proc-vmstat.nr_slab_unreclaimable 36381 ± 4% +1403.0% 546810 ± 14% proc-vmstat.nr_writeback 91.54 -22.1% 71.32 ± 7% vmstat.cpu.id 7.22 ± 4% +280.4% 27.47 ± 19% vmstat.cpu.wa 14.58 ± 4% +537.4% 92.92 ± 8% vmstat.procs.b 6140 ± 2% +90.1% 11673 ± 9% vmstat.system.cs 91.46 -20.9 70.56 ± 7% mpstat.cpu.all.idle% 7.52 ± 4% +20.8 28.29 ± 19% mpstat.cpu.all.iowait% 0.12 ± 7% +0.0 0.14 ± 7% mpstat.cpu.all.irq% 0.35 ± 6% +0.1 0.43 ± 5% mpstat.cpu.all.sys% 11.24 ± 8% +20.7% 13.56 ± 3% mpstat.max_utilization_pct 34947 ± 5% +14.3% 39928 ± 4% numa-vmstat.node0.nr_slab_unreclaimable 9001 ± 14% +1553.7% 148860 ± 19% numa-vmstat.node0.nr_writeback 1329092 ± 4% -20.9% 1051569 ± 9% numa-vmstat.node1.nr_dirty 10019 ± 7% +1490.0% 159311 ± 14% numa-vmstat.node1.nr_writeback 2808522 ± 8% -17.7% 2310216 ± 2% numa-vmstat.node2.nr_file_pages 2638799 ± 3% -12.7% 2304024 ± 2% numa-vmstat.node2.nr_inactive_file 7810 ± 9% +1035.8% 88707 ± 16% numa-vmstat.node2.nr_writeback 2638797 ± 3% -12.7% 2304025 ± 2% numa-vmstat.node2.nr_zone_inactive_file 29952 ± 3% +13.4% 33964 ± 4% numa-vmstat.node3.nr_slab_unreclaimable 10686 ± 9% +1351.3% 155091 ± 12% numa-vmstat.node3.nr_writeback 139656 ± 5% +14.2% 159539 ± 4% numa-meminfo.node0.SUnreclaim 35586 ± 13% +1565.8% 592799 ± 18% numa-meminfo.node0.Writeback 5304285 ± 4% -20.8% 4198452 ± 10% numa-meminfo.node1.Dirty 40011 ± 5% +1484.2% 633862 ± 14% numa-meminfo.node1.Writeback 11211668 ± 7% -17.7% 9222157 ± 2% numa-meminfo.node2.FilePages 10532776 ± 3% -12.7% 9197387 ± 2% numa-meminfo.node2.Inactive 10532776 ± 3% -12.7% 9197387 ± 2% numa-meminfo.node2.Inactive(file) 12378624 ± 7% -15.0% 10520827 ± 2% numa-meminfo.node2.MemUsed 29574 ± 9% +1087.0% 351055 ± 16% numa-meminfo.node2.Writeback 119679 ± 3% +13.4% 135718 ± 4% numa-meminfo.node3.SUnreclaim 41446 ± 10% +1380.1% 613443 ± 11% numa-meminfo.node3.Writeback 23.38 ± 2% -6.7 16.72 perf-stat.i.cache-miss-rate% 38590732 ± 3% +53.9% 59394561 ± 5% perf-stat.i.cache-references 5973 ± 2% +96.3% 11729 ± 10% perf-stat.i.context-switches 0.92 +7.4% 0.99 perf-stat.i.cpi 7.023e+09 ± 3% +12.5% 7.898e+09 ± 4% perf-stat.i.cpu-cycles 237.41 ± 2% +393.6% 1171 ± 20% perf-stat.i.cpu-migrations 1035 ± 3% +9.3% 1132 ± 4% perf-stat.i.cycles-between-cache-misses 1.15 -5.7% 1.09 perf-stat.i.ipc 25.41 ± 2% -7.4 18.03 ± 2% perf-stat.overall.cache-miss-rate% 0.94 +7.2% 1.01 perf-stat.overall.cpi 1.06 -6.8% 0.99 perf-stat.overall.ipc 38042801 ± 3% +54.0% 58576659 ± 5% perf-stat.ps.cache-references 5897 ± 2% +96.3% 11577 ± 10% perf-stat.ps.context-switches 6.925e+09 ± 3% +12.4% 7.787e+09 ± 4% perf-stat.ps.cpu-cycles 234.11 ± 2% +394.2% 1156 ± 20% perf-stat.ps.cpu-migrations 5.892e+11 +1.4% 5.977e+11 perf-stat.total.instructions 0.08 ± 6% -36.2% 0.05 ± 33% perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.folio_alloc_noprof.__filemap_get_folio 0.01 ± 73% +159.8% 0.04 ± 13% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.07 ± 10% -33.0% 0.05 ± 44% perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 0.01 ± 73% +695.6% 0.09 ± 25% perf-sched.sch_delay.avg.ms.io_schedule.blk_mq_get_tag.__blk_mq_alloc_requests.blk_mq_submit_bio 0.03 ± 30% -45.4% 0.02 ± 24% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] 0.06 ± 21% -100.0% 0.00 perf-sched.sch_delay.avg.ms.kjournald2.kthread.ret_from_fork.ret_from_fork_asm 0.05 ± 45% +305.0% 0.19 ± 66% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 0.01 ±210% +8303.3% 0.85 ±183% perf-sched.sch_delay.max.ms.__cond_resched.down_write.mpage_map_and_submit_extent.ext4_do_writepages.ext4_writepages 0.05 ± 49% +92.4% 0.10 ± 35% perf-sched.sch_delay.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork 0.03 ± 76% +165.2% 0.07 ± 28% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.07 ± 46% +83423.5% 59.86 ±146% perf-sched.sch_delay.max.ms.io_schedule.blk_mq_get_tag.__blk_mq_alloc_requests.blk_mq_submit_bio 0.06 ± 21% -100.0% 0.00 perf-sched.sch_delay.max.ms.kjournald2.kthread.ret_from_fork.ret_from_fork_asm 0.09 ± 7% +19.7% 0.10 ± 8% perf-sched.sch_delay.max.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm 0.10 ± 13% +3587.7% 3.60 ±172% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 18911 +109.0% 39526 ± 25% perf-sched.total_wait_and_delay.count.ms 4.42 ± 25% +77.8% 7.86 ± 15% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 150.39 ± 6% -14.9% 127.97 ± 8% perf-sched.wait_and_delay.avg.ms.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read 8.03 ± 89% -74.5% 2.05 ±143% perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 1.24 ± 8% +2018.4% 26.26 ± 26% perf-sched.wait_and_delay.avg.ms.io_schedule.blk_mq_get_tag.__blk_mq_alloc_requests.blk_mq_submit_bio 0.83 ± 2% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 86.14 ± 9% +33.0% 114.52 ± 7% perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 1047 ± 6% -8.2% 960.83 perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 171.50 ± 7% -69.1% 53.00 ±141% perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call 162.50 ± 7% -69.3% 49.83 ±141% perf-sched.wait_and_delay.count.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 3635 ± 6% +451.1% 20036 ± 35% perf-sched.wait_and_delay.count.io_schedule.blk_mq_get_tag.__blk_mq_alloc_requests.blk_mq_submit_bio 26.33 ± 5% -12.7% 23.00 ± 2% perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll 116.17 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 3938 ± 21% -66.2% 1332 ± 60% perf-sched.wait_and_delay.count.schedule_timeout.io_schedule_timeout.balance_dirty_pages.balance_dirty_pages_ratelimited_flags 4831 ± 5% +102.8% 9799 ± 23% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 87.47 ± 20% +823.3% 807.60 ±141% perf-sched.wait_and_delay.max.ms.io_schedule.blk_mq_get_tag.__blk_mq_alloc_requests.blk_mq_submit_bio 2.81 ± 4% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 77.99 ± 13% +2082.7% 1702 ± 73% perf-sched.wait_and_delay.max.ms.schedule_timeout.io_schedule_timeout.balance_dirty_pages.balance_dirty_pages_ratelimited_flags 396.00 ± 16% -34.0% 261.22 ± 33% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 3.89 ± 18% +100.6% 7.81 ± 15% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 0.13 ±157% +2.1e+05% 271.76 ±126% perf-sched.wait_time.avg.ms.__cond_resched.down_write.mpage_map_and_submit_extent.ext4_do_writepages.ext4_writepages 150.37 ± 6% -15.2% 127.52 ± 8% perf-sched.wait_time.avg.ms.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read 1.23 ± 8% +2030.8% 26.17 ± 26% perf-sched.wait_time.avg.ms.io_schedule.blk_mq_get_tag.__blk_mq_alloc_requests.blk_mq_submit_bio 31.17 ± 50% -100.0% 0.00 perf-sched.wait_time.avg.ms.kjournald2.kthread.ret_from_fork.ret_from_fork_asm 86.09 ± 9% +32.8% 114.34 ± 7% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 1086 ± 17% +309.4% 4449 ± 26% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 0.17 ±142% +7.2e+05% 1196 ±119% perf-sched.wait_time.max.ms.__cond_resched.down_write.mpage_map_and_submit_extent.ext4_do_writepages.ext4_writepages 7.27 ± 45% +1259.6% 98.80 ±139% perf-sched.wait_time.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork 262.77 ±113% +316.1% 1093 ± 52% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 87.45 ± 20% +823.5% 807.53 ±141% perf-sched.wait_time.max.ms.io_schedule.blk_mq_get_tag.__blk_mq_alloc_requests.blk_mq_submit_bio 0.04 ± 30% +992.0% 0.43 ± 92% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] 31.17 ± 50% -100.0% 0.00 perf-sched.wait_time.max.ms.kjournald2.kthread.ret_from_fork.ret_from_fork_asm 75.85 ± 16% +2144.2% 1702 ± 73% perf-sched.wait_time.max.ms.schedule_timeout.io_schedule_timeout.balance_dirty_pages.balance_dirty_pages_ratelimited_flags 395.95 ± 16% -34.0% 261.15 ± 33% perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread *************************************************************************************************** lkp-icl-2sp6: 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory ========================================================================================= compiler/cpufreq_governor/disk/fs/kconfig/rootfs/tbox_group/test/testcase: gcc-12/performance/1HDD/xfs/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/lkp-icl-2sp6/fivestreamwrite.f/filebench commit: a2dadb7ea8 ("nfs: add support in nfs to handle multiple writeback contexts") 2850eee23d ("writeback: set the num of writeback contexts to number of online cpus") a2dadb7ea862d5c1 2850eee23dbc4ff9878d88625b1 ---------------- --------------------------- %stddev %change %stddev \ | \ 2.06 ± 3% +1.5 3.58 mpstat.cpu.all.iowait% 8388855 ± 5% +17.7% 9875928 ± 5% numa-meminfo.node0.Dirty 0.02 ± 5% +48.0% 0.03 ± 3% sched_debug.cpu.nr_uninterruptible.avg 2.70 +72.6% 4.65 vmstat.procs.b 97.67 -1.5% 96.17 iostat.cpu.idle 2.04 ± 3% +73.9% 3.55 iostat.cpu.iowait 2094449 ± 5% +17.8% 2468063 ± 5% numa-vmstat.node0.nr_dirty 2113170 ± 5% +17.7% 2487005 ± 5% numa-vmstat.node0.nr_zone_write_pending 6.99 ± 3% +0.5 7.48 ± 2% perf-stat.i.cache-miss-rate% 1.82 +3.2% 1.88 perf-stat.i.cpi 0.64 -2.0% 0.62 perf-stat.i.ipc 2.88 ± 5% +9.5% 3.15 ± 4% perf-stat.overall.MPKI 464.45 +4.3% 484.58 filebench.sum_bytes_mb/s 27873 +4.3% 29084 filebench.sum_operations 464.51 +4.3% 484.66 filebench.sum_operations/s 10.76 -4.2% 10.31 filebench.sum_time_ms/op 464.67 +4.3% 484.67 filebench.sum_writes/s 57175040 +4.2% 59565397 filebench.time.file_system_outputs 7146880 +4.2% 7445674 proc-vmstat.nr_dirtied 4412053 +9.1% 4815253 proc-vmstat.nr_dirty 7485090 +5.0% 7855964 proc-vmstat.nr_file_pages 24899858 -1.5% 24530112 proc-vmstat.nr_free_pages 24705120 -1.5% 24343672 proc-vmstat.nr_free_pages_blocks 6573042 +5.6% 6943969 proc-vmstat.nr_inactive_file 34473 ± 3% +7.5% 37072 proc-vmstat.nr_writeback 6573042 +5.6% 6943969 proc-vmstat.nr_zone_inactive_file 4446526 +9.1% 4852325 proc-vmstat.nr_zone_write_pending 7963041 +3.8% 8262916 proc-vmstat.pgalloc_normal 0.02 ± 10% +45.0% 0.03 ± 5% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 317.88 ±166% -100.0% 0.08 ± 60% perf-sched.sch_delay.avg.ms.kthreadd.ret_from_fork.ret_from_fork_asm 474.99 ±141% -100.0% 0.10 ± 49% perf-sched.sch_delay.max.ms.kthreadd.ret_from_fork.ret_from_fork_asm 17.87 ± 13% +125.8% 40.36 ± 4% perf-sched.wait_and_delay.avg.ms.io_schedule.blk_mq_get_tag.__blk_mq_alloc_requests.blk_mq_submit_bio 47.75 +19.8% 57.20 ± 5% perf-sched.wait_and_delay.avg.ms.schedule_timeout.io_schedule_timeout.balance_dirty_pages.balance_dirty_pages_ratelimited_flags 517.00 -17.2% 427.83 ± 5% perf-sched.wait_and_delay.count.schedule_timeout.io_schedule_timeout.balance_dirty_pages.balance_dirty_pages_ratelimited_flags 54.05 ± 2% +253.0% 190.80 ± 18% perf-sched.wait_and_delay.max.ms.schedule_timeout.io_schedule_timeout.balance_dirty_pages.balance_dirty_pages_ratelimited_flags 4286 ± 4% -8.8% 3909 ± 8% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 17.77 ± 13% +126.0% 40.16 ± 4% perf-sched.wait_time.avg.ms.io_schedule.blk_mq_get_tag.__blk_mq_alloc_requests.blk_mq_submit_bio 47.63 +19.8% 57.06 ± 5% perf-sched.wait_time.avg.ms.schedule_timeout.io_schedule_timeout.balance_dirty_pages.balance_dirty_pages_ratelimited_flags 53.95 ± 2% +253.5% 190.70 ± 18% perf-sched.wait_time.max.ms.schedule_timeout.io_schedule_timeout.balance_dirty_pages.balance_dirty_pages_ratelimited_flags 4285 ± 4% -8.9% 3906 ± 7% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 0.77 ± 15% -0.3 0.43 ± 72% perf-profile.calltrace.cycles-pp.enqueue_task.ttwu_do_activate.sched_ttwu_pending.__flush_smp_call_function_queue.flush_smp_call_function_queue 1.76 ± 9% -0.3 1.44 ± 8% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 1.73 ± 10% -0.3 1.43 ± 8% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.85 ± 12% -0.2 0.66 ± 14% perf-profile.calltrace.cycles-pp.__pick_next_task.__schedule.schedule_idle.do_idle.cpu_startup_entry 4.74 ± 6% -0.6 4.17 ± 9% perf-profile.children.cycles-pp.__handle_mm_fault 4.92 ± 6% -0.6 4.36 ± 9% perf-profile.children.cycles-pp.handle_mm_fault 2.85 ± 5% -0.5 2.34 ± 8% perf-profile.children.cycles-pp.enqueue_task 2.60 ± 4% -0.4 2.15 ± 8% perf-profile.children.cycles-pp.enqueue_task_fair 2.58 ± 6% -0.4 2.22 ± 8% perf-profile.children.cycles-pp.do_pte_missing 2.04 ± 9% -0.3 1.70 ± 10% perf-profile.children.cycles-pp.do_read_fault 1.88 ± 5% -0.3 1.56 ± 7% perf-profile.children.cycles-pp.ttwu_do_activate 1.85 ± 10% -0.3 1.58 ± 12% perf-profile.children.cycles-pp.filemap_map_pages 1.20 ± 8% -0.2 0.98 ± 9% perf-profile.children.cycles-pp.__flush_smp_call_function_queue 0.49 ± 23% -0.2 0.29 ± 23% perf-profile.children.cycles-pp.set_next_task_fair 0.38 ± 32% -0.2 0.20 ± 39% perf-profile.children.cycles-pp.strnlen_user 0.44 ± 28% -0.2 0.26 ± 27% perf-profile.children.cycles-pp.set_next_entity 0.70 ± 7% -0.2 0.52 ± 7% perf-profile.children.cycles-pp.folios_put_refs 0.22 ± 20% -0.1 0.15 ± 27% perf-profile.children.cycles-pp.try_charge_memcg 0.02 ±141% +0.1 0.08 ± 44% perf-profile.children.cycles-pp.__blk_mq_alloc_driver_tag 0.09 ± 59% +0.1 0.18 ± 21% perf-profile.children.cycles-pp.irq_work_tick 0.26 ± 22% +0.2 0.41 ± 13% perf-profile.children.cycles-pp.cpu_stop_queue_work 0.34 ± 31% +0.2 0.57 ± 27% perf-profile.children.cycles-pp.perf_event_mmap_event 2.68 ± 10% +0.4 3.10 ± 12% perf-profile.children.cycles-pp.sched_balance_domains 0.38 ± 32% -0.2 0.19 ± 44% perf-profile.self.cycles-pp.strnlen_user 0.02 ±141% +0.1 0.08 ± 44% perf-profile.self.cycles-pp.ahci_single_level_irq_intr 0.22 ± 36% +0.2 0.43 ± 21% perf-profile.self.cycles-pp.sched_balance_rq 0.36 ± 41% +0.3 0.66 ± 18% perf-profile.self.cycles-pp._find_next_and_bit Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. -- 0-DAY CI Kernel Test Service https://github.com/intel/lkp-tests/wiki