Re: [LKP] [rcu] kernel BUG at include/linux/pagemap.h:149!
by Frederic Weisbecker
On Fri, Sep 11, 2015 at 10:19:47AM +0800, Boqun Feng wrote:
> Subject: [PATCH 01/27] rcu: Don't disable preemption for Tiny and Tree RCU
> readers
>
> Because preempt_disable() maps to barrier() for non-debug builds,
> it forces the compiler to spill and reload registers. Because Tree
> RCU and Tiny RCU now only appear in CONFIG_PREEMPT=n builds, these
> barrier() instances generate needless extra code for each instance of
> rcu_read_lock() and rcu_read_unlock(). This extra code slows down Tree
> RCU and bloats Tiny RCU.
>
> This commit therefore removes the preempt_disable() and preempt_enable()
> from the non-preemptible implementations of __rcu_read_lock() and
> __rcu_read_unlock(), respectively.
>
> For debug purposes, preempt_disable() and preempt_enable() are still
> kept if CONFIG_PREEMPT_COUNT=y, which makes the detection of sleeping
> inside atomic sections still work in non-preemptible kernels.
>
> Signed-off-by: Boqun Feng <boqun.feng(a)gmail.com>
> Signed-off-by: Paul E. McKenney <paulmck(a)linux.vnet.ibm.com>
> ---
> include/linux/rcupdate.h | 6 ++++--
> include/linux/rcutiny.h | 1 +
> kernel/rcu/tree.c | 9 +++++++++
> 3 files changed, 14 insertions(+), 2 deletions(-)
>
> diff --git a/include/linux/rcupdate.h b/include/linux/rcupdate.h
> index d63bb77..6c3cece 100644
> --- a/include/linux/rcupdate.h
> +++ b/include/linux/rcupdate.h
> @@ -297,12 +297,14 @@ void synchronize_rcu(void);
>
> static inline void __rcu_read_lock(void)
> {
> - preempt_disable();
> + if (IS_ENABLED(CONFIG_PREEMPT_COUNT))
> + preempt_disable();
preempt_disable() is a no-op when !CONFIG_PREEMPT_COUNT, right?
Or rather it's a barrier(), which is anyway implied by rcu_read_lock().
So perhaps we can get rid of the IS_ENABLED() check?
3 years
Test monitoring on custom github repo
by Thomas Garnier
Hi,
I am working on KASLR (PIE for x86_64). I previously used Kees (CCed)
branches for lkp bot testing but someone told be I could ask you to add a
custom github path to monitor all branches on it.
I pushed my changes to: https://github.com/thgarnie/linux (kasrl_pie_v2
right now)
Can you add it? Anything I need to do?
Thanks,
--
Thomas
3 years, 5 months
[lkp-robot] [brd] 316ba5736c: aim7.jobs-per-min -11.2% regression
by kernel test robot
Greeting,
FYI, we noticed a -11.2% regression of aim7.jobs-per-min due to commit:
commit: 316ba5736c9caa5dbcd84085989862d2df57431d ("brd: Mark as non-rotational")
https://git.kernel.org/cgit/linux/kernel/git/axboe/linux-block.git for-4.18/block
in testcase: aim7
on test machine: 40 threads Intel(R) Xeon(R) CPU E5-2690 v2 @ 3.00GHz with 384G memory
with following parameters:
disk: 1BRD_48G
fs: btrfs
test: disk_rw
load: 1500
cpufreq_governor: performance
test-description: AIM7 is a traditional UNIX system level benchmark suite which is used to test and measure the performance of multiuser system.
test-url: https://sourceforge.net/projects/aimbench/files/aim-suite7/
Details are as below:
-------------------------------------------------------------------------------------------------->
=========================================================================================
compiler/cpufreq_governor/disk/fs/kconfig/load/rootfs/tbox_group/test/testcase:
gcc-7/performance/1BRD_48G/btrfs/x86_64-rhel-7.2/1500/debian-x86_64-2016-08-31.cgz/lkp-ivb-ep01/disk_rw/aim7
commit:
522a777566 ("block: consolidate struct request timestamp fields")
316ba5736c ("brd: Mark as non-rotational")
522a777566f56696 316ba5736c9caa5dbcd8408598
---------------- --------------------------
%stddev %change %stddev
\ | \
28321 -11.2% 25147 aim7.jobs-per-min
318.19 +12.6% 358.23 aim7.time.elapsed_time
318.19 +12.6% 358.23 aim7.time.elapsed_time.max
1437526 ± 2% +14.6% 1646849 ± 2% aim7.time.involuntary_context_switches
11986 +14.2% 13691 aim7.time.system_time
73.06 ± 2% -3.6% 70.43 aim7.time.user_time
2449470 ± 2% -25.0% 1837521 ± 4% aim7.time.voluntary_context_switches
20.25 ± 58% +1681.5% 360.75 ±109% numa-meminfo.node1.Mlocked
456062 -16.3% 381859 softirqs.SCHED
9015 ± 7% -21.3% 7098 ± 22% meminfo.CmaFree
47.50 ± 58% +1355.8% 691.50 ± 92% meminfo.Mlocked
5.24 ± 3% -1.2 3.99 ± 2% mpstat.cpu.idle%
0.61 ± 2% -0.1 0.52 ± 2% mpstat.cpu.usr%
16627 +12.8% 18762 ± 4% slabinfo.Acpi-State.active_objs
16627 +12.9% 18775 ± 4% slabinfo.Acpi-State.num_objs
57.00 ± 2% +17.5% 67.00 vmstat.procs.r
20936 -24.8% 15752 ± 2% vmstat.system.cs
45474 -1.7% 44681 vmstat.system.in
6.50 ± 59% +1157.7% 81.75 ± 75% numa-vmstat.node0.nr_mlock
242870 ± 3% +13.2% 274913 ± 7% numa-vmstat.node0.nr_written
2278 ± 7% -22.6% 1763 ± 21% numa-vmstat.node1.nr_free_cma
4.75 ± 58% +1789.5% 89.75 ±109% numa-vmstat.node1.nr_mlock
88018135 ± 3% -48.9% 44980457 ± 7% cpuidle.C1.time
1398288 ± 3% -51.1% 683493 ± 9% cpuidle.C1.usage
3499814 ± 2% -38.5% 2153158 ± 5% cpuidle.C1E.time
52722 ± 4% -45.6% 28692 ± 6% cpuidle.C1E.usage
9865857 ± 3% -40.1% 5905155 ± 5% cpuidle.C3.time
69656 ± 2% -42.6% 39990 ± 5% cpuidle.C3.usage
590856 ± 2% -12.3% 517910 cpuidle.C6.usage
46160 ± 7% -53.7% 21372 ± 11% cpuidle.POLL.time
1716 ± 7% -46.6% 916.25 ± 14% cpuidle.POLL.usage
197656 +4.1% 205732 proc-vmstat.nr_active_file
191867 +4.1% 199647 proc-vmstat.nr_dirty
509282 +1.6% 517318 proc-vmstat.nr_file_pages
2282 ± 8% -24.4% 1725 ± 22% proc-vmstat.nr_free_cma
357.50 +10.6% 395.25 ± 2% proc-vmstat.nr_inactive_file
11.50 ± 58% +1397.8% 172.25 ± 93% proc-vmstat.nr_mlock
970355 ± 4% +14.6% 1111549 ± 8% proc-vmstat.nr_written
197984 +4.1% 206034 proc-vmstat.nr_zone_active_file
357.50 +10.6% 395.25 ± 2% proc-vmstat.nr_zone_inactive_file
192282 +4.1% 200126 proc-vmstat.nr_zone_write_pending
7901465 ± 3% -14.0% 6795016 ± 16% proc-vmstat.pgalloc_movable
886101 +10.2% 976329 proc-vmstat.pgfault
2.169e+12 +15.2% 2.497e+12 perf-stat.branch-instructions
0.41 -0.1 0.35 perf-stat.branch-miss-rate%
31.19 ± 2% +1.6 32.82 perf-stat.cache-miss-rate%
9.116e+09 +8.3% 9.869e+09 perf-stat.cache-misses
2.924e+10 +2.9% 3.008e+10 ± 2% perf-stat.cache-references
6712739 ± 2% -15.4% 5678643 ± 2% perf-stat.context-switches
4.02 +2.7% 4.13 perf-stat.cpi
3.761e+13 +17.3% 4.413e+13 perf-stat.cpu-cycles
606958 -13.7% 523758 ± 2% perf-stat.cpu-migrations
2.476e+12 +13.4% 2.809e+12 perf-stat.dTLB-loads
0.18 ± 2% -0.0 0.16 ± 9% perf-stat.dTLB-store-miss-rate%
1.079e+09 ± 2% -9.6% 9.755e+08 ± 9% perf-stat.dTLB-store-misses
5.933e+11 +1.6% 6.029e+11 perf-stat.dTLB-stores
9.349e+12 +14.2% 1.068e+13 perf-stat.instructions
11247 ± 11% +19.8% 13477 ± 9% perf-stat.instructions-per-iTLB-miss
0.25 -2.6% 0.24 perf-stat.ipc
865561 +10.3% 954350 perf-stat.minor-faults
2.901e+09 ± 3% +9.8% 3.186e+09 ± 3% perf-stat.node-load-misses
3.682e+09 ± 3% +11.0% 4.088e+09 ± 3% perf-stat.node-loads
3.778e+09 +4.8% 3.959e+09 ± 2% perf-stat.node-store-misses
5.079e+09 +6.4% 5.402e+09 perf-stat.node-stores
865565 +10.3% 954352 perf-stat.page-faults
51.75 ± 5% -12.5% 45.30 ± 10% sched_debug.cfs_rq:/.load_avg.avg
316.35 ± 3% +17.2% 370.81 ± 8% sched_debug.cfs_rq:/.util_est_enqueued.stddev
15294 ± 30% +234.9% 51219 ± 76% sched_debug.cpu.avg_idle.min
299443 ± 3% -7.3% 277566 ± 5% sched_debug.cpu.avg_idle.stddev
1182 ± 19% -26.3% 872.02 ± 13% sched_debug.cpu.nr_load_updates.stddev
1.22 ± 8% +21.7% 1.48 ± 6% sched_debug.cpu.nr_running.avg
2.75 ± 10% +26.2% 3.47 ± 6% sched_debug.cpu.nr_running.max
0.58 ± 7% +24.2% 0.73 ± 6% sched_debug.cpu.nr_running.stddev
77148 -20.0% 61702 ± 7% sched_debug.cpu.nr_switches.avg
70024 -24.8% 52647 ± 8% sched_debug.cpu.nr_switches.min
6662 ± 6% +61.9% 10789 ± 24% sched_debug.cpu.nr_switches.stddev
80.45 ± 18% -19.1% 65.05 ± 6% sched_debug.cpu.nr_uninterruptible.stddev
76819 -19.3% 62008 ± 8% sched_debug.cpu.sched_count.avg
70616 -23.5% 53996 ± 8% sched_debug.cpu.sched_count.min
5494 ± 9% +85.3% 10179 ± 26% sched_debug.cpu.sched_count.stddev
16936 -52.9% 7975 ± 9% sched_debug.cpu.sched_goidle.avg
19281 -49.9% 9666 ± 7% sched_debug.cpu.sched_goidle.max
15417 -54.8% 6962 ± 10% sched_debug.cpu.sched_goidle.min
875.00 ± 6% -35.0% 569.09 ± 13% sched_debug.cpu.sched_goidle.stddev
40332 -23.5% 30851 ± 7% sched_debug.cpu.ttwu_count.avg
35074 -26.3% 25833 ± 6% sched_debug.cpu.ttwu_count.min
3239 ± 8% +67.4% 5422 ± 28% sched_debug.cpu.ttwu_count.stddev
5232 +27.4% 6665 ± 13% sched_debug.cpu.ttwu_local.avg
15877 ± 12% +77.5% 28184 ± 27% sched_debug.cpu.ttwu_local.max
2530 ± 10% +95.9% 4956 ± 27% sched_debug.cpu.ttwu_local.stddev
2.52 ± 7% -0.6 1.95 ± 3% perf-profile.calltrace.cycles-pp.btrfs_dirty_pages.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write
1.48 ± 12% -0.5 1.01 ± 4% perf-profile.calltrace.cycles-pp.btrfs_get_extent.btrfs_dirty_pages.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write
1.18 ± 16% -0.4 0.76 ± 7% perf-profile.calltrace.cycles-pp.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_get_extent.btrfs_dirty_pages.__btrfs_buffered_write
1.18 ± 16% -0.4 0.76 ± 7% perf-profile.calltrace.cycles-pp.btrfs_lookup_file_extent.btrfs_get_extent.btrfs_dirty_pages.__btrfs_buffered_write.btrfs_file_write_iter
0.90 ± 17% -0.3 0.56 ± 4% perf-profile.calltrace.cycles-pp.__dentry_kill.dentry_kill.dput.__fput.task_work_run
0.90 ± 17% -0.3 0.56 ± 4% perf-profile.calltrace.cycles-pp.evict.__dentry_kill.dentry_kill.dput.__fput
0.90 ± 17% -0.3 0.56 ± 4% perf-profile.calltrace.cycles-pp.dentry_kill.dput.__fput.task_work_run.exit_to_usermode_loop
0.90 ± 18% -0.3 0.56 ± 4% perf-profile.calltrace.cycles-pp.btrfs_evict_inode.evict.__dentry_kill.dentry_kill.dput
0.90 ± 17% -0.3 0.57 ± 5% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.90 ± 17% -0.3 0.57 ± 5% perf-profile.calltrace.cycles-pp.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.90 ± 17% -0.3 0.57 ± 5% perf-profile.calltrace.cycles-pp.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.90 ± 17% -0.3 0.57 ± 5% perf-profile.calltrace.cycles-pp.dput.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64
1.69 -0.1 1.54 ± 2% perf-profile.calltrace.cycles-pp.lock_and_cleanup_extent_if_need.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write
0.87 ± 4% -0.1 0.76 ± 2% perf-profile.calltrace.cycles-pp.__clear_extent_bit.clear_extent_bit.lock_and_cleanup_extent_if_need.__btrfs_buffered_write.btrfs_file_write_iter
0.87 ± 4% -0.1 0.76 ± 2% perf-profile.calltrace.cycles-pp.clear_extent_bit.lock_and_cleanup_extent_if_need.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write
0.71 ± 6% -0.1 0.61 ± 2% perf-profile.calltrace.cycles-pp.clear_state_bit.__clear_extent_bit.clear_extent_bit.lock_and_cleanup_extent_if_need.__btrfs_buffered_write
0.69 ± 6% -0.1 0.60 ± 2% perf-profile.calltrace.cycles-pp.btrfs_clear_bit_hook.clear_state_bit.__clear_extent_bit.clear_extent_bit.lock_and_cleanup_extent_if_need
96.77 +0.6 97.33 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
0.00 +0.6 0.56 ± 3% perf-profile.calltrace.cycles-pp.can_overcommit.reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write.btrfs_file_write_iter
96.72 +0.6 97.29 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
43.13 +0.8 43.91 perf-profile.calltrace.cycles-pp.btrfs_inode_rsv_release.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write
42.37 +0.8 43.16 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.block_rsv_release_bytes.btrfs_inode_rsv_release.__btrfs_buffered_write
43.11 +0.8 43.89 perf-profile.calltrace.cycles-pp.block_rsv_release_bytes.btrfs_inode_rsv_release.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write
42.96 +0.8 43.77 perf-profile.calltrace.cycles-pp._raw_spin_lock.block_rsv_release_bytes.btrfs_inode_rsv_release.__btrfs_buffered_write.btrfs_file_write_iter
95.28 +0.9 96.23 perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
95.22 +1.0 96.18 perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
94.88 +1.0 95.85 perf-profile.calltrace.cycles-pp.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
94.83 +1.0 95.80 perf-profile.calltrace.cycles-pp.btrfs_file_write_iter.__vfs_write.vfs_write.ksys_write.do_syscall_64
94.51 +1.0 95.50 perf-profile.calltrace.cycles-pp.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write.ksys_write
42.44 +1.1 43.52 perf-profile.calltrace.cycles-pp._raw_spin_lock.reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write.btrfs_file_write_iter
42.09 +1.1 43.18 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write
44.07 +1.2 45.29 perf-profile.calltrace.cycles-pp.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write
43.42 +1.3 44.69 perf-profile.calltrace.cycles-pp.reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write
2.06 ± 18% -0.9 1.21 ± 6% perf-profile.children.cycles-pp.btrfs_search_slot
2.54 ± 7% -0.6 1.96 ± 3% perf-profile.children.cycles-pp.btrfs_dirty_pages
1.05 ± 24% -0.5 0.52 ± 9% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
1.50 ± 12% -0.5 1.03 ± 4% perf-profile.children.cycles-pp.btrfs_get_extent
1.22 ± 15% -0.4 0.79 ± 8% perf-profile.children.cycles-pp.btrfs_lookup_file_extent
0.81 ± 5% -0.4 0.41 ± 6% perf-profile.children.cycles-pp.btrfs_calc_reclaim_metadata_size
0.74 ± 24% -0.4 0.35 ± 9% perf-profile.children.cycles-pp.btrfs_lock_root_node
0.74 ± 24% -0.4 0.35 ± 9% perf-profile.children.cycles-pp.btrfs_tree_lock
0.90 ± 17% -0.3 0.56 ± 4% perf-profile.children.cycles-pp.__dentry_kill
0.90 ± 17% -0.3 0.56 ± 4% perf-profile.children.cycles-pp.evict
0.90 ± 17% -0.3 0.56 ± 4% perf-profile.children.cycles-pp.dentry_kill
0.90 ± 18% -0.3 0.56 ± 4% perf-profile.children.cycles-pp.btrfs_evict_inode
0.91 ± 18% -0.3 0.57 ± 4% perf-profile.children.cycles-pp.exit_to_usermode_loop
0.52 ± 20% -0.3 0.18 ± 14% perf-profile.children.cycles-pp.do_idle
0.90 ± 17% -0.3 0.57 ± 5% perf-profile.children.cycles-pp.task_work_run
0.90 ± 17% -0.3 0.57 ± 5% perf-profile.children.cycles-pp.__fput
0.90 ± 18% -0.3 0.57 ± 4% perf-profile.children.cycles-pp.dput
0.51 ± 20% -0.3 0.18 ± 14% perf-profile.children.cycles-pp.secondary_startup_64
0.51 ± 20% -0.3 0.18 ± 14% perf-profile.children.cycles-pp.cpu_startup_entry
0.50 ± 21% -0.3 0.17 ± 16% perf-profile.children.cycles-pp.start_secondary
0.47 ± 20% -0.3 0.16 ± 13% perf-profile.children.cycles-pp.cpuidle_enter_state
0.47 ± 19% -0.3 0.16 ± 13% perf-profile.children.cycles-pp.intel_idle
0.61 ± 20% -0.3 0.36 ± 11% perf-profile.children.cycles-pp.btrfs_tree_read_lock
0.47 ± 26% -0.3 0.21 ± 10% perf-profile.children.cycles-pp.prepare_to_wait_event
0.64 ± 18% -0.2 0.39 ± 9% perf-profile.children.cycles-pp.btrfs_read_lock_root_node
0.40 ± 22% -0.2 0.21 ± 5% perf-profile.children.cycles-pp.btrfs_clear_path_blocking
0.38 ± 23% -0.2 0.19 ± 13% perf-profile.children.cycles-pp.finish_wait
1.51 ± 3% -0.2 1.35 ± 2% perf-profile.children.cycles-pp.__clear_extent_bit
1.71 -0.1 1.56 ± 2% perf-profile.children.cycles-pp.lock_and_cleanup_extent_if_need
0.29 ± 25% -0.1 0.15 ± 10% perf-profile.children.cycles-pp.btrfs_orphan_del
0.27 ± 27% -0.1 0.12 ± 8% perf-profile.children.cycles-pp.btrfs_del_orphan_item
0.33 ± 18% -0.1 0.19 ± 9% perf-profile.children.cycles-pp.queued_read_lock_slowpath
0.33 ± 19% -0.1 0.20 ± 4% perf-profile.children.cycles-pp.__wake_up_common_lock
0.45 ± 15% -0.1 0.34 ± 2% perf-profile.children.cycles-pp.btrfs_alloc_data_chunk_ondemand
0.47 ± 16% -0.1 0.36 ± 4% perf-profile.children.cycles-pp.btrfs_check_data_free_space
0.91 ± 4% -0.1 0.81 ± 3% perf-profile.children.cycles-pp.clear_extent_bit
1.07 ± 5% -0.1 0.97 perf-profile.children.cycles-pp.__set_extent_bit
0.77 ± 6% -0.1 0.69 ± 3% perf-profile.children.cycles-pp.btrfs_clear_bit_hook
0.17 ± 20% -0.1 0.08 ± 10% perf-profile.children.cycles-pp.queued_write_lock_slowpath
0.16 ± 22% -0.1 0.08 ± 24% perf-profile.children.cycles-pp.btrfs_lookup_inode
0.21 ± 17% -0.1 0.14 ± 19% perf-profile.children.cycles-pp.__btrfs_update_delayed_inode
0.26 ± 12% -0.1 0.18 ± 13% perf-profile.children.cycles-pp.btrfs_async_run_delayed_root
0.52 ± 5% -0.1 0.45 perf-profile.children.cycles-pp.set_extent_bit
0.45 ± 5% -0.1 0.40 ± 3% perf-profile.children.cycles-pp.alloc_extent_state
0.11 ± 17% -0.1 0.06 ± 11% perf-profile.children.cycles-pp.btrfs_clear_lock_blocking_rw
0.28 ± 9% -0.0 0.23 ± 3% perf-profile.children.cycles-pp.btrfs_drop_pages
0.07 -0.0 0.03 ±100% perf-profile.children.cycles-pp.btrfs_set_lock_blocking_rw
0.39 ± 3% -0.0 0.34 ± 3% perf-profile.children.cycles-pp.get_alloc_profile
0.33 ± 7% -0.0 0.29 perf-profile.children.cycles-pp.btrfs_set_extent_delalloc
0.38 ± 2% -0.0 0.35 ± 4% perf-profile.children.cycles-pp.__set_page_dirty_nobuffers
0.49 ± 3% -0.0 0.46 ± 3% perf-profile.children.cycles-pp.pagecache_get_page
0.18 ± 4% -0.0 0.15 ± 2% perf-profile.children.cycles-pp.truncate_inode_pages_range
0.08 ± 5% -0.0 0.05 ± 9% perf-profile.children.cycles-pp.btrfs_set_path_blocking
0.08 ± 6% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.truncate_cleanup_page
0.80 ± 4% +0.2 0.95 ± 2% perf-profile.children.cycles-pp.can_overcommit
96.84 +0.5 97.37 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
96.80 +0.5 97.35 perf-profile.children.cycles-pp.do_syscall_64
43.34 +0.8 44.17 perf-profile.children.cycles-pp.btrfs_inode_rsv_release
43.49 +0.8 44.32 perf-profile.children.cycles-pp.block_rsv_release_bytes
95.32 +0.9 96.26 perf-profile.children.cycles-pp.ksys_write
95.26 +0.9 96.20 perf-profile.children.cycles-pp.vfs_write
94.91 +1.0 95.88 perf-profile.children.cycles-pp.__vfs_write
94.84 +1.0 95.81 perf-profile.children.cycles-pp.btrfs_file_write_iter
94.55 +1.0 95.55 perf-profile.children.cycles-pp.__btrfs_buffered_write
86.68 +1.0 87.70 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
44.08 +1.2 45.31 perf-profile.children.cycles-pp.btrfs_delalloc_reserve_metadata
43.49 +1.3 44.77 perf-profile.children.cycles-pp.reserve_metadata_bytes
87.59 +1.8 89.38 perf-profile.children.cycles-pp._raw_spin_lock
0.47 ± 19% -0.3 0.16 ± 13% perf-profile.self.cycles-pp.intel_idle
0.33 ± 6% -0.1 0.18 ± 6% perf-profile.self.cycles-pp.get_alloc_profile
0.27 ± 8% -0.0 0.22 ± 4% perf-profile.self.cycles-pp.btrfs_drop_pages
0.07 -0.0 0.03 ±100% perf-profile.self.cycles-pp.btrfs_set_lock_blocking_rw
0.14 ± 5% -0.0 0.12 ± 6% perf-profile.self.cycles-pp.clear_page_dirty_for_io
0.09 ± 5% -0.0 0.07 ± 10% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.17 ± 4% +0.1 0.23 ± 3% perf-profile.self.cycles-pp.reserve_metadata_bytes
0.31 ± 7% +0.1 0.45 ± 2% perf-profile.self.cycles-pp.can_overcommit
86.35 +1.0 87.39 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
aim7.jobs-per-min
29000 +-+-----------------------------------------------------------------+
28500 +-+ +.. + +..+.. +.. |
|..+ +.+..+.. : .. + .+.+..+..+.+.. .+..+.. + + + |
28000 +-+ + .. : + +. + + + |
27500 +-+ + + |
| |
27000 +-+ |
26500 +-+ |
26000 +-+ |
| |
25500 +-+ O O O O O |
25000 +-+ O O O O O O O O O
| O O O O O O O O |
24500 O-+O O O O |
24000 +-+-----------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Xiaolong
3 years, 6 months
[lkp-robot] [sched/fair] d519329f72: unixbench.score -9.9% regression
by kernel test robot
Greeting,
FYI, we noticed a -9.9% regression of unixbench.score due to commit:
commit: d519329f72a6f36bc4f2b85452640cfe583b4f81 ("sched/fair: Update util_est only on util_avg updates")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
in testcase: unixbench
on test machine: 8 threads Intel(R) Core(TM) i7 CPU 870 @ 2.93GHz with 6G memory
with following parameters:
runtime: 300s
nr_task: 100%
test: execl
test-description: UnixBench is the original BYTE UNIX benchmark suite aims to test performance of Unix-like system.
test-url: https://github.com/kdlucas/byte-unixbench
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
compiler/kconfig/nr_task/rootfs/runtime/tbox_group/test/testcase:
gcc-7/x86_64-rhel-7.2/100%/debian-x86_64-2016-08-31.cgz/300s/nhm-white/execl/unixbench
commit:
a07630b8b2 ("sched/cpufreq/schedutil: Use util_est for OPP selection")
d519329f72 ("sched/fair: Update util_est only on util_avg updates")
a07630b8b2c16f82 d519329f72a6f36bc4f2b85452
---------------- --------------------------
%stddev %change %stddev
\ | \
4626 -9.9% 4167 unixbench.score
3495362 ± 4% +70.4% 5957769 ± 2% unixbench.time.involuntary_context_switches
2.866e+08 -11.6% 2.534e+08 unixbench.time.minor_page_faults
666.75 -9.7% 602.25 unixbench.time.percent_of_cpu_this_job_got
1830 -9.7% 1653 unixbench.time.system_time
395.13 -5.2% 374.58 unixbench.time.user_time
8611715 -58.9% 3537314 ± 3% unixbench.time.voluntary_context_switches
6639375 -9.1% 6033775 unixbench.workload
26025 +3849.3% 1027825 interrupts.CAL:Function_call_interrupts
4856 ± 14% -27.4% 3523 ± 11% slabinfo.filp.active_objs
3534356 -8.8% 3223918 softirqs.RCU
77929 -11.2% 69172 vmstat.system.cs
19489 ± 2% +7.5% 20956 vmstat.system.in
9.05 ± 9% +11.0% 10.05 ± 8% boot-time.dhcp
131.63 ± 4% +8.6% 142.89 ± 7% boot-time.idle
9.07 ± 9% +11.0% 10.07 ± 8% boot-time.kernel_boot
76288 ± 3% -12.8% 66560 ± 3% meminfo.DirectMap4k
16606 -13.1% 14433 meminfo.Inactive
16515 -13.2% 14341 meminfo.Inactive(anon)
11.87 ± 5% +7.8 19.63 ± 4% mpstat.cpu.idle%
0.07 ± 35% -0.0 0.04 ± 17% mpstat.cpu.soft%
68.91 -6.1 62.82 mpstat.cpu.sys%
29291570 +325.4% 1.246e+08 cpuidle.C1.time
8629105 -36.1% 5513780 cpuidle.C1.usage
668733 ± 12% +11215.3% 75668902 ± 2% cpuidle.C1E.time
9763 ± 12% +16572.7% 1627882 ± 2% cpuidle.C1E.usage
1.834e+08 ± 9% +23.1% 2.258e+08 ± 11% cpuidle.C3.time
222674 ± 8% +133.4% 519690 ± 6% cpuidle.C3.usage
4129 -13.3% 3581 proc-vmstat.nr_inactive_anon
4129 -13.3% 3581 proc-vmstat.nr_zone_inactive_anon
2.333e+08 -12.2% 2.049e+08 proc-vmstat.numa_hit
2.333e+08 -12.2% 2.049e+08 proc-vmstat.numa_local
6625 -10.9% 5905 proc-vmstat.pgactivate
2.392e+08 -12.1% 2.102e+08 proc-vmstat.pgalloc_normal
2.936e+08 -12.6% 2.566e+08 proc-vmstat.pgfault
2.392e+08 -12.1% 2.102e+08 proc-vmstat.pgfree
2850 -15.3% 2413 turbostat.Avg_MHz
8629013 -36.1% 5513569 turbostat.C1
1.09 +3.5 4.61 turbostat.C1%
9751 ± 12% +16593.0% 1627864 ± 2% turbostat.C1E
0.03 ± 19% +2.8 2.80 turbostat.C1E%
222574 ± 8% +133.4% 519558 ± 6% turbostat.C3
6.84 ± 8% +1.5 8.34 ± 10% turbostat.C3%
2.82 ± 7% +250.3% 9.87 ± 2% turbostat.CPU%c1
6552773 ± 3% +23.8% 8111699 ± 2% turbostat.IRQ
2.02 ± 11% +28.3% 2.58 ± 9% turbostat.Pkg%pc3
7.635e+11 -12.5% 6.682e+11 perf-stat.branch-instructions
3.881e+10 -12.9% 3.381e+10 perf-stat.branch-misses
2.09 -0.3 1.77 ± 4% perf-stat.cache-miss-rate%
1.551e+09 -15.1% 1.316e+09 ± 4% perf-stat.cache-misses
26177920 -10.5% 23428188 perf-stat.context-switches
1.99 -2.8% 1.93 perf-stat.cpi
7.553e+12 -14.7% 6.446e+12 perf-stat.cpu-cycles
522523 ± 2% +628.3% 3805664 perf-stat.cpu-migrations
2.425e+10 ± 4% -14.3% 2.078e+10 perf-stat.dTLB-load-misses
1.487e+12 -11.3% 1.319e+12 perf-stat.dTLB-loads
1.156e+10 ± 3% -7.7% 1.066e+10 perf-stat.dTLB-store-misses
6.657e+11 -11.1% 5.915e+11 perf-stat.dTLB-stores
0.15 +0.0 0.15 perf-stat.iTLB-load-miss-rate%
5.807e+09 -11.0% 5.166e+09 perf-stat.iTLB-load-misses
3.799e+12 -12.1% 3.34e+12 perf-stat.iTLB-loads
3.803e+12 -12.2% 3.338e+12 perf-stat.instructions
654.99 -1.4% 646.07 perf-stat.instructions-per-iTLB-miss
0.50 +2.8% 0.52 perf-stat.ipc
2.754e+08 -11.6% 2.435e+08 perf-stat.minor-faults
1.198e+08 ± 7% +73.1% 2.074e+08 ± 4% perf-stat.node-stores
2.754e+08 -11.6% 2.435e+08 perf-stat.page-faults
572928 -3.4% 553258 perf-stat.path-length
unixbench.score
4800 +-+------------------------------------------------------------------+
|+ + + |
4700 +-+ + + :+ +. :+ + + |
| + + + +. : + + + + + + + .+++++ .+ +|
4600 +-+ +++ :+++ + ++: : :+ +++ ++.++++ + ++++ ++ |
| + + + ++ ++ + |
4500 +-+ |
| |
4400 +-+ |
| |
4300 +-+ |
O |
4200 +-O O O OOOO OO OOO OOOO OOOO O O |
|O OO OOOOO O O OO O O O O O OO |
4100 +-+------------------------------------------------------------------+
unixbench.workload
9e+06 +-+---------------------------------------------------------------+
| : |
8.5e+06 +-+ : |
| : |
8e+06 +-+ : |
| :: |
7.5e+06 +-+ : : + |
| +: : : + |
7e+06 +-+ + + :: : :: + + : + + + + + |
|:+ + + : :: : : :: : :+ : : ::+ :+ .+ :+ ++ ++ + ++ ::++|
6.5e+06 +-O+ +++ ++++ +++ + ++ +.+ + ++ + + + + + + + +.+++ + |
O O O O O O O |
6e+06 +O+OOO O OOOOOOOO OOOO OO OOOOOOOOO O O O OO |
| O |
5.5e+06 +-+---------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Xiaolong
3 years, 8 months
[mailbox] afd0b1fb22: kmsg.Error_parsing_PCC_subspaces_from_PCCT
by kernel test robot
FYI, we noticed the following commit (built with gcc-7):
commit: afd0b1fb22269f48d68fdf269891c653818c8047 ("mailbox: PCC: handle parse error")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
in testcase: ftq
with following parameters:
nr_task: 100%
samples: 6000ss
test: cache
freq: 20
ucode: 0x200004d
cpufreq_governor: performance
test-description: The FTQ benchmarks measure hardware and software interference or 'noise' on a node from the applications perspective.
test-url: https://github.com/rminnich/ftq
on test machine: 104 threads Intel(R) Xeon(R) Platinum 8170 CPU @ 2.10GHz with 192G memory
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
2018-09-21 23:31:26 ./ftq_cache -f 20 -n 6000 -t 104 -a 360448
ticks per ns 2.09508
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Thanks,
Rong Chen
3 years, 8 months
[mm/swap] d884021f52: will-it-scale.per_process_ops -2.4% regression
by kernel test robot
Greeting,
FYI, we noticed a -2.4% regression of will-it-scale.per_process_ops due to commit:
commit: d884021f52609407c7943705b3e54b1642fa10cb ("[PATCH 1/2] mm/swap: Add pagevec locking")
url: https://github.com/0day-ci/linux/commits/Sebastian-Andrzej-Siewior/mm-swa...
in testcase: will-it-scale
on test machine: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory
with following parameters:
nr_task: 50%
mode: process
test: brk1
cpufreq_governor: performance
test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
test-url: https://github.com/antonblanchard/will-it-scale
In addition to that, the commit also has significant impact on the following tests:
+------------------+-----------------------------------------------------------------------+
| testcase: change | fio-basic: |
| test machine | 48 threads Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz with 64G memory |
| test parameters | bs=4k |
| | cpufreq_governor=performance |
| | disk=1HDD |
| | fs2=nfsv4 |
| | fs=ext4 |
| | ioengine=sync |
| | nr_task=100% |
| | runtime=300s |
| | rw=write |
| | test_size=128G |
| | ucode=0x42d |
+------------------+-----------------------------------------------------------------------+
| testcase: change | fio-basic: fio.latency_100us% 1.6% improvement |
| test machine | 72 threads Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz with 128G memory |
| test parameters | bs=4k |
| | cpufreq_governor=performance |
| | disk=1SSD |
| | fs=ext4 |
| | ioengine=sync |
| | nr_task=64 |
| | runtime=300s |
| | rw=randwrite |
| | test_size=512g |
| | ucode=0x3d |
+------------------+-----------------------------------------------------------------------+
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-7/performance/x86_64-rhel-7.2/process/50%/debian-x86_64-2018-04-03.cgz/lkp-bdw-ep3d/brk1/will-it-scale
commit:
f3c0b8ce48 (" dmaengine fix for dmaengine-fix-4.19-rc4")
d884021f52 ("mm/swap: Add pagevec locking")
f3c0b8ce48408b74 d884021f52609407c7943705b3
---------------- --------------------------
fail:runs %reproduction fail:runs
| | |
:4 25% 1:4 dmesg.WARNING:at#for_ip_interrupt_entry/0x
:4 75% 3:4 dmesg.WARNING:at_ip_perf_event_mmap_output/0x
%stddev %change %stddev
\ | \
743536 -2.4% 725610 will-it-scale.per_process_ops
32715630 -2.4% 31926873 will-it-scale.workload
0.00 ± 71% +0.0 0.01 ± 34% mpstat.cpu.soft%
2457 ± 5% +12.2% 2757 ± 5% slabinfo.kmalloc-512.active_objs
67.33 ± 9% -17.3% 55.67 ± 4% sched_debug.cpu.cpu_load[0].max
10817 ± 11% -16.7% 9008 ± 5% sched_debug.cpu.load.stddev
1.26 -0.0 1.23 perf-stat.branch-miss-rate%
4.635e+10 -3.0% 4.496e+10 perf-stat.branch-misses
3.046e+12 +1.7% 3.097e+12 perf-stat.dTLB-stores
11.55 ± 2% +11.5 23.10 ± 6% perf-stat.iTLB-load-miss-rate%
1.57e+09 +125.1% 3.534e+09 ± 7% perf-stat.iTLB-load-misses
11183 -55.4% 4991 ± 7% perf-stat.instructions-per-iTLB-miss
55.41 ± 19% +21.4 76.78 ± 15% perf-stat.node-store-miss-rate%
3387268 ± 21% -36.2% 2161442 ± 7% perf-stat.node-stores
536500 +2.4% 549279 perf-stat.path-length
31.36 -1.1 30.24 perf-profile.calltrace.cycles-pp.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
8.85 -0.6 8.21 perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
7.70 -0.4 7.33 perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk
11.19 -0.2 10.96 perf-profile.calltrace.cycles-pp.syscall_return_via_sysret
5.01 -0.2 4.79 perf-profile.calltrace.cycles-pp.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
3.41 -0.2 3.21 perf-profile.calltrace.cycles-pp.down_write_killable.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
5.23 -0.2 5.05 perf-profile.calltrace.cycles-pp.find_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
4.62 -0.2 4.44 perf-profile.calltrace.cycles-pp.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.67 -0.2 1.48 perf-profile.calltrace.cycles-pp.find_vma.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
3.59 -0.2 3.42 perf-profile.calltrace.cycles-pp.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
3.23 -0.1 3.08 perf-profile.calltrace.cycles-pp.selinux_vm_enough_memory.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk.do_syscall_64
2.05 -0.1 1.91 perf-profile.calltrace.cycles-pp.vmacache_find.find_vma.do_munmap.__x64_sys_brk.do_syscall_64
3.94 -0.1 3.79 perf-profile.calltrace.cycles-pp.vm_area_alloc.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
7.96 -0.1 7.84 perf-profile.calltrace.cycles-pp.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
2.42 -0.1 2.30 perf-profile.calltrace.cycles-pp.cred_has_capability.selinux_vm_enough_memory.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk
2.50 -0.1 2.38 perf-profile.calltrace.cycles-pp.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
2.33 -0.1 2.23 ± 2% perf-profile.calltrace.cycles-pp.kmem_cache_alloc.vm_area_alloc.do_brk_flags.__x64_sys_brk.do_syscall_64
1.16 -0.1 1.07 ± 2% perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
2.25 -0.1 2.15 perf-profile.calltrace.cycles-pp.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64
1.49 -0.1 1.40 perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk
1.06 -0.1 0.97 perf-profile.calltrace.cycles-pp.__vm_enough_memory.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.63 -0.1 1.55 perf-profile.calltrace.cycles-pp.avc_has_perm_noaudit.cred_has_capability.selinux_vm_enough_memory.security_vm_enough_memory_mm.do_brk_flags
1.15 -0.1 1.07 ± 2% perf-profile.calltrace.cycles-pp.up_write.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.91 -0.1 0.84 ± 2% perf-profile.calltrace.cycles-pp.tlb_gather_mmu.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
0.88 ± 3% -0.1 0.81 perf-profile.calltrace.cycles-pp.___might_sleep.down_write_killable.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.74 ± 2% -0.1 0.68 perf-profile.calltrace.cycles-pp.free_pgd_range.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
1.16 -0.1 1.10 perf-profile.calltrace.cycles-pp.vma_merge.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.27 -0.1 1.22 ± 2% perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
1.13 -0.1 1.08 perf-profile.calltrace.cycles-pp.__vma_rb_erase.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.37 ± 2% -0.0 1.32 perf-profile.calltrace.cycles-pp.cap_vm_enough_memory.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk.do_syscall_64
1.58 -0.0 1.54 perf-profile.calltrace.cycles-pp.arch_get_unmapped_area_topdown.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64
0.94 -0.0 0.90 perf-profile.calltrace.cycles-pp.kmem_cache_free.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64
1.14 -0.0 1.10 perf-profile.calltrace.cycles-pp.perf_event_mmap_output.perf_iterate_sb.perf_event_mmap.do_brk_flags.__x64_sys_brk
0.73 -0.0 0.69 ± 2% perf-profile.calltrace.cycles-pp.cap_mmap_addr.security_mmap_addr.get_unmapped_area.do_brk_flags.__x64_sys_brk
0.96 -0.0 0.93 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_stage2
74.01 +0.1 74.13 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
72.69 +0.2 72.89 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
69.82 +0.3 70.16 perf-profile.calltrace.cycles-pp.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +0.7 0.73 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.lru_add_drain_cpu.unmap_region.do_munmap.__x64_sys_brk
0.00 +0.8 0.83 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_unlock_irqrestore.lru_add_drain_cpu.unmap_region.do_munmap.__x64_sys_brk
0.00 +2.2 2.23 perf-profile.calltrace.cycles-pp._raw_spin_lock.lru_add_drain_cpu.unmap_region.do_munmap.__x64_sys_brk
27.94 +2.4 30.30 perf-profile.calltrace.cycles-pp.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
14.50 +2.9 17.43 perf-profile.calltrace.cycles-pp.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +4.4 4.43 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
31.57 -1.1 30.43 perf-profile.children.cycles-pp.do_brk_flags
0.82 ± 3% -0.8 0.07 perf-profile.children.cycles-pp.lru_add_drain
8.91 -0.7 8.26 perf-profile.children.cycles-pp.unmap_vmas
7.98 -0.4 7.59 perf-profile.children.cycles-pp.unmap_page_range
7.18 -0.4 6.81 perf-profile.children.cycles-pp.find_vma
12.90 -0.2 12.65 perf-profile.children.cycles-pp.syscall_return_via_sysret
3.27 -0.2 3.03 perf-profile.children.cycles-pp.vma_compute_subtree_gap
0.60 ± 3% -0.2 0.37 ± 3% perf-profile.children.cycles-pp.unmap_single_vma
3.62 ± 2% -0.2 3.40 perf-profile.children.cycles-pp.down_write_killable
5.06 -0.2 4.84 perf-profile.children.cycles-pp.security_vm_enough_memory_mm
1.71 -0.2 1.52 perf-profile.children.cycles-pp._cond_resched
4.74 -0.2 4.55 perf-profile.children.cycles-pp.get_unmapped_area
2.41 -0.2 2.25 ± 2% perf-profile.children.cycles-pp.vmacache_find
3.74 -0.2 3.58 perf-profile.children.cycles-pp.vma_link
3.27 -0.1 3.12 perf-profile.children.cycles-pp.selinux_vm_enough_memory
3.95 -0.1 3.80 ± 2% perf-profile.children.cycles-pp.vm_area_alloc
8.28 -0.1 8.14 perf-profile.children.cycles-pp.perf_event_mmap
2.94 -0.1 2.81 perf-profile.children.cycles-pp.___might_sleep
2.47 -0.1 2.35 perf-profile.children.cycles-pp.__vma_link_rb
2.43 -0.1 2.31 perf-profile.children.cycles-pp.cred_has_capability
0.34 ± 2% -0.1 0.22 ± 5% perf-profile.children.cycles-pp.arch_tlb_gather_mmu
2.64 -0.1 2.53 perf-profile.children.cycles-pp.remove_vma
0.62 ± 3% -0.1 0.53 ± 4% perf-profile.children.cycles-pp.vmacache_update
1.14 -0.1 1.04 perf-profile.children.cycles-pp.__vm_enough_memory
0.87 ± 2% -0.1 0.79 ± 3% perf-profile.children.cycles-pp.rcu_all_qs
2.46 -0.1 2.37 ± 2% perf-profile.children.cycles-pp.kmem_cache_alloc
1.22 -0.1 1.14 ± 2% perf-profile.children.cycles-pp.up_write
0.92 -0.1 0.84 ± 2% perf-profile.children.cycles-pp.tlb_gather_mmu
1.63 -0.1 1.55 perf-profile.children.cycles-pp.avc_has_perm_noaudit
1.26 ± 2% -0.1 1.19 perf-profile.children.cycles-pp.__vma_rb_erase
0.78 ± 2% -0.1 0.72 perf-profile.children.cycles-pp.free_pgd_range
1.19 -0.1 1.12 perf-profile.children.cycles-pp.vma_merge
1.41 ± 2% -0.1 1.35 perf-profile.children.cycles-pp.cap_vm_enough_memory
0.82 -0.1 0.77 perf-profile.children.cycles-pp.percpu_counter_add_batch
1.30 -0.0 1.25 ± 2% perf-profile.children.cycles-pp.tlb_finish_mmu
1.64 -0.0 1.59 perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown
0.95 -0.0 0.90 perf-profile.children.cycles-pp.kmem_cache_free
0.44 -0.0 0.40 perf-profile.children.cycles-pp.strlen
1.22 -0.0 1.18 perf-profile.children.cycles-pp.perf_event_mmap_output
0.53 ± 3% -0.0 0.49 ± 3% perf-profile.children.cycles-pp.cap_capable
0.96 ± 2% -0.0 0.93 perf-profile.children.cycles-pp.entry_SYSCALL_64_stage2
0.37 ± 2% -0.0 0.35 ± 4% perf-profile.children.cycles-pp.tlb_flush_mmu_free
74.20 +0.1 74.31 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
72.84 +0.2 73.04 perf-profile.children.cycles-pp.do_syscall_64
70.18 +0.3 70.48 perf-profile.children.cycles-pp.__x64_sys_brk
0.00 +0.8 0.84 ± 3% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
0.00 +1.0 0.97 perf-profile.children.cycles-pp._raw_spin_lock_irqsave
0.72 +2.3 3.02 perf-profile.children.cycles-pp._raw_spin_lock
28.33 +2.3 30.68 perf-profile.children.cycles-pp.do_munmap
14.65 +2.9 17.56 perf-profile.children.cycles-pp.unmap_region
0.66 ± 4% +3.9 4.55 perf-profile.children.cycles-pp.lru_add_drain_cpu
3.19 -0.4 2.78 ± 2% perf-profile.self.cycles-pp.__x64_sys_brk
5.09 -0.4 4.71 perf-profile.self.cycles-pp.unmap_page_range
12.87 -0.2 12.62 perf-profile.self.cycles-pp.syscall_return_via_sysret
0.55 ± 4% -0.2 0.34 ± 4% perf-profile.self.cycles-pp.unmap_single_vma
2.81 -0.2 2.64 perf-profile.self.cycles-pp.vma_compute_subtree_gap
2.28 -0.2 2.12 perf-profile.self.cycles-pp.vmacache_find
2.84 -0.1 2.70 perf-profile.self.cycles-pp.___might_sleep
0.80 -0.1 0.69 ± 2% perf-profile.self.cycles-pp._cond_resched
3.24 -0.1 3.12 perf-profile.self.cycles-pp.do_brk_flags
0.33 ± 3% -0.1 0.22 ± 5% perf-profile.self.cycles-pp.arch_tlb_gather_mmu
4.25 -0.1 4.14 perf-profile.self.cycles-pp.find_vma
2.52 -0.1 2.41 perf-profile.self.cycles-pp.do_munmap
2.43 ± 2% -0.1 2.33 perf-profile.self.cycles-pp.do_syscall_64
1.42 -0.1 1.32 perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
0.56 ± 3% -0.1 0.47 ± 5% perf-profile.self.cycles-pp.vmacache_update
1.62 -0.1 1.54 perf-profile.self.cycles-pp.avc_has_perm_noaudit
1.14 -0.1 1.07 perf-profile.self.cycles-pp.up_write
1.14 -0.1 1.07 perf-profile.self.cycles-pp.vma_merge
0.77 ± 2% -0.1 0.71 ± 2% perf-profile.self.cycles-pp.free_pgd_range
0.78 -0.1 0.73 ± 2% perf-profile.self.cycles-pp.security_mmap_addr
0.77 -0.1 0.72 ± 3% perf-profile.self.cycles-pp.get_unmapped_area
0.77 -0.1 0.72 ± 4% perf-profile.self.cycles-pp.cred_has_capability
0.39 ± 2% -0.0 0.34 ± 3% perf-profile.self.cycles-pp.unmap_vmas
1.44 ± 2% -0.0 1.39 perf-profile.self.cycles-pp.vm_area_alloc
0.89 -0.0 0.84 perf-profile.self.cycles-pp.__vma_rb_erase
0.78 -0.0 0.73 perf-profile.self.cycles-pp.__vm_enough_memory
1.12 ± 2% -0.0 1.07 perf-profile.self.cycles-pp.__might_sleep
0.78 -0.0 0.73 perf-profile.self.cycles-pp.percpu_counter_add_batch
1.16 -0.0 1.12 perf-profile.self.cycles-pp.perf_event_mmap_output
0.94 -0.0 0.90 perf-profile.self.cycles-pp.kmem_cache_free
0.66 -0.0 0.62 perf-profile.self.cycles-pp.rcu_all_qs
0.49 ± 3% -0.0 0.46 ± 2% perf-profile.self.cycles-pp.cap_capable
0.40 ± 2% -0.0 0.36 ± 2% perf-profile.self.cycles-pp.strlen
0.65 -0.0 0.63 ± 2% perf-profile.self.cycles-pp.cap_mmap_addr
0.63 ± 4% +0.1 0.75 perf-profile.self.cycles-pp.lru_add_drain_cpu
0.00 +0.8 0.81 ± 2% perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
0.00 +0.9 0.94 perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.68 ± 2% +2.2 2.92 perf-profile.self.cycles-pp._raw_spin_lock
will-it-scale.per_process_ops
760000 +-+----------------------------------------------------------------+
| .+. + +. : |
755000 +-+ +. .+.+.+ : |
750000 +-+ : +. : |
|.+. : +. .+.. .+.+. |
745000 +-+ +..+.+ + + +..+. .+.|
| +.+.+..+ |
740000 +-+ |
| |
735000 +-+ |
730000 +-+ O |
| O O O |
725000 O-O O O O O O O O O O O O |
| O O O O O O O O |
720000 +-+----------------------------------------------------------------+
will-it-scale.workload
3.36e+07 +-+--------------------------------------------------------------+
3.34e+07 +-+ .+. .+. .+.+ |
| +.+..+.+ + +. : |
3.32e+07 +-+ +. + : |
3.3e+07 +-+ + + : |
|.+.+..+.+ +.+.+.+..+.+. |
3.28e+07 +-+ +.+.+.+..+.+.+.|
3.26e+07 +-+ |
3.24e+07 +-+ |
| |
3.22e+07 +-+ O O |
3.2e+07 +-O O O O O O O |
O O O O O O O O O O O O O O |
3.18e+07 +-+ O O |
3.16e+07 +-+--------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
***************************************************************************************************
ivb44: 48 threads Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz with 64G memory
=========================================================================================
bs/compiler/cpufreq_governor/disk/fs2/fs/ioengine/kconfig/nr_task/rootfs/runtime/rw/tbox_group/test_size/testcase/ucode:
4k/gcc-7/performance/1HDD/nfsv4/ext4/sync/x86_64-rhel-7.2/100%/debian-x86_64-2018-04-03.cgz/300s/write/ivb44/128G/fio-basic/0x42d
commit:
f3c0b8ce48 (" dmaengine fix for dmaengine-fix-4.19-rc4")
d884021f52 ("mm/swap: Add pagevec locking")
f3c0b8ce48408b74 d884021f52609407c7943705b3
---------------- --------------------------
fail:runs %reproduction fail:runs
| | |
:2 50% 1:4 kmsg.usb#-#:device_descriptor_read,error
:2 150% 3:4 dmesg.WARNING:at#for_ip_swapgs_restore_regs_and_return_to_usermode/0x
:2 150% 3:4 dmesg.WARNING:stack_recursion
0:2 15% 0:4 perf-profile.children.cycles-pp.io_schedule_timeout
0:2 17% 0:4 perf-profile.children.cycles-pp.schedule_timeout
%stddev %change %stddev
\ | \
4.1e+09 ± 76% -50.6% 2.025e+09 ±134% cpuidle.C3.time
7382534 ± 46% -61.3% 2854923 ± 99% cpuidle.C3.usage
29610 -1.1% 29281 proc-vmstat.nr_slab_unreclaimable
797236 +6.3% 847312 ± 4% proc-vmstat.pgfree
7382736 ± 46% -61.3% 2854836 ± 99% turbostat.C3
28.31 ± 76% -14.3 13.97 ±134% turbostat.C3%
2520 ± 2% -20.2% 2010 ± 14% slabinfo.eventpoll_pwq.active_objs
2520 ± 2% -20.2% 2010 ± 14% slabinfo.eventpoll_pwq.num_objs
350.00 ± 9% -27.1% 255.00 ± 17% slabinfo.secpath_cache.active_objs
350.00 ± 9% -27.1% 255.00 ± 17% slabinfo.secpath_cache.num_objs
1972 ± 2% -27.8% 1425 ± 24% slabinfo.task_group.active_objs
1972 ± 2% -27.8% 1425 ± 24% slabinfo.task_group.num_objs
86415 +65.8% 143246 ± 2% numa-meminfo.node0.Active
80095 ± 7% +69.3% 135603 ± 3% numa-meminfo.node0.Active(anon)
78653 ± 9% +69.4% 133251 ± 3% numa-meminfo.node0.AnonPages
8516 ± 75% +1396.8% 127471 ± 65% numa-meminfo.node0.Inactive(anon)
9990 ± 53% +1200.9% 129966 ± 63% numa-meminfo.node0.Shmem
198827 -27.4% 144403 ± 2% numa-meminfo.node1.Active
190365 ± 3% -27.9% 137259 ± 2% numa-meminfo.node1.Active(anon)
187188 ± 4% -28.1% 134639 ± 2% numa-meminfo.node1.AnonPages
231370 ± 2% -51.6% 112031 ± 74% numa-meminfo.node1.Inactive(anon)
240802 -48.7% 123588 ± 66% numa-meminfo.node1.Mapped
234636 ± 2% -51.3% 114274 ± 72% numa-meminfo.node1.Shmem
165.66 ± 44% -55.4% 73.84 ± 33% sched_debug.cfs_rq:/.load_avg.avg
304.80 ± 34% -37.0% 191.89 ± 6% sched_debug.cfs_rq:/.removed.load_avg.max
14073 ± 33% -36.8% 8887 ± 6% sched_debug.cfs_rq:/.removed.runnable_sum.max
125.80 ± 39% -60.3% 49.99 ± 26% sched_debug.cfs_rq:/.removed.util_avg.max
21.17 ± 48% -50.6% 10.45 ± 57% sched_debug.cfs_rq:/.removed.util_avg.stddev
506292 ± 2% -33.1% 338712 ± 18% sched_debug.cpu.avg_idle.min
103884 ± 5% +368.2% 486437 ± 71% sched_debug.cpu.avg_idle.stddev
7.96 +14.2% 9.10 ± 5% sched_debug.cpu.cpu_load[3].avg
246.40 ± 19% +37.1% 337.89 ± 9% sched_debug.cpu.cpu_load[3].max
37.79 ± 12% +29.3% 48.86 ± 9% sched_debug.cpu.cpu_load[3].stddev
160.10 ± 4% +43.3% 229.43 ± 13% sched_debug.cpu.cpu_load[4].max
24.69 +34.6% 33.25 ± 13% sched_debug.cpu.cpu_load[4].stddev
0.00 ± 6% +32.1% 0.00 ± 11% sched_debug.cpu.next_balance.stddev
20024 ± 7% +69.3% 33898 ± 3% numa-vmstat.node0.nr_active_anon
19661 ± 9% +69.4% 33311 ± 3% numa-vmstat.node0.nr_anon_pages
2128 ± 75% +1395.3% 31827 ± 65% numa-vmstat.node0.nr_inactive_anon
2499 ± 52% +1198.6% 32451 ± 63% numa-vmstat.node0.nr_shmem
20024 ± 7% +69.3% 33898 ± 3% numa-vmstat.node0.nr_zone_active_anon
2128 ± 75% +1395.3% 31827 ± 65% numa-vmstat.node0.nr_zone_inactive_anon
16608 ± 4% -26.0% 12293 ± 28% numa-vmstat.node1
47590 ± 3% -27.9% 34315 ± 2% numa-vmstat.node1.nr_active_anon
46795 ± 4% -28.1% 33660 ± 2% numa-vmstat.node1.nr_anon_pages
57747 ± 2% -51.5% 27994 ± 74% numa-vmstat.node1.nr_inactive_anon
60219 -48.6% 30929 ± 66% numa-vmstat.node1.nr_mapped
58564 ± 2% -51.2% 28556 ± 72% numa-vmstat.node1.nr_shmem
47590 ± 3% -27.9% 34315 ± 2% numa-vmstat.node1.nr_zone_active_anon
57747 ± 2% -51.5% 27994 ± 74% numa-vmstat.node1.nr_zone_inactive_anon
75.04 ± 6% -3.9 71.14 ± 4% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
87.77 -2.6 85.21 perf-profile.calltrace.cycles-pp.secondary_startup_64
1.31 ± 3% -0.2 1.11 ± 12% perf-profile.calltrace.cycles-pp.iov_iter_copy_from_user_atomic.generic_perform_write.__generic_file_write_iter.ext4_file_write_iter.__vfs_write
1.24 ± 4% -0.2 1.08 ± 13% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyin.iov_iter_copy_from_user_atomic.generic_perform_write.__generic_file_write_iter
1.24 ± 4% -0.1 1.09 ± 12% perf-profile.calltrace.cycles-pp.copyin.iov_iter_copy_from_user_atomic.generic_perform_write.__generic_file_write_iter.ext4_file_write_iter
1.25 +0.1 1.33 ± 3% perf-profile.calltrace.cycles-pp.lapic_next_deadline.clockevents_program_event.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt
78.72 -3.3 75.42 ± 3% perf-profile.children.cycles-pp.cpuidle_enter_state
87.77 -2.6 85.21 perf-profile.children.cycles-pp.secondary_startup_64
87.77 -2.6 85.21 perf-profile.children.cycles-pp.cpu_startup_entry
87.82 -2.5 85.29 perf-profile.children.cycles-pp.do_idle
1.37 ± 49% -0.7 0.62 ± 19% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
1.31 ± 3% -0.2 1.11 ± 12% perf-profile.children.cycles-pp.iov_iter_copy_from_user_atomic
0.30 ± 9% -0.2 0.14 ± 30% perf-profile.children.cycles-pp.__hrtimer_get_next_event
1.25 ± 4% -0.1 1.10 ± 12% perf-profile.children.cycles-pp.copyin
1.25 ± 3% -0.1 1.11 ± 14% perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
0.34 ± 18% -0.1 0.25 ± 28% perf-profile.children.cycles-pp.timerqueue_del
0.18 ± 13% -0.1 0.10 ± 58% perf-profile.children.cycles-pp.nr_iowait_cpu
0.24 ± 19% -0.1 0.15 ± 28% perf-profile.children.cycles-pp.interrupt_entry
0.26 ± 17% -0.1 0.19 ± 16% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
0.10 ± 40% -0.1 0.03 ±105% perf-profile.children.cycles-pp.__list_del_entry_valid
0.10 ± 40% -0.1 0.03 ±105% perf-profile.children.cycles-pp.pick_next_task_fair
0.07 ± 20% -0.0 0.04 ± 57% perf-profile.children.cycles-pp.mem_cgroup_try_charge
0.10 ± 14% -0.0 0.07 ± 26% perf-profile.children.cycles-pp.can_stop_idle_tick
0.20 ± 2% +0.0 0.24 ± 3% perf-profile.children.cycles-pp.radix_tree_next_chunk
0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.alloc_pages_current
0.00 +0.1 0.07 ± 23% perf-profile.children.cycles-pp.sys_imageblit
0.00 +0.1 0.07 ± 23% perf-profile.children.cycles-pp.drm_fb_helper_sys_imageblit
0.00 +0.1 0.07 ± 23% perf-profile.children.cycles-pp.mga_imageblit
0.10 ± 30% +0.1 0.17 ± 4% perf-profile.children.cycles-pp.rcu_needs_cpu
0.06 ± 16% +0.1 0.13 ± 42% perf-profile.children.cycles-pp.__wake_up_common_lock
0.08 ± 5% +0.1 0.17 ± 28% perf-profile.children.cycles-pp.__x64_sys_execve
0.08 ± 5% +0.1 0.17 ± 25% perf-profile.children.cycles-pp.execve
0.08 ± 6% +0.1 0.16 ± 26% perf-profile.children.cycles-pp.__do_execve_file
0.00 +0.1 0.09 ± 40% perf-profile.children.cycles-pp.proc_reg_read
0.00 +0.1 0.11 ± 37% perf-profile.children.cycles-pp.seq_read
0.00 +0.1 0.13 ± 42% perf-profile.children.cycles-pp.__mutex_lock
0.30 ± 6% +0.2 0.46 ± 28% perf-profile.children.cycles-pp._raw_spin_lock
0.26 +0.2 0.45 ± 23% perf-profile.children.cycles-pp._raw_spin_lock_irq
0.21 ± 23% -0.1 0.08 ± 15% perf-profile.self.cycles-pp.__hrtimer_get_next_event
0.18 ± 13% -0.1 0.09 ± 59% perf-profile.self.cycles-pp.nr_iowait_cpu
0.24 ± 19% -0.1 0.14 ± 30% perf-profile.self.cycles-pp.interrupt_entry
0.12 ± 28% -0.1 0.04 ±110% perf-profile.self.cycles-pp.timerqueue_del
0.24 ± 16% -0.1 0.16 ± 16% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
0.22 ± 4% -0.1 0.15 ± 18% perf-profile.self.cycles-pp.get_next_timer_interrupt
0.22 ± 4% -0.0 0.18 ± 12% perf-profile.self.cycles-pp.timerqueue_add
0.10 ± 14% -0.0 0.07 ± 26% perf-profile.self.cycles-pp.can_stop_idle_tick
0.20 ± 2% +0.0 0.24 ± 3% perf-profile.self.cycles-pp.radix_tree_next_chunk
0.10 ± 14% +0.1 0.17 ± 15% perf-profile.self.cycles-pp.do_syscall_64
0.09 +0.1 0.15 ± 24% perf-profile.self.cycles-pp.__update_load_avg_se
0.00 +0.1 0.07 ± 23% perf-profile.self.cycles-pp.sys_imageblit
0.00 +0.1 0.07 ± 31% perf-profile.self.cycles-pp.ext4_block_write_begin
0.10 ± 26% +0.1 0.17 ± 4% perf-profile.self.cycles-pp.rcu_needs_cpu
0.00 +0.1 0.07 ± 38% perf-profile.self.cycles-pp.hrtimer_get_next_event
0.00 +0.1 0.08 ± 19% perf-profile.self.cycles-pp.__mutex_lock
0.23 ± 2% +0.2 0.39 ± 23% perf-profile.self.cycles-pp._raw_spin_lock_irq
0.29 +0.2 0.44 ± 28% perf-profile.self.cycles-pp._raw_spin_lock
0.50 ± 2% +0.2 0.68 ± 16% perf-profile.self.cycles-pp.__next_timer_interrupt
***************************************************************************************************
lkp-hsw-ep2: 72 threads Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz with 128G memory
=========================================================================================
bs/compiler/cpufreq_governor/disk/fs/ioengine/kconfig/nr_task/rootfs/runtime/rw/tbox_group/test_size/testcase/ucode:
4k/gcc-7/performance/1SSD/ext4/sync/x86_64-rhel-7.2/64/debian-x86_64-2018-04-03.cgz/300s/randwrite/lkp-hsw-ep2/512g/fio-basic/0x3d
commit:
f3c0b8ce48 (" dmaengine fix for dmaengine-fix-4.19-rc4")
d884021f52 ("mm/swap: Add pagevec locking")
f3c0b8ce48408b74 d884021f52609407c7943705b3
---------------- --------------------------
fail:runs %reproduction fail:runs
| | |
:2 50% 1:4 dmesg.WARNING:at#for_ip_interrupt_entry/0x
:2 50% 1:4 dmesg.WARNING:at_ip_native_sched_clock/0x
0:2 22% 1:4 perf-profile.children.cycles-pp.schedule_timeout
1:2 60% 2:4 perf-profile.children.cycles-pp.io_schedule_timeout
%stddev %change %stddev
\ | \
5.06 +1.6 6.65 ± 11% fio.latency_100us%
0.86 ± 3% +0.1 0.99 ± 3% fio.latency_250us%
256.63 +7.7% 276.50 fio.time.system_time
45.00 +7.8% 48.50 ± 3% fio.write_clat_90%_us
58.00 +7.8% 62.50 ± 3% fio.write_clat_95%_us
14118389 ± 51% +79.1% 25292065 ± 9% cpuidle.C6.usage
2571 -1.1% 2543 proc-vmstat.nr_page_table_pages
19228 ± 79% -80.4% 3774 ± 3% proc-vmstat.numa_pte_updates
14115889 ± 51% +79.2% 25289182 ± 9% turbostat.C6
0.21 ± 17% -64.6% 0.07 ± 40% turbostat.Pkg%pc3
7321922 ± 2% -8.3% 6717403 ± 6% numa-numastat.node0.local_node
436778 ± 29% -67.9% 140332 ±173% numa-numastat.node0.numa_foreign
7336061 ± 2% -8.4% 6721069 ± 6% numa-numastat.node0.numa_hit
436778 ± 29% -67.9% 140332 ±173% numa-numastat.node1.numa_miss
865.00 ± 13% -20.7% 686.00 ± 26% slabinfo.nfs_commit_data.active_objs
865.00 ± 13% -20.7% 686.00 ± 26% slabinfo.nfs_commit_data.num_objs
2627 ± 24% -22.1% 2045 ± 34% slabinfo.scsi_sense_cache.active_objs
2628 ± 24% -21.4% 2065 ± 32% slabinfo.scsi_sense_cache.num_objs
51178 ± 4% -7.7% 47251 ± 4% slabinfo.vm_area_struct.active_objs
1287 ± 4% -7.8% 1186 ± 4% slabinfo.vm_area_struct.active_slabs
51507 ± 4% -7.8% 47465 ± 4% slabinfo.vm_area_struct.num_objs
1287 ± 4% -7.8% 1186 ± 4% slabinfo.vm_area_struct.num_slabs
2194 ± 3% -119.1% -419.80 sched_debug.cfs_rq:/.spread0.avg
5.71 ± 3% -14.8% 4.87 ± 12% sched_debug.cfs_rq:/.util_est_enqueued.avg
324306 ± 10% -9.0% 295152 ± 4% sched_debug.cpu.nr_load_updates.max
5621 ± 68% -69.7% 1702 ± 12% sched_debug.cpu.nr_load_updates.stddev
11.00 ± 13% -18.6% 8.95 ± 20% sched_debug.cpu.nr_uninterruptible.max
1027 ± 7% +45.0% 1490 ± 19% sched_debug.cpu.ttwu_count.min
546.55 -30.2% 381.53 ± 28% sched_debug.cpu.ttwu_local.min
0.00 ± 49% +122.3% 0.00 ± 30% sched_debug.rt_rq:/.rt_time.avg
0.02 ± 49% +122.3% 0.03 ± 30% sched_debug.rt_rq:/.rt_time.max
0.00 ± 49% +122.3% 0.00 ± 30% sched_debug.rt_rq:/.rt_time.stddev
289208 ± 15% -26.5% 212596 ± 23% numa-meminfo.node0.Active
190569 -33.4% 126938 ± 23% numa-meminfo.node0.Active(anon)
157915 ± 2% -36.2% 100710 ± 26% numa-meminfo.node0.AnonHugePages
190832 -35.5% 123030 ± 22% numa-meminfo.node0.AnonPages
7548 -16.5% 6301 ± 4% numa-meminfo.node0.KernelStack
8824 ± 2% -64.8% 3105 ± 78% numa-meminfo.node0.PageTables
69548 -11.1% 61798 ± 8% numa-meminfo.node0.SUnreclaim
87222 +74.1% 151868 ± 18% numa-meminfo.node1.Active(anon)
62739 ± 2% +93.2% 121230 ± 22% numa-meminfo.node1.AnonHugePages
79918 ± 2% +85.9% 148605 ± 18% numa-meminfo.node1.AnonPages
15315 +531.7% 96745 ± 49% numa-meminfo.node1.Inactive(anon)
4801 ± 4% +23.8% 5941 ± 5% numa-meminfo.node1.KernelStack
16370 +510.7% 99979 ± 48% numa-meminfo.node1.Mapped
1468 ± 12% +381.8% 7073 ± 34% numa-meminfo.node1.PageTables
47642 -33.4% 31737 ± 23% numa-vmstat.node0.nr_active_anon
47710 -35.5% 30760 ± 22% numa-vmstat.node0.nr_anon_pages
7549 -16.5% 6301 ± 4% numa-vmstat.node0.nr_kernel_stack
2206 ± 2% -64.8% 776.75 ± 78% numa-vmstat.node0.nr_page_table_pages
17386 -11.1% 15449 ± 8% numa-vmstat.node0.nr_slab_unreclaimable
3677874 -11.3% 3261486 ± 3% numa-vmstat.node0.nr_written
47642 -33.4% 31737 ± 23% numa-vmstat.node0.nr_zone_active_anon
432470 ± 29% -67.9% 138936 ±173% numa-vmstat.node0.numa_foreign
6791323 -12.1% 5968091 ± 4% numa-vmstat.node0.numa_hit
6776182 -12.0% 5962755 ± 4% numa-vmstat.node0.numa_local
21807 +74.1% 37971 ± 18% numa-vmstat.node1.nr_active_anon
19981 ± 2% +86.0% 37155 ± 18% numa-vmstat.node1.nr_anon_pages
3829 +532.5% 24221 ± 49% numa-vmstat.node1.nr_inactive_anon
4804 ± 4% +23.6% 5939 ± 5% numa-vmstat.node1.nr_kernel_stack
4094 +513.7% 25127 ± 48% numa-vmstat.node1.nr_mapped
367.50 ± 12% +381.1% 1768 ± 34% numa-vmstat.node1.nr_page_table_pages
3117498 ± 3% +14.3% 3562760 ± 4% numa-vmstat.node1.nr_written
21807 +74.1% 37971 ± 18% numa-vmstat.node1.nr_zone_active_anon
3829 +532.5% 24221 ± 49% numa-vmstat.node1.nr_zone_inactive_anon
5695266 +15.9% 6602176 ± 6% numa-vmstat.node1.numa_hit
5557113 +16.1% 6453474 ± 6% numa-vmstat.node1.numa_local
432474 ± 29% -67.9% 138938 ±173% numa-vmstat.node1.numa_miss
19.04 -4.1 14.96 ± 16% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary
16.23 ± 4% -3.6 12.66 ± 18% perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter_state.do_idle.cpu_startup_entry
6.30 ± 11% -2.2 4.15 ± 34% perf-profile.calltrace.cycles-pp.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter_state.do_idle
4.96 ± 14% -1.7 3.30 ± 39% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter_state
3.58 ± 3% -1.0 2.55 ± 19% perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt
5.03 ± 5% -0.8 4.21 ± 12% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter_state
3.06 ± 2% -0.8 2.26 ± 18% perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt
2.86 ± 3% -0.8 2.09 ± 18% perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
1.50 ± 23% -0.7 0.76 ± 80% perf-profile.calltrace.cycles-pp.rcu_process_callbacks.__softirqentry_text_start.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt
2.65 ± 22% -0.7 1.92 ± 38% perf-profile.calltrace.cycles-pp.clockevents_program_event.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter_state
1.66 ± 34% -0.7 0.96 ± 75% perf-profile.calltrace.cycles-pp.ktime_get.clockevents_program_event.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt
3.17 -0.6 2.54 ± 15% perf-profile.calltrace.cycles-pp.tick_nohz_get_sleep_length.menu_select.do_idle.cpu_startup_entry.start_secondary
2.56 -0.6 2.00 ± 16% perf-profile.calltrace.cycles-pp.tick_nohz_next_event.tick_nohz_get_sleep_length.menu_select.do_idle.cpu_startup_entry
1.58 -0.4 1.15 ± 24% perf-profile.calltrace.cycles-pp.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues
0.77 ± 9% -0.4 0.33 ±104% perf-profile.calltrace.cycles-pp.load_balance.rebalance_domains.__softirqentry_text_start.irq_exit.smp_apic_timer_interrupt
0.78 ± 3% -0.3 0.46 ± 58% perf-profile.calltrace.cycles-pp.run_timer_softirq.__softirqentry_text_start.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt
18.20 ± 4% -3.9 14.30 ± 16% perf-profile.children.cycles-pp.apic_timer_interrupt
16.71 ± 6% -3.7 13.05 ± 18% perf-profile.children.cycles-pp.smp_apic_timer_interrupt
8.34 ± 14% -1.6 6.73 ± 21% perf-profile.children.cycles-pp.irq_exit
8.96 ± 4% -1.5 7.42 ± 14% perf-profile.children.cycles-pp.hrtimer_interrupt
3.20 ± 19% -1.2 2.01 ± 36% perf-profile.children.cycles-pp.ktime_get
3.79 -1.1 2.68 ± 19% perf-profile.children.cycles-pp.tick_sched_timer
3.26 ± 2% -0.9 2.38 ± 18% perf-profile.children.cycles-pp.tick_sched_handle
3.05 ± 2% -0.8 2.21 ± 18% perf-profile.children.cycles-pp.update_process_times
5.28 -0.8 4.47 ± 11% perf-profile.children.cycles-pp.__hrtimer_run_queues
5.86 ± 2% -0.8 5.08 ± 13% perf-profile.children.cycles-pp.menu_select
2.71 ± 23% -0.7 1.97 ± 37% perf-profile.children.cycles-pp.clockevents_program_event
3.24 -0.7 2.57 ± 15% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
1.54 ± 24% -0.6 0.90 ± 55% perf-profile.children.cycles-pp.rcu_process_callbacks
2.62 -0.6 2.03 ± 16% perf-profile.children.cycles-pp.tick_nohz_next_event
0.86 ± 11% -0.5 0.40 ± 31% perf-profile.children.cycles-pp.tick_nohz_irq_exit
0.53 ± 34% -0.4 0.09 ± 39% perf-profile.children.cycles-pp.timekeeping_max_deferment
0.71 ± 14% -0.3 0.42 ± 51% perf-profile.children.cycles-pp.note_gp_changes
0.80 -0.2 0.58 ± 19% perf-profile.children.cycles-pp.run_timer_softirq
0.23 ± 53% -0.2 0.06 ± 64% perf-profile.children.cycles-pp.page_fault
0.34 ± 49% -0.2 0.16 ±111% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
0.23 ± 52% -0.2 0.05 ± 62% perf-profile.children.cycles-pp.do_page_fault
0.22 ± 55% -0.2 0.05 ± 62% perf-profile.children.cycles-pp.__do_page_fault
0.57 ± 12% -0.1 0.43 ± 31% perf-profile.children.cycles-pp.schedule
0.67 ± 5% -0.1 0.53 ± 23% perf-profile.children.cycles-pp.rcu_check_callbacks
0.68 ± 4% -0.1 0.54 ± 7% perf-profile.children.cycles-pp._raw_spin_lock
0.52 ± 6% -0.1 0.42 ± 15% perf-profile.children.cycles-pp.unwind_next_frame
0.19 ± 21% -0.1 0.10 ± 24% perf-profile.children.cycles-pp.raise_softirq
0.16 ± 9% -0.1 0.08 ± 15% perf-profile.children.cycles-pp.cpu_load_update
0.56 -0.1 0.49 ± 10% perf-profile.children.cycles-pp.__entry_SYSCALL_64_trampoline
0.17 ± 23% -0.1 0.10 ± 29% perf-profile.children.cycles-pp.irq_entries_start
0.11 ± 4% -0.1 0.05 ± 67% perf-profile.children.cycles-pp.orc_find
0.23 ± 6% -0.1 0.18 ± 8% perf-profile.children.cycles-pp.rcu_eqs_enter
0.17 ± 3% -0.0 0.12 ± 25% perf-profile.children.cycles-pp.calc_global_load_tick
0.09 ± 17% -0.0 0.04 ±103% perf-profile.children.cycles-pp.blk_queue_start_tag
0.15 ± 6% -0.0 0.12 ± 23% perf-profile.children.cycles-pp.rcu_nmi_exit
0.09 -0.0 0.06 ± 59% perf-profile.children.cycles-pp.percpu_counter_add_batch
0.09 ± 22% -0.0 0.06 ± 22% perf-profile.children.cycles-pp.update_rt_rq_load_avg
0.10 ± 5% -0.0 0.08 ± 5% perf-profile.children.cycles-pp.__alloc_pages_nodemask
0.08 ± 12% -0.0 0.06 ± 17% perf-profile.children.cycles-pp.task_tick_idle
0.10 ± 5% -0.0 0.08 ± 5% perf-profile.children.cycles-pp.get_iowait_load
0.10 ± 5% +0.0 0.14 ± 15% perf-profile.children.cycles-pp.nr_iowait_cpu
0.00 +0.1 0.06 ± 20% perf-profile.children.cycles-pp.blk_account_io_done
0.09 ± 17% +0.1 0.16 ± 21% perf-profile.children.cycles-pp.blk_finish_request
0.10 ± 26% +0.1 0.22 ± 45% perf-profile.children.cycles-pp.kmem_cache_free
0.30 ± 6% +0.2 0.48 ± 39% perf-profile.children.cycles-pp.ext4_end_bio
0.00 +0.2 0.18 ± 27% perf-profile.children.cycles-pp.pipe_read
0.75 ± 6% +0.2 0.93 ± 8% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
0.07 ± 7% +0.2 0.29 ± 26% perf-profile.children.cycles-pp.__vfs_read
0.07 ± 20% +0.2 0.32 ± 24% perf-profile.children.cycles-pp.vfs_read
0.07 ± 20% +0.3 0.34 ± 24% perf-profile.children.cycles-pp.ksys_read
0.03 ±100% +0.3 0.34 ± 23% perf-profile.children.cycles-pp.read
2.52 ± 25% -1.2 1.35 ± 56% perf-profile.self.cycles-pp.ktime_get
0.53 ± 35% -0.4 0.08 ± 42% perf-profile.self.cycles-pp.timekeeping_max_deferment
2.04 ± 9% -0.3 1.72 ± 6% perf-profile.self.cycles-pp.cpuidle_enter_state
0.42 ± 12% -0.2 0.23 ± 46% perf-profile.self.cycles-pp.__softirqentry_text_start
0.44 ± 21% -0.2 0.25 ± 50% perf-profile.self.cycles-pp.note_gp_changes
0.34 ± 49% -0.2 0.16 ±111% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
0.41 -0.2 0.24 ± 22% perf-profile.self.cycles-pp.run_timer_softirq
0.29 -0.2 0.13 ± 34% perf-profile.self.cycles-pp.tick_sched_timer
0.67 ± 3% -0.1 0.53 ± 8% perf-profile.self.cycles-pp._raw_spin_lock
0.38 ± 6% -0.1 0.25 ± 35% perf-profile.self.cycles-pp.update_blocked_averages
0.18 ± 18% -0.1 0.09 ± 27% perf-profile.self.cycles-pp.raise_softirq
0.21 ± 2% -0.1 0.12 ± 21% perf-profile.self.cycles-pp.load_balance
0.16 ± 9% -0.1 0.08 ± 15% perf-profile.self.cycles-pp.cpu_load_update
0.28 ± 14% -0.1 0.21 ± 24% perf-profile.self.cycles-pp.rcu_process_callbacks
0.26 ± 21% -0.1 0.18 ± 10% perf-profile.self.cycles-pp.idle_cpu
0.54 -0.1 0.46 ± 12% perf-profile.self.cycles-pp.__entry_SYSCALL_64_trampoline
0.18 ± 8% -0.1 0.12 ± 20% perf-profile.self.cycles-pp.scheduler_tick
0.23 ± 6% -0.1 0.17 ± 4% perf-profile.self.cycles-pp.rcu_eqs_enter
0.11 -0.1 0.05 ± 70% perf-profile.self.cycles-pp.orc_find
0.09 ± 17% -0.1 0.03 ±102% perf-profile.self.cycles-pp.update_rt_rq_load_avg
0.09 ± 22% -0.1 0.04 ±102% perf-profile.self.cycles-pp.tick_nohz_get_sleep_length
0.15 ± 17% -0.1 0.09 ± 24% perf-profile.self.cycles-pp.rcu_needs_cpu
0.35 ± 9% -0.1 0.30 ± 9% perf-profile.self.cycles-pp.tick_nohz_next_event
0.15 ± 6% -0.0 0.12 ± 23% perf-profile.self.cycles-pp.rcu_nmi_exit
0.07 ± 20% -0.0 0.04 ± 63% perf-profile.self.cycles-pp.ext4_bio_write_page
0.10 ± 5% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.get_iowait_load
0.09 ± 11% +0.0 0.14 ± 15% perf-profile.self.cycles-pp.nr_iowait_cpu
0.15 ± 33% +0.1 0.24 ± 17% perf-profile.self.cycles-pp.perf_mux_hrtimer_handler
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Rong Chen
3 years, 9 months
[x86/pti/64] bf904d2762: will-it-scale.per_thread_ops 1.7% improvement
by kernel test robot
Greeting,
FYI, we noticed a 1.7% improvement of will-it-scale.per_thread_ops due to commit:
commit: bf904d2762ee6fc1e4acfcb0772bbfb4a27ad8a6 ("x86/pti/64: Remove the SYSCALL64 entry trampoline")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
in testcase: will-it-scale
on test machine: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory
with following parameters:
nr_task: 16
mode: thread
test: pwrite1
cpufreq_governor: performance
test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
test-url: https://github.com/antonblanchard/will-it-scale
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-7/performance/x86_64-rhel-7.2/thread/16/debian-x86_64-2018-04-03.cgz/lkp-bdw-ep3d/pwrite1/will-it-scale
commit:
98f05b5138 ("x86/entry/64: Use the TSS sp2 slot for SYSCALL/SYSRET scratch space")
bf904d2762 ("x86/pti/64: Remove the SYSCALL64 entry trampoline")
98f05b5138f0a9b5 bf904d2762ee6fc1e4acfcb077
---------------- --------------------------
fail:runs %reproduction fail:runs
| | |
1:4 -25% :4 dmesg.WARNING:at#for_ip_interrupt_entry/0x
2:4 -50% :4 dmesg.WARNING:at_ip_fsnotify/0x
%stddev %change %stddev
\ | \
1221307 +1.7% 1242132 will-it-scale.per_thread_ops
7349 ± 3% +3.6% 7616 will-it-scale.time.minor_page_faults
675.23 +1.8% 687.28 will-it-scale.time.user_time
19540927 +1.7% 19874128 will-it-scale.workload
4323 ± 16% -54.7% 1958 ±103% numa-numastat.node0.other_node
98872 ± 24% +33.4% 131877 ± 10% numa-meminfo.node0.AnonPages
2292 ± 8% -10.5% 2050 ± 7% numa-meminfo.node1.PageTables
24718 ± 24% +33.4% 32969 ± 10% numa-vmstat.node0.nr_anon_pages
7864 ± 12% +21.7% 9568 ± 15% numa-vmstat.node1
573.00 ± 8% -10.6% 512.50 ± 7% numa-vmstat.node1.nr_page_table_pages
2.25 ± 15% -50.0% 1.12 ± 60% sched_debug.cfs_rq:/.load_avg.min
418.57 ± 87% -81.2% 78.54 ±173% sched_debug.cfs_rq:/.removed.runnable_sum.avg
7842 ± 70% -76.0% 1885 ±173% sched_debug.cfs_rq:/.removed.runnable_sum.max
1734 ± 77% -78.3% 376.68 ±173% sched_debug.cfs_rq:/.removed.runnable_sum.stddev
-2477409 -0.1% -2474518 sched_debug.cfs_rq:/.spread0.min
209211 ± 19% -30.2% 146101 ± 30% sched_debug.cpu.avg_idle.min
70.04 ± 7% -18.4% 57.17 ± 7% sched_debug.cpu.cpu_load[2].max
66.92 ± 5% -11.1% 59.46 ± 6% sched_debug.cpu.cpu_load[3].max
6736 ± 23% +37.8% 9285 ± 9% sched_debug.cpu.ttwu_local.max
1672 ± 12% +32.2% 2210 ± 11% sched_debug.cpu.ttwu_local.stddev
1.81 -0.3 1.56 perf-stat.branch-miss-rate%
4.262e+10 -13.3% 3.696e+10 perf-stat.branch-misses
1.27 -1.3% 1.25 perf-stat.cpi
0.01 ± 7% -0.0 0.00 ± 2% perf-stat.dTLB-load-miss-rate%
5.163e+08 ± 7% -62.1% 1.958e+08 ± 2% perf-stat.dTLB-load-misses
4.318e+12 +1.4% 4.38e+12 perf-stat.dTLB-loads
0.01 ± 6% -0.0 0.00 ± 4% perf-stat.dTLB-store-miss-rate%
4.264e+08 ± 6% -69.6% 1.294e+08 ± 4% perf-stat.dTLB-store-misses
2.915e+12 +1.1% 2.947e+12 perf-stat.dTLB-stores
2.21 ± 3% +95.5 97.67 perf-stat.iTLB-load-miss-rate%
2.564e+08 ± 3% +2372.0% 6.338e+09 perf-stat.iTLB-load-misses
1.136e+10 -98.7% 1.509e+08 perf-stat.iTLB-loads
1.18e+13 +1.4% 1.196e+13 perf-stat.instructions
46053 ± 3% -95.9% 1887 ± 2% perf-stat.instructions-per-iTLB-miss
0.79 +1.4% 0.80 perf-stat.ipc
8.65 ± 4% -8.7 0.00 perf-profile.calltrace.cycles-pp.__entry_SYSCALL_64_trampoline
0.57 ± 4% -0.2 0.39 ± 57% perf-profile.calltrace.cycles-pp.___might_sleep.down_write.generic_file_write_iter.__vfs_write.vfs_write
0.00 +8.4 8.41 ± 2% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64
9.48 ± 5% -9.5 0.00 perf-profile.children.cycles-pp.__entry_SYSCALL_64_trampoline
0.03 ±100% +0.0 0.07 ± 17% perf-profile.children.cycles-pp.clockevents_program_event
0.01 ±173% +0.1 0.07 ± 23% perf-profile.children.cycles-pp.ktime_get
0.31 ± 6% +0.1 0.37 ± 6% perf-profile.children.cycles-pp.smp_apic_timer_interrupt
0.35 ± 8% +0.1 0.42 ± 6% perf-profile.children.cycles-pp.apic_timer_interrupt
0.00 +0.2 0.17 ± 4% perf-profile.children.cycles-pp.__x86_indirect_thunk_r10
0.00 +1.0 0.96 perf-profile.children.cycles-pp.__x86_indirect_thunk_rax
0.00 +8.4 8.42 ± 2% perf-profile.children.cycles-pp.entry_SYSCALL_64
9.31 ± 4% -9.3 0.00 perf-profile.self.cycles-pp.__entry_SYSCALL_64_trampoline
1.55 ± 6% -0.6 0.97 perf-profile.self.cycles-pp.do_syscall_64
1.03 ± 5% -0.2 0.81 ± 4% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
0.00 +0.1 0.12 ± 3% perf-profile.self.cycles-pp.__x86_indirect_thunk_r10
0.00 +0.8 0.81 ± 2% perf-profile.self.cycles-pp.__x86_indirect_thunk_rax
0.00 +8.4 8.42 ± 2% perf-profile.self.cycles-pp.entry_SYSCALL_64
will-it-scale.per_thread_ops
1.255e+06 +-+-------------------------------------------------------------+
1.25e+06 +-+ O |
| O O O O |
1.245e+06 O-+ O O O O O O O O OO O O O |
1.24e+06 +-O O O OO O O OO O OO |
| OO O O |
1.235e+06 +-+ |
1.23e+06 +-+ |
1.225e+06 +-+.+ .+.+ |
| +.+. +. .+.+.++ + .+ .|
1.22e+06 +-+ +.+ +. .+. + +. .+ .+.+.+.++.+.+ +.+ |
1.215e+06 +-+ + + + .+ + |
| :.+ |
1.21e+06 +-+ + |
1.205e+06 +-+-------------------------------------------------------------+
will-it-scale.workload
2.01e+07 +-+--------------------------------------------------------------+
| |
2e+07 +-+ OO O O |
O O O O O O O |
1.99e+07 +-O O O O O O O O O O O O OO O O O OO |
1.98e+07 +-+ O O O O O |
| |
1.97e+07 +-+ |
| +.+ |
1.96e+07 +-+.+ .+. .+ +. .+. + : .|
1.95e+07 +-+ + +.+ +. .+ : + + : .+.++.+. .+.++.+ |
| + + .+ +.+ .++.+.+ + |
1.94e+07 +-+ + : + |
| : + |
1.93e+07 +-+--------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Rong Chen
3 years, 9 months
[fsnotify] 60f7ed8c7c: will-it-scale.per_thread_ops -5.9% regression
by kernel test robot
Greeting,
FYI, we noticed a -5.9% regression of will-it-scale.per_thread_ops due to commit:
commit: 60f7ed8c7c4d06aeda448c6da74621552ee739aa ("fsnotify: send path type events to group with super block marks")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
in testcase: will-it-scale
on test machine: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory
with following parameters:
nr_task: 16
mode: thread
test: unlink2
cpufreq_governor: performance
test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
test-url: https://github.com/antonblanchard/will-it-scale
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-7/performance/x86_64-rhel-7.2/thread/16/debian-x86_64-2018-04-03.cgz/lkp-bdw-ep3d/unlink2/will-it-scale
commit:
1e6cb72399 ("fsnotify: add super block object type")
60f7ed8c7c ("fsnotify: send path type events to group with super block marks")
1e6cb72399fd58b3 60f7ed8c7c4d06aeda448c6da7
---------------- --------------------------
%stddev %change %stddev
\ | \
54483 -5.9% 51256 will-it-scale.per_thread_ops
46266 ± 2% -4.3% 44270 ± 2% will-it-scale.time.involuntary_context_switches
103.21 -7.8% 95.17 will-it-scale.time.user_time
871749 -5.9% 820115 will-it-scale.workload
10888 +22.2% 13303 ± 17% numa-meminfo.node0.Mapped
2001 ± 12% -16.8% 1665 ± 16% numa-meminfo.node0.PageTables
865.75 ± 32% +42.7% 1235 ± 22% slabinfo.dmaengine-unmap-16.active_objs
865.75 ± 32% +42.7% 1235 ± 22% slabinfo.dmaengine-unmap-16.num_objs
10974 ± 34% +60.2% 17584 ± 13% numa-vmstat.node0
2826 ± 3% +24.6% 3523 ± 16% numa-vmstat.node0.nr_mapped
500.00 ± 12% -16.9% 415.75 ± 16% numa-vmstat.node0.nr_page_table_pages
20375718 -6.3% 19092155 proc-vmstat.numa_hit
20370933 -6.3% 19087362 proc-vmstat.numa_local
69383484 -6.3% 65029677 proc-vmstat.pgalloc_normal
69362606 -6.3% 65008530 proc-vmstat.pgfree
8.39 ±109% +7.9e+05% 66298 ±140% sched_debug.cfs_rq:/.MIN_vruntime.avg
201.35 ±109% +2.2e+05% 436949 ± 61% sched_debug.cfs_rq:/.MIN_vruntime.max
40.24 ±109% +3.4e+05% 135845 ± 97% sched_debug.cfs_rq:/.MIN_vruntime.stddev
8.39 ±109% +7.9e+05% 66298 ±140% sched_debug.cfs_rq:/.max_vruntime.avg
201.35 ±109% +2.2e+05% 436949 ± 61% sched_debug.cfs_rq:/.max_vruntime.max
40.24 ±109% +3.4e+05% 135845 ± 97% sched_debug.cfs_rq:/.max_vruntime.stddev
43805 ± 6% +35.5% 59365 ± 31% sched_debug.cpu.load.avg
108694 ± 62% +154.7% 276883 ± 25% sched_debug.cpu.load.max
33945 ± 37% +119.6% 74560 ± 42% sched_debug.cpu.load.stddev
34287 ± 3% +10.1% 37761 ± 4% sched_debug.cpu.nr_switches.max
15993 ± 2% +10.8% 17727 ± 4% sched_debug.cpu.sched_goidle.max
1.36 ± 2% -0.1 1.24 perf-stat.branch-miss-rate%
1.546e+10 ± 2% -10.3% 1.387e+10 perf-stat.branch-misses
3.025e+08 ± 8% -17.7% 2.489e+08 ± 14% perf-stat.dTLB-load-misses
1.603e+12 ± 2% -2.6% 1.561e+12 perf-stat.dTLB-loads
0.01 ± 7% -0.0 0.01 ± 6% perf-stat.dTLB-store-miss-rate%
1.02e+08 ± 5% -34.7% 66552058 ± 6% perf-stat.dTLB-store-misses
9.269e+11 -5.8% 8.729e+11 perf-stat.dTLB-stores
4.885e+08 ± 33% -25.3% 3.649e+08 ± 9% perf-stat.iTLB-load-misses
6.92e+08 ± 5% -9.7% 6.251e+08 ± 2% perf-stat.node-loads
3.66e+09 ± 2% -8.5% 3.347e+09 ± 2% perf-stat.node-store-misses
2.464e+09 ± 3% -11.4% 2.184e+09 ± 2% perf-stat.node-stores
6419017 +3.9% 6671008 perf-stat.path-length
11.13 ± 16% -7.2 3.93 ± 6% perf-profile.calltrace.cycles-pp.d_instantiate.shmem_mknod.path_openat.do_filp_open.do_sys_open
11.80 ± 14% -7.0 4.76 ± 4% perf-profile.calltrace.cycles-pp.__destroy_inode.destroy_inode.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
11.96 ± 14% -7.0 4.94 ± 5% perf-profile.calltrace.cycles-pp.destroy_inode.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
10.89 ± 15% -7.0 3.88 ± 4% perf-profile.calltrace.cycles-pp.security_inode_free.__destroy_inode.destroy_inode.do_unlinkat.do_syscall_64
10.03 ± 17% -6.7 3.33 ± 7% perf-profile.calltrace.cycles-pp.inode_doinit_with_dentry.security_d_instantiate.d_instantiate.shmem_mknod.path_openat
10.07 ± 16% -6.7 3.37 ± 6% perf-profile.calltrace.cycles-pp.security_d_instantiate.d_instantiate.shmem_mknod.path_openat.do_filp_open
9.91 ± 16% -6.7 3.23 ± 5% perf-profile.calltrace.cycles-pp.selinux_inode_free_security.security_inode_free.__destroy_inode.destroy_inode.do_unlinkat
9.17 ± 17% -6.5 2.66 ± 6% perf-profile.calltrace.cycles-pp._raw_spin_lock.selinux_inode_free_security.security_inode_free.__destroy_inode.destroy_inode
9.24 ± 18% -6.4 2.81 ± 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.inode_doinit_with_dentry.security_d_instantiate.d_instantiate.shmem_mknod
8.56 ± 19% -6.3 2.31 ± 10% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.inode_doinit_with_dentry.security_d_instantiate.d_instantiate
8.57 ± 18% -6.2 2.33 ± 7% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.selinux_inode_free_security.security_inode_free.__destroy_inode
1.89 ± 16% -0.6 1.28 ± 14% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.shmem_reserve_inode.shmem_get_inode.shmem_mknod
3.08 ± 11% -0.6 2.48 ± 10% perf-profile.calltrace.cycles-pp.shmem_evict_inode.evict.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.80 ± 16% -0.5 1.26 ± 16% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.shmem_free_inode.shmem_evict_inode.evict
0.96 ± 11% -0.3 0.62 ± 5% perf-profile.calltrace.cycles-pp.__call_rcu.security_inode_free.__destroy_inode.destroy_inode.do_unlinkat
0.86 ± 10% -0.3 0.55 ± 5% perf-profile.calltrace.cycles-pp.rcu_segcblist_enqueue.__call_rcu.security_inode_free.__destroy_inode.destroy_inode
0.75 ± 8% -0.3 0.47 ± 59% perf-profile.calltrace.cycles-pp.security_inode_init_security.shmem_mknod.path_openat.do_filp_open.do_sys_open
0.70 ± 8% -0.3 0.43 ± 58% perf-profile.calltrace.cycles-pp.selinux_inode_init_security.security_inode_init_security.shmem_mknod.path_openat.do_filp_open
1.07 ± 6% -0.2 0.82 ± 7% perf-profile.calltrace.cycles-pp.security_inode_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
1.01 ± 6% -0.2 0.77 ± 8% perf-profile.calltrace.cycles-pp.may_create.security_inode_create.path_openat.do_filp_open.do_sys_open
0.26 ±100% +0.4 0.63 ± 7% perf-profile.calltrace.cycles-pp._raw_spin_lock.new_inode_pseudo.new_inode.shmem_get_inode.shmem_mknod
0.77 ± 7% +0.4 1.16 ± 2% perf-profile.calltrace.cycles-pp.inode_wait_for_writeback.evict.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.74 ± 7% +0.4 1.14 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock.inode_wait_for_writeback.evict.do_unlinkat.do_syscall_64
1.30 ± 9% +0.7 2.00 ± 8% perf-profile.calltrace.cycles-pp.do_dentry_open.path_openat.do_filp_open.do_sys_open.do_syscall_64
0.00 +0.8 0.78 ± 6% perf-profile.calltrace.cycles-pp.fsnotify.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.94 ± 6% +0.8 1.72 ± 3% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.86 ± 6% +0.8 1.64 ± 3% perf-profile.calltrace.cycles-pp.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +0.8 0.79 ± 17% perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.native_queued_spin_lock_slowpath._raw_spin_lock.evict
0.00 +0.8 0.79 ± 4% perf-profile.calltrace.cycles-pp.fsnotify.do_dentry_open.path_openat.do_filp_open.do_sys_open
0.00 +0.8 0.79 ± 17% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.native_queued_spin_lock_slowpath._raw_spin_lock.evict.do_unlinkat
0.00 +0.8 0.80 ± 18% perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.native_queued_spin_lock_slowpath._raw_spin_lock.inode_sb_list_add
0.00 +0.8 0.81 ± 18% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.native_queued_spin_lock_slowpath._raw_spin_lock.inode_sb_list_add.new_inode
0.00 +0.9 0.85 perf-profile.calltrace.cycles-pp.fsnotify.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64
0.27 ±100% +1.0 1.29 ± 2% perf-profile.calltrace.cycles-pp.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.27 ±173% +1.2 1.50 ± 19% perf-profile.calltrace.cycles-pp.rcu_process_callbacks.__softirqentry_text_start.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt
0.27 ±173% +1.2 1.51 ± 18% perf-profile.calltrace.cycles-pp.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt.native_queued_spin_lock_slowpath._raw_spin_lock
0.27 ±173% +1.2 1.51 ± 18% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt.native_queued_spin_lock_slowpath
6.36 ± 9% +8.3 14.64 ± 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.inode_sb_list_add.new_inode.shmem_get_inode
14.67 ± 8% +8.4 23.08 ± 5% perf-profile.calltrace.cycles-pp.shmem_get_inode.shmem_mknod.path_openat.do_filp_open.do_sys_open
6.72 ± 9% +8.4 15.14 ± 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.inode_sb_list_add.new_inode.shmem_get_inode.shmem_mknod
11.87 ± 7% +8.4 20.30 ± 5% perf-profile.calltrace.cycles-pp.evict.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
6.19 ± 7% +8.5 14.67 ± 7% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.evict.do_unlinkat.do_syscall_64
7.41 ± 9% +8.6 15.96 ± 7% perf-profile.calltrace.cycles-pp.inode_sb_list_add.new_inode.shmem_get_inode.shmem_mknod.path_openat
10.33 ± 8% +8.6 18.95 ± 6% perf-profile.calltrace.cycles-pp.new_inode.shmem_get_inode.shmem_mknod.path_openat.do_filp_open
6.93 ± 7% +8.8 15.71 ± 7% perf-profile.calltrace.cycles-pp._raw_spin_lock.evict.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
11.14 ± 15% -7.2 3.93 ± 6% perf-profile.children.cycles-pp.d_instantiate
11.81 ± 14% -7.0 4.76 ± 4% perf-profile.children.cycles-pp.__destroy_inode
11.97 ± 14% -7.0 4.95 ± 5% perf-profile.children.cycles-pp.destroy_inode
10.89 ± 15% -7.0 3.88 ± 4% perf-profile.children.cycles-pp.security_inode_free
10.04 ± 17% -6.7 3.33 ± 7% perf-profile.children.cycles-pp.inode_doinit_with_dentry
10.07 ± 17% -6.7 3.37 ± 6% perf-profile.children.cycles-pp.security_d_instantiate
9.91 ± 16% -6.7 3.24 ± 5% perf-profile.children.cycles-pp.selinux_inode_free_security
3.09 ± 11% -0.6 2.49 ± 11% perf-profile.children.cycles-pp.shmem_evict_inode
0.84 ± 9% -0.4 0.46 ± 46% perf-profile.children.cycles-pp.selinux_determine_inode_label
1.32 ± 9% -0.3 0.99 ± 4% perf-profile.children.cycles-pp.__call_rcu
0.98 ± 9% -0.3 0.66 ± 5% perf-profile.children.cycles-pp.rcu_segcblist_enqueue
1.07 ± 6% -0.2 0.83 ± 8% perf-profile.children.cycles-pp.security_inode_create
0.94 ± 7% -0.2 0.69 ± 9% perf-profile.children.cycles-pp.__list_del_entry_valid
1.01 ± 7% -0.2 0.77 ± 8% perf-profile.children.cycles-pp.may_create
0.37 ± 6% -0.2 0.16 ± 6% perf-profile.children.cycles-pp.__fd_install
0.75 ± 8% -0.2 0.58 ± 16% perf-profile.children.cycles-pp.security_inode_init_security
0.70 ± 8% -0.2 0.54 ± 17% perf-profile.children.cycles-pp.selinux_inode_init_security
0.42 ± 8% -0.1 0.36 ± 4% perf-profile.children.cycles-pp.d_delete
0.21 ± 8% -0.1 0.16 ± 13% perf-profile.children.cycles-pp._atomic_dec_and_lock
0.34 ± 6% -0.1 0.29 ± 5% perf-profile.children.cycles-pp.fsnotify_destroy_marks
0.24 ± 8% -0.0 0.21 ± 7% perf-profile.children.cycles-pp.down_write
0.06 ± 14% +0.0 0.08 ± 15% perf-profile.children.cycles-pp.prandom_u32_state
0.12 ± 8% +0.0 0.16 ± 7% perf-profile.children.cycles-pp.__d_instantiate
0.77 ± 7% +0.4 1.17 ± 2% perf-profile.children.cycles-pp.inode_wait_for_writeback
1.30 ± 9% +0.7 2.00 ± 8% perf-profile.children.cycles-pp.do_dentry_open
0.94 ± 6% +0.8 1.73 ± 3% perf-profile.children.cycles-pp.exit_to_usermode_loop
0.86 ± 6% +0.8 1.65 ± 2% perf-profile.children.cycles-pp.task_work_run
0.51 ± 8% +0.8 1.30 ± 2% perf-profile.children.cycles-pp.__fput
0.23 ± 13% +2.3 2.52 ± 4% perf-profile.children.cycles-pp.fsnotify
14.69 ± 8% +8.4 23.11 ± 5% perf-profile.children.cycles-pp.shmem_get_inode
11.88 ± 7% +8.4 20.30 ± 5% perf-profile.children.cycles-pp.evict
7.42 ± 9% +8.6 15.97 ± 7% perf-profile.children.cycles-pp.inode_sb_list_add
10.33 ± 8% +8.6 18.95 ± 6% perf-profile.children.cycles-pp.new_inode
0.74 ± 9% -0.4 0.37 ± 59% perf-profile.self.cycles-pp.selinux_determine_inode_label
0.97 ± 9% -0.3 0.66 ± 5% perf-profile.self.cycles-pp.rcu_segcblist_enqueue
0.92 ± 7% -0.2 0.68 ± 9% perf-profile.self.cycles-pp.__list_del_entry_valid
0.36 ± 5% -0.2 0.16 ± 7% perf-profile.self.cycles-pp.__fd_install
0.41 ± 18% -0.2 0.22 ± 14% perf-profile.self.cycles-pp.inode_doinit_with_dentry
0.15 ± 12% -0.0 0.10 ± 10% perf-profile.self.cycles-pp._atomic_dec_and_lock
0.14 ± 8% -0.0 0.11 ± 6% perf-profile.self.cycles-pp.down_write
0.08 ± 11% +0.0 0.11 ± 4% perf-profile.self.cycles-pp.__d_instantiate
0.10 ± 10% +0.0 0.14 ± 7% perf-profile.self.cycles-pp.inode_sb_list_add
0.22 ± 12% +2.2 2.47 ± 3% perf-profile.self.cycles-pp.fsnotify
will-it-scale.workload
940000 +-+----------------------------------------------------------------+
| + + :: |
920000 +-+ : :: .+ : : |
900000 +-+ +. + : : : : + + : : |
| + : +. : + : : .+ .+. : :+ +. .+.: + +. |
880000 +-+: + : +. : + + : +. .+ + + + + + : : + |
|: : + + + + :+ + : : +|
860000 +-+ + + :: |
| + |
840000 +-+ O |
820000 O-+ O O O O O O O O O |
| O O O O |
800000 +-+ O O O O O O O |
| O O O O O |
780000 +-+----------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Rong Chen
3 years, 9 months
[platform/x86] 65d69a0593: WARNING:Could_NOT_find_tracepoint_structs_for_some_tracepoints
by kernel test robot
FYI, we noticed the following commit (built with gcc-7):
commit: 65d69a05938b05836f270501ba217ff521748312 ("platform/x86: Add SoCWatch driver for v2.6")
2018-09-30 11:20:01
in testcase: trinity
with following parameters:
runtime: 300s
test-description: Trinity is a linux system call fuzz tester.
test-url: http://codemonkey.org.uk/projects/trinity/
on test machine: qemu-system-x86_64 -enable-kvm -cpu SandyBridge -m 256M
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
+----------------------------------------------------------------+-----------+------------+
| | v4.19-rc5 | 65d69a0593 |
+----------------------------------------------------------------+-----------+------------+
| boot_successes | 38 | 0 |
| boot_failures | 0 | 6 |
| WARNING:Could_NOT_find_tracepoint_structs_for_some_tracepoints | 0 | 6 |
+----------------------------------------------------------------+-----------+------------+
[ 13.385284] WARNING: Could NOT find tracepoint structs for some tracepoints!
[ 13.386543] -----------------------------------------
[ 13.387472] OK: LOADED SoC Watch Driver
[ 13.388183] -----------------------------------------
[ 13.397539] NET: Registered protocol family 26
[ 13.398667] NET: Registered protocol family 10
[ 13.399578] _warn_unseeded_randomness: 2 callbacks suppressed
[ 13.399585] random: get_random_u32 called from neigh_hash_alloc+0x79/0x8d with crng_init=0
[ 13.402803] random: get_random_bytes called from ipv6_regen_rndid+0x16/0x9b with crng_init=0
[ 13.404535] random: get_random_u32 called from bucket_table_alloc+0xfa/0x118 with crng_init=0
[ 13.406080] Segment Routing with IPv6
[ 13.406800] sit: IPv6, IPv4 and MPLS over IPv4 tunneling driver
[ 13.408580] NET: Registered protocol family 17
[ 13.409404] NET: Registered protocol family 15
[ 13.410209] NET: Registered protocol family 5
[ 13.411054] NET: Registered protocol family 9
[ 13.411868] X25: Linux Version 0.2
[ 13.412503] NET4: DECnet for Linux: V.2.5.68s (C) 1995-2003 Linux DECnet Project Team
[ 13.414017] DECnet: Routing cache hash table of 256 buckets, 11Kbytes
[ 13.415219] NET: Registered protocol family 12
[ 13.416041] 8021q: 802.1Q VLAN Support v1.8
[ 13.416815] 9pnet: Installing 9P2000 support
[ 13.417693] NET: Registered protocol family 36
[ 13.418998] batman_adv: B.A.T.M.A.N. advanced 2018.2 (compatibility version 15) loaded
[ 13.421089] ... APIC ID: 00000000 (0)
[ 13.421784] ... APIC VERSION: 01050014
[ 13.421832] 0000000000000000000000000000000000000000000000000000000000000000
[ 13.421832] 0000000000000000000000000000000000000000000000000000000000000000
[ 13.421832] 0000000000000000000000000000000000000000000000000000000000001000
[ 13.426313] number of MP IRQ sources: 15.
[ 13.427051] number of IO-APIC #0 registers: 24.
[ 13.427855] testing the IO APIC.......................
[ 13.428776] IO APIC #0......
[ 13.429328] .... register #00: 00000000
[ 13.430025] ....... : physical APIC id: 00
[ 13.430772] ....... : Delivery Type: 0
[ 13.431488] ....... : LTS : 0
[ 13.432209] .... register #01: 00170011
[ 13.432893] ....... : max redirection entries: 17
[ 13.433769] ....... : PRQ implemented: 0
[ 13.434540] ....... : IO APIC version: 11
[ 13.435316] .... register #02: 00000000
[ 13.436024] ....... : arbitration: 00
[ 13.436727] .... IRQ redirection table:
[ 13.437437] IOAPIC 0:
[ 13.437895] pin00, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.439271] pin01, enabled , edge , high, V(23), IRR(0), S(0), logical , D(01), M(0)
[ 13.440637] pin02, enabled , edge , high, V(30), IRR(0), S(0), logical , D(01), M(0)
[ 13.442014] pin03, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.443375] pin04, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.444742] pin05, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.446125] pin06, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.447519] pin07, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.448892] pin08, enabled , edge , high, V(24), IRR(0), S(0), logical , D(01), M(0)
[ 13.450294] pin09, enabled , level, high, V(21), IRR(0), S(0), logical , D(01), M(0)
[ 13.451656] pin0a, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.453056] pin0b, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.454443] pin0c, enabled , edge , high, V(22), IRR(0), S(0), logical , D(01), M(0)
[ 13.455842] pin0d, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.457239] pin0e, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.458624] pin0f, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.460038] pin10, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.461432] pin11, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.462811] pin12, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.464192] pin13, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.465563] pin14, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.466938] pin15, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.468335] pin16, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.469708] pin17, disabled, edge , high, V(00), IRR(0), S(0), physical, D(00), M(0)
[ 13.471064] IRQ to pin mappings:
[ 13.471622] IRQ0 -> 0:2
[ 13.472092] IRQ1 -> 0:1
[ 13.472549] IRQ3 -> 0:3
[ 13.473029] IRQ4 -> 0:4
[ 13.473476] IRQ5 -> 0:5
[ 13.473944] IRQ6 -> 0:6
[ 13.474388] IRQ7 -> 0:7
[ 13.474817] IRQ8 -> 0:8
[ 13.475285] IRQ9 -> 0:9
[ 13.475729] IRQ10 -> 0:10
[ 13.476225] IRQ11 -> 0:11
[ 13.476691] IRQ12 -> 0:12
[ 13.477198] IRQ13 -> 0:13
[ 13.477667] IRQ14 -> 0:14
[ 13.478171] IRQ15 -> 0:15
[ 13.478643] .................................... done.
[ 13.479558] Using IPI Shortcut mode
[ 13.481221] sched_clock: Marking stable (13283604071, 197228228)->(13552011153, -71178854)
[ 13.483335] page_owner is disabled
[ 13.484476] Key type trusted registered
[ 13.485511] Key type encrypted registered
[ 13.486274] AppArmor: AppArmor sha1 policy hashing enabled
[ 13.487801] BIOS EDD facility v0.16 2004-Jun-25, 2 devices found
[ 13.494276] IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready
[ 13.496588] 8021q: adding VLAN 0 to HW filter on device eth0
[ 15.538457] e1000: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: RX
[ 15.540593] IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
[ 15.542696] _warn_unseeded_randomness: 6 callbacks suppressed
[ 15.542703] random: get_random_bytes called from addrconf_dad_kick+0x6e/0x9b with crng_init=0
[ 15.551871] Sending DHCP requests .
[ 15.551884] random: get_random_bytes called from ip_auto_config+0x6c8/0x1302 with crng_init=0
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp qemu -k <bzImage> job-script # job-script is attached in this email
Thanks,
eywa
3 years, 9 months
[flow_dissector] d58e468b11: BUG:unable_to_handle_kernel
by kernel test robot
FYI, we noticed the following commit (built with gcc-7):
commit: d58e468b1112dcd1d5193c0a89ff9f98b5a3e8b9 ("flow_dissector: implements flow dissector BPF hook")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
in testcase: test_bpf
with following parameters:
test: jit
on test machine: qemu-system-x86_64 -enable-kvm -cpu host -smp 2 -m 2G
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
+------------------------------------------+------------+------------+
| | 1edb6e035e | d58e468b11 |
+------------------------------------------+------------+------------+
| boot_successes | 16 | 0 |
| boot_failures | 0 | 4 |
| BUG:unable_to_handle_kernel | 0 | 4 |
| Oops:#[##] | 0 | 4 |
| RIP:__skb_flow_dissect | 0 | 4 |
| Kernel_panic-not_syncing:Fatal_exception | 0 | 4 |
+------------------------------------------+------------+------------+
[ 11.372447] BUG: unable to handle kernel paging request at 0000000000001288
[ 11.375142] PGD 0 P4D 0
[ 11.375884] Oops: 0000 [#1] SMP PTI
[ 11.376822] CPU: 1 PID: 446 Comm: modprobe Not tainted 4.19.0-rc2-00341-gd58e468 #1
[ 11.378646] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.10.2-1 04/01/2014
[ 11.380605] RIP: 0010:__skb_flow_dissect+0x8e/0x1690
[ 11.381777] Code: 05 f7 d3 79 7e 4d 8d 14 07 41 0f b7 44 24 06 66 89 85 3e ff ff ff 48 85 db 0f 84 12 02 00 00 48 8b 43 10 48 8b 80 e8 04 00 00 <48> 8b 90 88 12 00 00 48 85 d2 0f 84 f7 01 00 00 48 8d 4d 92 31 c0
[ 11.385799] RSP: 0018:ffffc900008c7a80 EFLAGS: 00010286
[ 11.387040] RAX: 0000000000000000 RBX: ffff88007e80d600 RCX: 0000000000000000
[ 11.388612] RDX: ffffc900008c7bb4 RSI: ffffffff828b3cc0 RDI: ffff88007e80d600
[ 11.390178] RBP: ffffc900008c7b80 R08: 0000000000000000 R09: 000000000000000e
[ 11.391747] R10: ffffc900008c7bb4 R11: ffff88007ea34c00 R12: ffffffff828b3cc0
[ 11.393315] R13: 0000000000000008 R14: 000000000000001e R15: ffffc900008c7bb4
[ 11.394870] FS: 00007f64be17e700(0000) GS:ffff880072900000(0000) knlGS:0000000000000000
[ 11.396785] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 11.398112] CR2: 0000000000001288 CR3: 000000007ee8e002 CR4: 00000000000606e0
[ 11.399693] Call Trace:
[ 11.400425] ? vsnprintf+0xf2/0x4b0
[ 11.401335] ? up+0x12/0x60
[ 11.402116] ? _cond_resched+0x19/0x30
[ 11.403069] ? skb_get_poff+0x4b/0xa0
[ 11.404010] ? __kmalloc_reserve+0x2e/0x80
[ 11.405145] skb_get_poff+0x4b/0xa0
[ 11.406050] bpf_skb_get_pay_offset+0xa/0x10
[ 11.407117] ? test_bpf_init+0x496/0x7d2 [test_bpf]
[ 11.408386] ? shmem_getpage_gfp+0x8d6/0xca0
[ 11.409425] ? test_bpf_init+0x7d2/0x7d2 [test_bpf]
[ 11.410693] ? test_bpf_init+0x7d2/0x7d2 [test_bpf]
[ 11.411968] ? do_one_initcall+0x46/0x1e4
[ 11.412974] ? _cond_resched+0x19/0x30
[ 11.413931] ? kmem_cache_alloc_trace+0x3e/0x1e0
[ 11.415043] ? do_init_module+0x5b/0x200
[ 11.416029] ? load_module+0x183c/0x1ce0
[ 11.417014] ? ima_post_read_file+0xe2/0x120
[ 11.418061] ? __do_sys_finit_module+0xe9/0x110
[ 11.419160] ? __do_sys_finit_module+0xe9/0x110
[ 11.420266] ? do_syscall_64+0x5b/0x180
[ 11.421241] ? entry_SYSCALL_64_after_hwframe+0x44/0xa9
[ 11.422471] Modules linked in: test_bpf(+) sr_mod cdrom sg ata_generic pata_acpi sb_edac crct10dif_pclmul crc32_pclmul crc32c_intel ghash_clmulni_intel pcbc ppdev aesni_intel crypto_simd cryptd glue_helper ata_piix serio_raw snd_pcm snd_timer snd soundcore pcspkr libata parport_pc i2c_piix4 floppy parport ip_tables
[ 11.428351] CR2: 0000000000001288
[ 11.429237] ---[ end trace dcd739f33f17d026 ]---
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp qemu -k <bzImage> job-script # job-script is attached in this email
Thanks,
Rong Chen
3 years, 9 months