[dm-devel] [PATCH 8/9] dm: Fix two race conditions related to stopping and starting queues
Mike Snitzer
snitzer at redhat.com
Thu Sep 1 20:33:33 UTC 2016
On Thu, Sep 01 2016 at 4:15pm -0400,
Bart Van Assche <bart.vanassche at sandisk.com> wrote:
> On 09/01/2016 12:05 PM, Mike Snitzer wrote:
> >On Thu, Sep 01 2016 at 1:59pm -0400,
> >Bart Van Assche <bart.vanassche at sandisk.com> wrote:
> >>On 09/01/2016 09:12 AM, Mike Snitzer wrote:
> >>>Please see/test the dm-4.8 and dm-4.9 branches (dm-4.9 being rebased
> >>>ontop of dm-4.8):
> >>>https://git.kernel.org/cgit/linux/kernel/git/device-mapper/linux-dm.git/log/?h=dm-4.8
> >>>https://git.kernel.org/cgit/linux/kernel/git/device-mapper/linux-dm.git/log/?h=dm-4.9
> >>
> >>Hello Mike,
> >>
> >>The result of my tests of the dm-4.9 branch is as follows:
> >>* With patch "dm mpath: check if path's request_queue is dying in
> >>activate_path()" I still see every now and then that CPU usage of
> >>one of the kworker threads jumps to 100%.
> >
> >So you're saying that the dying queue check is still needed in the path
> >selector? Would be useful to know why the 100% is occuring. Can you
> >get a stack trace during this time?
>
> Hello Mike,
>
> A few days ago I had already tried to obtain a stack trace with perf
> but the information reported by perf wasn't entirely accurate. What
> I know about that 100% CPU usage is as follows:
> * "dmsetup table" showed three SRP SCSI device nodes but these SRP SCSI
> device nodes were not visible in /sys/block. This means that
> scsi_remove_host() had already removed these from sysfs.
> * hctx->run_work kept being requeued over and over again on the kernel
> thread with name "kworker/3:1H". I assume this means that
> blk_mq_run_hw_queue() was called with the second argument (async) set
> to true. This probably means that the following dm-rq code was
> triggered:
>
> if (map_request(tio, rq, md) == DM_MAPIO_REQUEUE) {
> /* Undo dm_start_request() before requeuing */
> rq_end_stats(md, rq);
> rq_completed(md, rq_data_dir(rq), false);
> return BLK_MQ_RQ_QUEUE_BUSY;
> }
I'm able to easily reproduce this 100% cpu usage using mptest's
test_02_sdev_delete.
'dmsetup suspend --nolockfs --noflush mp' hangs, seems rooted in your
use of blk_mq_freeze_queue():
[ 298.136930] dmsetup D ffff880142cb3b70 0 9478 9414 0x00000080
[ 298.144831] ffff880142cb3b70 ffff880142cb3b28 ffff880330d6cb00 ffff88032d0022f8
[ 298.153132] ffff880142cb4000 ffff88032d0022f8 ffff88032b161800 0000000000000001
[ 298.161438] 0000000000000001 ffff880142cb3b88 ffffffff816c06e5 ffff88032d001aa0
[ 298.169740] Call Trace:
[ 298.172473] [<ffffffff816c06e5>] schedule+0x35/0x80
[ 298.178019] [<ffffffff8131b937>] blk_mq_freeze_queue_wait+0x57/0xc0
[ 298.185116] [<ffffffff810c58c0>] ? prepare_to_wait_event+0xf0/0xf0
[ 298.192117] [<ffffffff8131d92a>] blk_mq_freeze_queue+0x1a/0x20
[ 298.198734] [<ffffffffa000e910>] dm_stop_queue+0x50/0xc0 [dm_mod]
[ 298.205644] [<ffffffffa0001824>] __dm_suspend+0x134/0x1f0 [dm_mod]
[ 298.212649] [<ffffffffa00035b8>] dm_suspend+0xb8/0xd0 [dm_mod]
[ 298.219270] [<ffffffffa000882e>] dev_suspend+0x18e/0x240 [dm_mod]
[ 298.226175] [<ffffffffa00086a0>] ? table_load+0x380/0x380 [dm_mod]
[ 298.233180] [<ffffffffa0009027>] ctl_ioctl+0x1e7/0x4d0 [dm_mod]
[ 298.239890] [<ffffffff81197f00>] ? lru_cache_add_active_or_unevictable+0x10/0xb0
[ 298.248253] [<ffffffffa0009323>] dm_ctl_ioctl+0x13/0x20 [dm_mod]
[ 298.255049] [<ffffffff81227937>] do_vfs_ioctl+0xa7/0x5d0
[ 298.261081] [<ffffffff8112787f>] ? __audit_syscall_entry+0xaf/0x100
[ 298.268178] [<ffffffff8100365d>] ? syscall_trace_enter+0x1dd/0x2c0
[ 298.275179] [<ffffffff81227ed9>] SyS_ioctl+0x79/0x90
[ 298.280821] [<ffffffff81003a47>] do_syscall_64+0x67/0x160
[ 298.286950] [<ffffffff816c4921>] entry_SYSCALL64_slow_path+0x25/0x25
More information about the dm-devel
mailing list