[PATCH V2 5/5] blk-mq: support concurrent queue quiesce/unquiesce
yukuai (C)
yukuai3 at huawei.com
Fri Oct 8 00:13:02 PDT 2021
On 2021/10/08 14:35, Ming Lei wrote:
> On Fri, Oct 08, 2021 at 02:22:39PM +0800, yukuai (C) wrote:
>> On 2021/10/08 13:10, Ming Lei wrote:
>>> Hello yukuai,
>>>
>>> On Fri, Oct 08, 2021 at 11:22:38AM +0800, yukuai (C) wrote:
>>>> On 2021/10/05 10:31, Ming Lei wrote:
>>>>> On Thu, Sep 30, 2021 at 08:56:29AM -0700, Bart Van Assche wrote:
>>>>>> On 9/30/21 5:56 AM, Ming Lei wrote:
>>>>>>> Turns out that blk_mq_freeze_queue() isn't stronger[1] than
>>>>>>> blk_mq_quiesce_queue() because dispatch may still be in-progress after
>>>>>>> queue is frozen, and in several cases, such as switching io scheduler,
>>>>>>> updating nr_requests & wbt latency, we still need to quiesce queue as a
>>>>>>> supplement of freezing queue.
>>>>>>
>>>>>> Is there agreement about this? If not, how about leaving out the above from the
>>>>>> patch description?
>>>>>
>>>>> Yeah, actually the code has been merged, please see the related
>>>>> functions: elevator_switch(), queue_wb_lat_store() and
>>>>> blk_mq_update_nr_requests().
>>>>>
>>>>>>
>>>>>>> As we need to extend uses of blk_mq_quiesce_queue(), it is inevitable
>>>>>>> for us to need support nested quiesce, especially we can't let
>>>>>>> unquiesce happen when there is quiesce originated from other contexts.
>>>>>>>
>>>>>>> This patch introduces q->mq_quiesce_depth to deal concurrent quiesce,
>>>>>>> and we only unquiesce queue when it is the last/outer-most one of all
>>>>>>> contexts.
>>>>>>>
>>>>>>> One kernel panic issue has been reported[2] when running stress test on
>>>>>>> dm-mpath's updating nr_requests and suspending queue, and the similar
>>>>>>> issue should exist on almost all drivers which use quiesce/unquiesce.
>>>>>>>
>>>>>>> [1] https://marc.info/?l=linux-block&m=150993988115872&w=2
>>>>>>> [2] https://listman.redhat.com/archives/dm-devel/2021-September/msg00189.html
>>>>>>
>>>>>> Please share the call stack of the kernel oops fixed by [2] since that
>>>>>> call stack is not in the patch description.
>>>>>
>>>>> OK, it is something like the following:
>>>>>
>>>>> [ 145.453672] Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.12.0-2.fc30 04/01/2014
>>>>> [ 145.454104] RIP: 0010:dm_softirq_done+0x46/0x220 [dm_mod]
>>>>> [ 145.454536] Code: 85 ed 0f 84 40 01 00 00 44 0f b6 b7 70 01 00 00 4c 8b a5 18 01 00 00 45 89 f5 f6 47 1d 04 75 57 49 8b 7c 24 08 48 85 ff 74 4d <48> 8b 47 08 48 8b 40 58 48 85 c0 74 40 49 8d 4c 24 50 44 89 f2 48
>>>>> [ 145.455423] RSP: 0000:ffffa88600003ef8 EFLAGS: 00010282
>>>>> [ 145.455865] RAX: ffffffffc03fbd10 RBX: ffff979144c00010 RCX: dead000000000200
>>>>> [ 145.456321] RDX: ffffa88600003f30 RSI: ffff979144c00068 RDI: ffffa88600d01040
>>>>> [ 145.456764] RBP: ffff979150eb7990 R08: ffff9791bbc27de0 R09: 0000000000000100
>>>>> [ 145.457205] R10: 0000000000000068 R11: 000000000000004c R12: ffff979144c00138
>>>>> [ 145.457647] R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000010
>>>>> [ 145.458080] FS: 00007f57e5d13180(0000) GS:ffff9791bbc00000(0000) knlGS:0000000000000000
>>>>> [ 145.458516] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
>>>>> [ 145.458945] CR2: ffffa88600d01048 CR3: 0000000106cf8003 CR4: 0000000000370ef0
>>>>> [ 145.459382] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
>>>>> [ 145.459815] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
>>>>> [ 145.460250] Call Trace:
>>>>> [ 145.460779] <IRQ>
>>>>> [ 145.461453] blk_done_softirq+0xa1/0xd0
>>>>> [ 145.462138] __do_softirq+0xd7/0x2d6
>>>>> [ 145.462814] irq_exit+0xf7/0x100
>>>>> [ 145.463480] do_IRQ+0x7f/0xd0
>>>>> [ 145.464131] common_interrupt+0xf/0xf
>>>>> [ 145.464797] </IRQ>
>>>>
>>>> Hi, out test can repoduce this problem:
>>>>
>>>> [ 139.158093] BUG: kernel NULL pointer dereference, address:
>>>> 0000000000000008
>>>> [ 139.160285] #PF: supervisor read access in kernel mode
>>>> [ 139.161905] #PF: error_code(0x0000) - not-present page
>>>> [ 139.163513] PGD 172745067 P4D 172745067 PUD 17fa88067 PMD 0
>>>> [ 139.164506] Oops: 0000 [#1] PREEMPT SMP
>>>> [ 139.165034] CPU: 17 PID: 1083 Comm: nbd-client Not tainted
>>>> 5.15.0-rc4-next-20211007-dirty #94
>>>> [ 139.166179] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS
>>>> ?-20190727_073836-buildvm-p4
>>>> [ 139.167962] RIP: 0010:kyber_has_work+0x31/0xb0
>>>> [ 139.168571] Code: 41 bd 48 00 00 00 41 54 45 31 e4 55 53 48 8b 9f b0 00
>>>> 00 00 48 8d 6b 08 49 63 c4 d
>>>> [ 139.171039] RSP: 0018:ffffc90000f479c8 EFLAGS: 00010246
>>>> [ 139.171740] RAX: 0000000000000000 RBX: 0000000000000000 RCX:
>>>> ffff888176218f40
>>>> [ 139.172680] RDX: ffffffffffffffff RSI: ffffc90000f479f4 RDI:
>>>> ffff888175310000
>>>> [ 139.173611] RBP: 0000000000000008 R08: 0000000000000000 R09:
>>>> ffff88882fa6c0a8
>>>> [ 139.174541] R10: 000000000000030e R11: ffff88817fbcfa10 R12:
>>>> 0000000000000000
>>>> [ 139.175482] R13: 0000000000000048 R14: ffffffff99b7e340 R15:
>>>> ffff8881783edc00
>>>> [ 139.176402] FS: 00007fa8e62e4b40(0000) GS:ffff88882fa40000(0000)
>>>> knlGS:0000000000000000
>>>> [ 139.177434] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
>>>> [ 139.178190] CR2: 0000000000000008 CR3: 00000001796ac000 CR4:
>>>> 00000000000006e0
>>>> [ 139.179127] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
>>>> 0000000000000000
>>>> [ 139.180066] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7:
>>>> 0000000000000400
>>>> [ 139.181000] Call Trace:
>>>> [ 139.182338] <TASK>
>>>> [ 139.182638] blk_mq_run_hw_queue+0x135/0x180
>>>> [ 139.183207] blk_mq_run_hw_queues+0x80/0x150
>>>> [ 139.183766] blk_mq_unquiesce_queue+0x33/0x40
>>>> [ 139.184329] nbd_clear_que+0x52/0xb0 [nbd]
>>>> [ 139.184869] nbd_disconnect_and_put+0x6b/0xe0 [nbd]
>>>> [ 139.185499] nbd_genl_disconnect+0x125/0x290 [nbd]
>>>> [ 139.186123] genl_family_rcv_msg_doit.isra.0+0x102/0x1b0
>>>> [ 139.186821] genl_rcv_msg+0xfc/0x2b0
>>>> [ 139.187300] ? nbd_ioctl+0x500/0x500 [nbd]
>>>> [ 139.187847] ? genl_family_rcv_msg_doit.isra.0+0x1b0/0x1b0
>>>> [ 139.188564] netlink_rcv_skb+0x62/0x180
>>>> [ 139.189075] genl_rcv+0x34/0x60
>>>> [ 139.189490] netlink_unicast+0x26d/0x590
>>>> [ 139.190006] netlink_sendmsg+0x3a1/0x6d0
>>>> [ 139.190513] ? netlink_rcv_skb+0x180/0x180
>>>> [ 139.191039] ____sys_sendmsg+0x1da/0x320
>>>> [ 139.191556] ? ____sys_recvmsg+0x130/0x220
>>>> [ 139.192095] ___sys_sendmsg+0x8e/0xf0
>>>> [ 139.192591] ? ___sys_recvmsg+0xa2/0xf0
>>>> [ 139.193102] ? __wake_up_common_lock+0xac/0xe0
>>>> [ 139.193699] __sys_sendmsg+0x6d/0xe0
>>>> [ 139.194167] __x64_sys_sendmsg+0x23/0x30
>>>> [ 139.194675] do_syscall_64+0x35/0x80
>>>> [ 139.195145] entry_SYSCALL_64_after_hwframe+0x44/0xae
>>>> [ 139.195806] RIP: 0033:0x7fa8e59ebb87
>>>> [ 139.196281] Code: 64 89 02 48 c7 c0 ff ff ff ff eb b9 0f 1f 80 00 00 00
>>>> 00 8b 05 6a 2b 2c 00 48 63 8
>>>> [ 139.198715] RSP: 002b:00007ffd50573c38 EFLAGS: 00000246 ORIG_RAX:
>>>> 000000000000002e
>>>> [ 139.199710] RAX: ffffffffffffffda RBX: 0000000001318120 RCX:
>>>> 00007fa8e59ebb87
>>>> [ 139.200643] RDX: 0000000000000000 RSI: 00007ffd50573c70 RDI:
>>>> 0000000000000003
>>>> [ 139.201583] RBP: 00000000013181f0 R08: 0000000000000014 R09:
>>>> 0000000000000002
>>>> [ 139.202512] R10: 0000000000000006 R11: 0000000000000246 R12:
>>>> 0000000001318030
>>>> [ 139.203434] R13: 00007ffd50573c70 R14: 0000000000000001 R15:
>>>> 00000000ffffffff
>>>> [ 139.204364] </TASK>
>>>> [ 139.204652] Modules linked in: nbd
>>>> [ 139.205101] CR2: 0000000000000008
>>>> [ 139.205580] ---[ end trace 0248c57101a02431 ]---
>>>>
>>>> hope the call stack can be helpful.
>>>
>>> Can you share the following info?
>>>
>>> 1) is the above panic triggered with this quiesce patchset or without
>>> it?
>>
>> Without it, of course.
>
> Got it, any chance to test this patchset V2 and see if your nbd issue can be
> fixed?
>
I aready test it and the problem can be fixed by your patch 5.
In fact, I was going to send a patch similar to your patch 5 today,
and found your patch set.
Thanks,
Kuai
More information about the Linux-nvme
mailing list