blktests nvme/052 failure

John Meneghini jmeneghi at redhat.com
Tue Jul 2 12:38:40 PDT 2024


Looks like I didn't test my queue-depth patches with blktests.

I'll pull the new nvme-6.11 branch and try running this test now.

/John

On 6/27/24 01:24, Chaitanya Kulkarni wrote:
> Hi,
> 
> blktests/nvme/052 is failing everytime on nvme-6.11, HEAD :-
> 
> commit 62eaa15c6aeef5011d3d41b69b63e02cf280324c (origin/nvme-6.11)
> Author: Thomas Song <tsong at purestorage.com>
> Date:   Tue Jun 25 08:26:05 2024 -0400
> 
>       nvme-multipath: implement "queue-depth" iopolicy
> 
> Below is the log.
> 
> -ck
> 
> [   36.481199] BUG: unable to handle page fault for address:
> 00000031004600c1
> [   36.482213] #PF: supervisor read access in kernel mode
> [   36.482730] #PF: error_code(0x0000) - not-present page
> [   36.483254] PGD 0 P4D 0
> [   36.483533] Oops: Oops: 0000 [#1] PREEMPT SMP NOPTI
> [   36.484045] CPU: 42 PID: 1122 Comm: kworker/u235:3 Tainted:
> G           O     N 6.10.0-rc3nvme+ #69
> [   36.484933] Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS
> rel-1.14.0-0-g155821a1990b-prebuilt.qemu.org 04/01/2014
> [   36.486024] Workqueue: nvme-wq nvme_scan_work [nvme_core]
> [   36.486803] RIP: 0010:lockref_get+0x4/0x60
> [   36.487233] Code: bc 9f ff b8 01 00 00 00 eb ad e8 d7 27 72 00 0f 1f
> 80 00 00 00 00 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 f3 0f 1e
> fa <48> 8b 07 85 c0 75 33 48 89 c2 be 64 00 00 00 48 89 d1 89 d2 48 c1
> [   36.488619] RSP: 0018:ffffc900023cbcd0 EFLAGS: 00010202
> [   36.488887] RAX: 0000000000000000 RBX: ffff888104a52438 RCX:
> 0000000000000002
> [   36.489281] RDX: 0000000000037abc RSI: ffffffff8159fa30 RDI:
> 00000031004600c1
> [   36.489671] RBP: ffffc900023cbd20 R08: 00000031004600c1 R09:
> ffff88981fcb2330
> [   36.490070] R10: 0000000000000001 R11: fffffffffff21a29 R12:
> ffff888104a523c8
> [   36.490478] R13: ffff888104a52448 R14: 0000000000000001 R15:
> ffff888104a52438
> [   36.490861] FS:  0000000000000000(0000) GS:ffff88981fc80000(0000)
> knlGS:0000000000000000
> [   36.491286] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> [   36.491592] CR2: 00000031004600c1 CR3: 0000000002a1c000 CR4:
> 0000000000350ef0
> [   36.492004] DR0: ffffffff837ce5c8 DR1: ffffffff837ce5c9 DR2:
> ffffffff837ce5ca
> [   36.492405] DR3: ffffffff837ce5cb DR6: 00000000ffff0ff0 DR7:
> 0000000000000600
> [   36.493359] Call Trace:
> [   36.493515]  <TASK>
> [   36.493639]  ? __die+0x24/0x70
> [   36.493831]  ? page_fault_oops+0x158/0x4e0
> [   36.494061]  ? __schedule+0x354/0xb00
> [   36.494262]  ? exc_page_fault+0x77/0x170
> [   36.494476]  ? asm_exc_page_fault+0x26/0x30
> [   36.494710]  ? __pfx_remove_one+0x10/0x10
> [   36.494979]  ? lockref_get+0x4/0x60
> [   36.495172]  simple_recursive_removal+0x37/0x2d0
> [   36.495448]  ? __pfx_remove_one+0x10/0x10
> [   36.495698]  debugfs_remove+0x44/0x70
> [   36.495910]  nvme_ns_remove+0x3a/0x200 [nvme_core]
> [   36.496192]  nvme_remove_invalid_namespaces+0xfd/0x130 [nvme_core]
> [   36.496566]  nvme_scan_work+0x2bc/0x5e0 [nvme_core]
> [   36.496837]  ? ttwu_do_activate+0x5d/0x1e0
> [   36.497062]  process_one_work+0x158/0x360
> [   36.497295]  worker_thread+0x2fd/0x410
> [   36.497527]  ? __pfx_worker_thread+0x10/0x10
> [   36.497766]  kthread+0xd0/0x100
> [   36.497956]  ? __pfx_kthread+0x10/0x10
> [   36.498188]  ret_from_fork+0x31/0x50
> [   36.498417]  ? __pfx_kthread+0x10/0x10
> [   36.498628]  ret_from_fork_asm+0x1a/0x30
> [   36.498839]  </TASK>
> [   36.498975] Modules linked in: loop nvme_loop(O) nvmet(O)
> nvme_keyring nvme_fabrics(O) nvme(O) nvme_core(O) nvme_auth
> snd_seq_dummy snd_hrtimer snd_seq snd_seq_device snd_timer snd soundcore
> bridge stp llc ip6table_mangle ip6table_raw ip6table_security
> iptable_mangle iptable_raw iptable_security ip_set rfkill nf_tables
> nfnetlink ip6table_filter ip6_tables iptable_filter tun sunrpc xfs
> intel_rapl_msr intel_rapl_common ppdev kvm_amd ccp iTCO_wdt
> iTCO_vendor_support parport_pc kvm i2c_i801 joydev parport pcspkr
> i2c_smbus lpc_ich ip_tables crct10dif_pclmul crc32_pclmul crc32c_intel
> ghash_clmulni_intel bochs sha512_ssse3 drm_vram_helper drm_kms_helper
> drm_ttm_helper sha256_ssse3 virtio_net ttm sha1_ssse3 net_failover
> serio_raw failover drm dimlib qemu_fw_cfg ipmi_devintf ipmi_msghandler
> fuse [last unloaded: nvme_auth]
> 
> Entering kdb (current=0xffff88817b1a5100, pid 1122) on processor 42
> Oops: (null)
> due to oops @ 0xffffffff817aae44
> CPU: 42 PID: 1122 Comm: kworker/u235:3 Tainted: G           O N
> 6.10.0-rc3nvme+ #69
> Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS
> rel-1.14.0-0-g155821a1990b-prebuilt.qemu.org 04/01/2014
> Workqueue: nvme-wq nvme_scan_work [nvme_core]
> RIP: 0010:lockref_get+0x4/0x60
> Code: bc 9f ff b8 01 00 00 00 eb ad e8 d7 27 72 00 0f 1f 80 00 00 00 00
> 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 f3 0f 1e fa <48> 8b 07
> 85 c0 75 33 48 89 c2 be 64 00 00 00 48 89 d1 89 d2 48 c1
> RSP: 0018:ffffc900023cbcd0 EFLAGS: 00010202
> RAX: 0000000000000000 RBX: ffff888104a52438 RCX: 0000000000000002
> RDX: 0000000000037abc RSI: ffffffff8159fa30 RDI: 00000031004600c1
> RBP: ffffc900023cbd20 R08: 00000031004600c1 R09: ffff88981fcb2330
> R10: 0000000000000001 R11: fffffffffff21a29 R12: ffff888104a523c8
> R13: ffff888104a52448 R14: 0000000000000001 R15: ffff888104a52438
> FS:  0000000000000000(0000) GS:ffff88981fc80000(0000) knlGS:0000000000000000
> CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> CR2: 00000031004600c1 CR3: 0000000002a1c000 CR4: 0000000000350ef0
> DR0: ffffffff837ce5c8 DR1: ffffffff837ce5c9 DR2: ffffffff837ce5ca
> DR3: ffffffff837ce5cb DR6: 00000000ffff0ff0 DR7: 0000000000000400
> Call Trace:
>    <TASK>
>    ? kdb_main_loop+0x31f/0x960
>    ? io_napi_sqpoll_busy_poll+0xf0/0x120
>    ? kdb_stub+0x1ae/0x3f0
>    ? kgdb_cpu_enter+0x2b3/0x610
>    ? kgdb_handle_exception+0xbd/0x100
>    ? __kgdb_notify+0x30/0xd0
>    ? kgdb_notify+0x21/0x40
>    ? notifier_call_chain+0x5b/0xd0
>    ? notify_die+0x53/0x80
>    ? __die+0x51/0x70
>    ? page_fault_oops+0x158/0x4e0
>    ? __schedule+0x354/0xb00
>    ? exc_page_fault+0x77/0x170
>    ? asm_exc_page_fault+0x26/0x30
>    ? __pfx_remove_one+0x10/0x10
>    ? lockref_get+0x4/0x60
>    simple_recursive_removal+0x37/0x2d0
>    ? __pfx_remove_one+0x10/0x10
>    debugfs_remove+0x44/0x70
>    nvme_ns_remove+0x3a/0x200 [nvme_core]
>    nvme_remove_invalid_namespaces+0xfd/0x130 [nvme_core]
>    nvme_scan_work+0x2bc/0x5e0 [nvme_core]
>    ? ttwu_do_activate+0x5d/0x1e0
>    process_one_work+0x158/0x360
>    worker_thread+0x2fd/0x410
>    ? __pfx_worker_thread+0x10/0x10
>    kthread+0xd0/0x100
>    ? __pfx_kthread+0x10/0x10
>    ret_from_fork+0x31/0x50
>    ? __pfx_kthread+0x10/0x10
>    ret_from_fork_asm+0x1a/0x30
>    </TASK>
> 
> 




More information about the Linux-nvme mailing list