blktests nvme/052 failure

Chaitanya Kulkarni chaitanyak at nvidia.com
Wed Jun 26 22:24:34 PDT 2024


Hi,

blktests/nvme/052 is failing everytime on nvme-6.11, HEAD :-

commit 62eaa15c6aeef5011d3d41b69b63e02cf280324c (origin/nvme-6.11)
Author: Thomas Song <tsong at purestorage.com>
Date:   Tue Jun 25 08:26:05 2024 -0400

     nvme-multipath: implement "queue-depth" iopolicy

Below is the log.

-ck

[   36.481199] BUG: unable to handle page fault for address: 
00000031004600c1
[   36.482213] #PF: supervisor read access in kernel mode
[   36.482730] #PF: error_code(0x0000) - not-present page
[   36.483254] PGD 0 P4D 0
[   36.483533] Oops: Oops: 0000 [#1] PREEMPT SMP NOPTI
[   36.484045] CPU: 42 PID: 1122 Comm: kworker/u235:3 Tainted: 
G           O     N 6.10.0-rc3nvme+ #69
[   36.484933] Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 
rel-1.14.0-0-g155821a1990b-prebuilt.qemu.org 04/01/2014
[   36.486024] Workqueue: nvme-wq nvme_scan_work [nvme_core]
[   36.486803] RIP: 0010:lockref_get+0x4/0x60
[   36.487233] Code: bc 9f ff b8 01 00 00 00 eb ad e8 d7 27 72 00 0f 1f 
80 00 00 00 00 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 f3 0f 1e 
fa <48> 8b 07 85 c0 75 33 48 89 c2 be 64 00 00 00 48 89 d1 89 d2 48 c1
[   36.488619] RSP: 0018:ffffc900023cbcd0 EFLAGS: 00010202
[   36.488887] RAX: 0000000000000000 RBX: ffff888104a52438 RCX: 
0000000000000002
[   36.489281] RDX: 0000000000037abc RSI: ffffffff8159fa30 RDI: 
00000031004600c1
[   36.489671] RBP: ffffc900023cbd20 R08: 00000031004600c1 R09: 
ffff88981fcb2330
[   36.490070] R10: 0000000000000001 R11: fffffffffff21a29 R12: 
ffff888104a523c8
[   36.490478] R13: ffff888104a52448 R14: 0000000000000001 R15: 
ffff888104a52438
[   36.490861] FS:  0000000000000000(0000) GS:ffff88981fc80000(0000) 
knlGS:0000000000000000
[   36.491286] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[   36.491592] CR2: 00000031004600c1 CR3: 0000000002a1c000 CR4: 
0000000000350ef0
[   36.492004] DR0: ffffffff837ce5c8 DR1: ffffffff837ce5c9 DR2: 
ffffffff837ce5ca
[   36.492405] DR3: ffffffff837ce5cb DR6: 00000000ffff0ff0 DR7: 
0000000000000600
[   36.493359] Call Trace:
[   36.493515]  <TASK>
[   36.493639]  ? __die+0x24/0x70
[   36.493831]  ? page_fault_oops+0x158/0x4e0
[   36.494061]  ? __schedule+0x354/0xb00
[   36.494262]  ? exc_page_fault+0x77/0x170
[   36.494476]  ? asm_exc_page_fault+0x26/0x30
[   36.494710]  ? __pfx_remove_one+0x10/0x10
[   36.494979]  ? lockref_get+0x4/0x60
[   36.495172]  simple_recursive_removal+0x37/0x2d0
[   36.495448]  ? __pfx_remove_one+0x10/0x10
[   36.495698]  debugfs_remove+0x44/0x70
[   36.495910]  nvme_ns_remove+0x3a/0x200 [nvme_core]
[   36.496192]  nvme_remove_invalid_namespaces+0xfd/0x130 [nvme_core]
[   36.496566]  nvme_scan_work+0x2bc/0x5e0 [nvme_core]
[   36.496837]  ? ttwu_do_activate+0x5d/0x1e0
[   36.497062]  process_one_work+0x158/0x360
[   36.497295]  worker_thread+0x2fd/0x410
[   36.497527]  ? __pfx_worker_thread+0x10/0x10
[   36.497766]  kthread+0xd0/0x100
[   36.497956]  ? __pfx_kthread+0x10/0x10
[   36.498188]  ret_from_fork+0x31/0x50
[   36.498417]  ? __pfx_kthread+0x10/0x10
[   36.498628]  ret_from_fork_asm+0x1a/0x30
[   36.498839]  </TASK>
[   36.498975] Modules linked in: loop nvme_loop(O) nvmet(O) 
nvme_keyring nvme_fabrics(O) nvme(O) nvme_core(O) nvme_auth 
snd_seq_dummy snd_hrtimer snd_seq snd_seq_device snd_timer snd soundcore 
bridge stp llc ip6table_mangle ip6table_raw ip6table_security 
iptable_mangle iptable_raw iptable_security ip_set rfkill nf_tables 
nfnetlink ip6table_filter ip6_tables iptable_filter tun sunrpc xfs 
intel_rapl_msr intel_rapl_common ppdev kvm_amd ccp iTCO_wdt 
iTCO_vendor_support parport_pc kvm i2c_i801 joydev parport pcspkr 
i2c_smbus lpc_ich ip_tables crct10dif_pclmul crc32_pclmul crc32c_intel 
ghash_clmulni_intel bochs sha512_ssse3 drm_vram_helper drm_kms_helper 
drm_ttm_helper sha256_ssse3 virtio_net ttm sha1_ssse3 net_failover 
serio_raw failover drm dimlib qemu_fw_cfg ipmi_devintf ipmi_msghandler 
fuse [last unloaded: nvme_auth]

Entering kdb (current=0xffff88817b1a5100, pid 1122) on processor 42 
Oops: (null)
due to oops @ 0xffffffff817aae44
CPU: 42 PID: 1122 Comm: kworker/u235:3 Tainted: G           O N 
6.10.0-rc3nvme+ #69
Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 
rel-1.14.0-0-g155821a1990b-prebuilt.qemu.org 04/01/2014
Workqueue: nvme-wq nvme_scan_work [nvme_core]
RIP: 0010:lockref_get+0x4/0x60
Code: bc 9f ff b8 01 00 00 00 eb ad e8 d7 27 72 00 0f 1f 80 00 00 00 00 
90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 f3 0f 1e fa <48> 8b 07 
85 c0 75 33 48 89 c2 be 64 00 00 00 48 89 d1 89 d2 48 c1
RSP: 0018:ffffc900023cbcd0 EFLAGS: 00010202
RAX: 0000000000000000 RBX: ffff888104a52438 RCX: 0000000000000002
RDX: 0000000000037abc RSI: ffffffff8159fa30 RDI: 00000031004600c1
RBP: ffffc900023cbd20 R08: 00000031004600c1 R09: ffff88981fcb2330
R10: 0000000000000001 R11: fffffffffff21a29 R12: ffff888104a523c8
R13: ffff888104a52448 R14: 0000000000000001 R15: ffff888104a52438
FS:  0000000000000000(0000) GS:ffff88981fc80000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00000031004600c1 CR3: 0000000002a1c000 CR4: 0000000000350ef0
DR0: ffffffff837ce5c8 DR1: ffffffff837ce5c9 DR2: ffffffff837ce5ca
DR3: ffffffff837ce5cb DR6: 00000000ffff0ff0 DR7: 0000000000000400
Call Trace:
  <TASK>
  ? kdb_main_loop+0x31f/0x960
  ? io_napi_sqpoll_busy_poll+0xf0/0x120
  ? kdb_stub+0x1ae/0x3f0
  ? kgdb_cpu_enter+0x2b3/0x610
  ? kgdb_handle_exception+0xbd/0x100
  ? __kgdb_notify+0x30/0xd0
  ? kgdb_notify+0x21/0x40
  ? notifier_call_chain+0x5b/0xd0
  ? notify_die+0x53/0x80
  ? __die+0x51/0x70
  ? page_fault_oops+0x158/0x4e0
  ? __schedule+0x354/0xb00
  ? exc_page_fault+0x77/0x170
  ? asm_exc_page_fault+0x26/0x30
  ? __pfx_remove_one+0x10/0x10
  ? lockref_get+0x4/0x60
  simple_recursive_removal+0x37/0x2d0
  ? __pfx_remove_one+0x10/0x10
  debugfs_remove+0x44/0x70
  nvme_ns_remove+0x3a/0x200 [nvme_core]
  nvme_remove_invalid_namespaces+0xfd/0x130 [nvme_core]
  nvme_scan_work+0x2bc/0x5e0 [nvme_core]
  ? ttwu_do_activate+0x5d/0x1e0
  process_one_work+0x158/0x360
  worker_thread+0x2fd/0x410
  ? __pfx_worker_thread+0x10/0x10
  kthread+0xd0/0x100
  ? __pfx_kthread+0x10/0x10
  ret_from_fork+0x31/0x50
  ? __pfx_kthread+0x10/0x10
  ret_from_fork_asm+0x1a/0x30
  </TASK>




More information about the Linux-nvme mailing list