[bug report][bisected] sysfs: cannot create duplicate filename '/devices/virtual/nvme-subsystem'

Nilay Shroff nilay at linux.ibm.com
Thu Jun 19 09:30:29 PDT 2025



On 6/13/25 9:16 AM, Yi Zhang wrote:
> Hi
> 
> I did bisecting, seems the issue was introduced by this commit:
> 
> 62188639ec16 nvme-multipath: introduce delayed removal of the
> multipath head node
> 
> On Tue, Jun 10, 2025 at 9:22 PM Yi Zhang <yi.zhang at redhat.com> wrote:
>>
>> Hi
>>
>> I reproduced this issue on the latest linux-block/for-next with
>> blktests nvme/fc nvme/061, please help check it and let me know if you
>> need any info/test, thanks.
>>
>> commit: 38f4878b9463 (HEAD, origin/for-next) Merge branch 'block-6.16'
>> into for-next
>>
>> [ 4810.793156] run blktests nvme/061 at 2025-06-10 08:52:48
>> [ 4811.164767] loop0: detected capacity change from 0 to 2097152
>> [ 4811.254997] nvmet: adding nsid 1 to subsystem blktests-subsystem-1
>> [ 4811.746222] nvme nvme0: NVME-FC{0}: create association : host wwpn
>> 0x20001100aa000001  rport wwpn 0x20001100ab000001: NQN
>> "blktests-subsystem-1"
>> [ 4811.750328] (NULL device *): {0:0} Association created
>> [ 4811.751363] nvmet: Created nvm controller 1 for subsystem
>> blktests-subsystem-1 for NQN
>> nqn.2014-08.org.nvmexpress:uuid:0f01fb42-9f7f-4856-b0b3-51e60b8de349.
>> [ 4811.754969] sysfs: cannot create duplicate filename
>> '/devices/virtual/nvme-subsystem'
>> [ 4811.755435] CPU: 18 UID: 0 PID: 14937 Comm: kworker/u290:0 Tainted:
>> G        W           6.15.0+ #1 PREEMPT(voluntary)
>> [ 4811.755451] Tainted: [W]=WARN
>> [ 4811.755455] Hardware name: HP ProLiant DL380 Gen9/ProLiant DL380
>> Gen9, BIOS P89 10/05/2016
>> [ 4811.755462] Workqueue: nvme-wq nvme_fc_connect_ctrl_work [nvme_fc]
>> [ 4811.755495] Call Trace:
>> [ 4811.755500]  <TASK>
>> [ 4811.755507]  dump_stack_lvl+0xac/0xc0
>> [ 4811.755531]  sysfs_warn_dup+0x72/0x90
>> [ 4811.755549]  sysfs_create_dir_ns+0x1f2/0x260
>> [ 4811.755561]  ? __pfx_sysfs_create_dir_ns+0x10/0x10
>> [ 4811.755570]  ? __pfx_do_raw_spin_trylock+0x10/0x10
>> [ 4811.755591]  ? kobject_add_internal+0x218/0x890
>> [ 4811.755605]  ? do_raw_spin_unlock+0x55/0x1f0
>> [ 4811.755617]  kobject_add_internal+0x272/0x890
>> [ 4811.755630]  kobject_add+0x11f/0x1f0
>> [ 4811.755641]  ? __pfx_kobject_add+0x10/0x10
>> [ 4811.755652]  ? __kmalloc_cache_noprof+0x3b6/0x4e0
>> [ 4811.755667]  ? __pfx_do_raw_spin_trylock+0x10/0x10
>> [ 4811.755686]  get_device_parent+0x325/0x430
>> [ 4811.755700]  ? __pfx_klist_children_put+0x10/0x10
>> [ 4811.755717]  device_add+0x203/0x10f0
>> [ 4811.755727]  ? lockdep_init_map_type+0x51/0x270
>> [ 4811.755738]  ? __pfx_device_add+0x10/0x10
>> [ 4811.755748]  ? __init_waitqueue_head+0xcb/0x150
>> [ 4811.755768]  nvme_init_subsystem+0xa5d/0x1470 [nvme_core]
>> [ 4811.755857]  ? __pfx_nvme_identify_ctrl+0x10/0x10 [nvme_core]
>> [ 4811.755912]  ? __pfx_nvme_init_subsystem+0x10/0x10 [nvme_core]
>> [ 4811.755965]  ? blk_mq_run_hw_queue+0x35e/0x530
>> [ 4811.755986]  ? rcu_is_watching+0x11/0xb0
>> [ 4811.756002]  nvme_init_identify+0x21c/0x2290 [nvme_core]
>> [ 4811.756057]  ? rcu_is_watching+0x11/0xb0
>> [ 4811.756064]  ? __pfx_do_raw_spin_trylock+0x10/0x10
>> [ 4811.756075]  ? blk_mq_hw_queue_need_run+0x271/0x3a0
>> [ 4811.756084]  ? rcu_is_watching+0x11/0xb0
>> [ 4811.756092]  ? trace_irq_enable.constprop.0+0x14a/0x1b0
>> [ 4811.756107]  ? rcu_is_watching+0x11/0xb0
>> [ 4811.756113]  ? blk_mq_run_hw_queue+0x2e3/0x530
>> [ 4811.756121]  ? percpu_ref_put_many.constprop.0+0x7b/0x1b0
>> [ 4811.756130]  ? rcu_is_watching+0x11/0xb0
>> [ 4811.756138]  ? __pfx_nvme_init_identify+0x10/0x10 [nvme_core]
>> [ 4811.756192]  ? percpu_ref_put_many.constprop.0+0x80/0x1b0
>> [ 4811.756201]  ? __nvme_submit_sync_cmd+0x1e6/0x330 [nvme_core]
>> [ 4811.756255]  ? nvmf_reg_read32+0xd1/0x1e0 [nvme_fabrics]
>> [ 4811.756276]  ? __pfx_nvmf_reg_read32+0x10/0x10 [nvme_fabrics]
>> [ 4811.756291]  ? percpu_ref_put_many.constprop.0+0x80/0x1b0
>> [ 4811.756305]  ? nvme_wait_ready+0x13f/0x2d0 [nvme_core]
>> [ 4811.756359]  nvme_init_ctrl_finish+0x1c8/0x810 [nvme_core]
>> [ 4811.756414]  ? __pfx_nvme_init_ctrl_finish+0x10/0x10 [nvme_core]
>> [ 4811.756472]  ? nvme_enable_ctrl+0x3e6/0x620 [nvme_core]
>> [ 4811.756523]  ? __pfx_nvme_enable_ctrl+0x10/0x10 [nvme_core]
>> [ 4811.756575]  ? nvme_fc_connect_admin_queue.constprop.0+0xa1a/0xd20 [nvme_fc]
>> [ 4811.756596]  nvme_fc_create_association+0x880/0x1b60 [nvme_fc]
>> [ 4811.756618]  ? __pfx_debug_object_deactivate+0x10/0x10
>> [ 4811.756630]  ? __pfx_nvme_fc_create_association+0x10/0x10 [nvme_fc]
>> [ 4811.756649]  ? rcu_is_watching+0x11/0xb0
>> [ 4811.756660]  nvme_fc_connect_ctrl_work+0x1d/0xb0 [nvme_fc]
>> [ 4811.756676]  ? rcu_is_watching+0x11/0xb0
>> [ 4811.756684]  process_one_work+0x8cd/0x1950
>> [ 4811.756704]  ? __pfx_process_one_work+0x10/0x10
>> [ 4811.756720]  ? assign_work+0x16c/0x240
>> [ 4811.756732]  worker_thread+0x58d/0xcf0
>> [ 4811.756748]  ? __pfx_worker_thread+0x10/0x10
>> [ 4811.756759]  kthread+0x3d8/0x7a0
>> [ 4811.756770]  ? __pfx_kthread+0x10/0x10
>> [ 4811.756781]  ? rcu_is_watching+0x11/0xb0
>> [ 4811.756789]  ? __pfx_kthread+0x10/0x10
>> [ 4811.756799]  ret_from_fork+0x406/0x510
>> [ 4811.756810]  ? __pfx_kthread+0x10/0x10
>> [ 4811.756819]  ret_from_fork_asm+0x1a/0x30
>> [ 4811.756843]  </TASK>
>> [ 4811.786652] kobject: kobject_add_internal failed for nvme-subsystem
>> with -EEXIST, don't try to register things with the same name in the
>> same directory.
>> [ 4811.787439] nvme nvme0: failed to register subsystem device.
>> [ 4811.788282] nvme nvme0: NVME-FC{0}: create_assoc failed, assoc_id
>> 428489c9c100000 ret -17
>> [ 4811.788849] nvme nvme0: NVME-FC{0}: reset: Reconnect attempt failed (-17)
>> [ 4811.789259] nvme nvme0: NVME-FC{0}: Reconnect attempt in 1 seconds
>> [ 4811.790106] nvme nvme0: NVME-FC{0}: new ctrl: NQN
>> "blktests-subsystem-1", hostnqn:
>> nqn.2014-08.org.nvmexpress:uuid:0f01fb42-9f7f-4856-b0b3-51e60b8de349
>> [ 4811.819765] (NULL device *): {0:0} Association deleted
>> [ 4811.835407] (NULL device *): {0:0} Association freed
>> [ 4811.835791] (NULL device *): Disconnect LS failed: No Association
>> [ 4812.216419] nvme nvme1: NVME-FC{1}: create association : host wwpn
>> 0x20001100aa000001  rport wwpn 0x20001100ab000001: NQN
>> "nqn.2014-08.org.nvmexpress.discovery"
>> [ 4812.220686] (NULL device *): {0:0} Association created
>> [ 4812.221710] nvmet: Created discovery controller 1 for subsystem
>> nqn.2014-08.org.nvmexpress.discovery for NQN
>> nqn.2014-08.org.nvmexpress:uuid:0f01fb42-9f7f-4856-b0b3-51e60b8de349.
>> [ 4812.225298] sysfs: cannot create duplicate filename
>> '/devices/virtual/nvme-subsystem'
>> [ 4812.225791] CPU: 3 UID: 0 PID: 10479 Comm: kworker/u289:13 Tainted:
>> G        W           6.15.0+ #1 PREEMPT(voluntary)
>> [ 4812.225808] Tainted: [W]=WARN
>> [ 4812.225811] Hardware name: HP ProLiant DL380 Gen9/ProLiant DL380
>> Gen9, BIOS P89 10/05/2016
>> [ 4812.225817] Workqueue: nvme-wq nvme_fc_connect_ctrl_work [nvme_fc]
>> [ 4812.225848] Call Trace:
>> [ 4812.225852]  <TASK>
>> [ 4812.225859]  dump_stack_lvl+0xac/0xc0
>> [ 4812.225880]  sysfs_warn_dup+0x72/0x90
>> [ 4812.225896]  sysfs_create_dir_ns+0x1f2/0x260
>> [ 4812.225908]  ? __pfx_sysfs_create_dir_ns+0x10/0x10
>> [ 4812.225918]  ? __pfx_do_raw_spin_trylock+0x10/0x10
>> [ 4812.225936]  ? kobject_add_internal+0x218/0x890
>> [ 4812.225949]  ? do_raw_spin_unlock+0x55/0x1f0
>> [ 4812.225962]  kobject_add_internal+0x272/0x890
>> [ 4812.225975]  kobject_add+0x11f/0x1f0
>> [ 4812.225986]  ? __pfx_kobject_add+0x10/0x10
>> [ 4812.225997]  ? __kmalloc_cache_noprof+0x3b6/0x4e0
>> [ 4812.226010]  ? __pfx_do_raw_spin_trylock+0x10/0x10
>> [ 4812.226029]  get_device_parent+0x325/0x430
>> [ 4812.226042]  ? __pfx_klist_children_put+0x10/0x10
>> [ 4812.226058]  device_add+0x203/0x10f0
>> [ 4812.226067]  ? lockdep_init_map_type+0x51/0x270
>> [ 4812.226079]  ? __pfx_device_add+0x10/0x10
>> [ 4812.226089]  ? __init_waitqueue_head+0xcb/0x150
>> [ 4812.226107]  nvme_init_subsystem+0xa5d/0x1470 [nvme_core]
>> [ 4812.226191]  ? __pfx_nvme_identify_ctrl+0x10/0x10 [nvme_core]
>> [ 4812.226245]  ? __pfx_nvme_init_subsystem+0x10/0x10 [nvme_core]
>> [ 4812.226298]  ? blk_mq_run_hw_queue+0x35e/0x530
>> [ 4812.226310]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.226325]  nvme_init_identify+0x21c/0x2290 [nvme_core]
>> [ 4812.226380]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.226386]  ? __pfx_do_raw_spin_trylock+0x10/0x10
>> [ 4812.226398]  ? blk_mq_hw_queue_need_run+0x271/0x3a0
>> [ 4812.226407]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.226414]  ? trace_irq_enable.constprop.0+0x14a/0x1b0
>> [ 4812.226430]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.226436]  ? blk_mq_run_hw_queue+0x2e3/0x530
>> [ 4812.226444]  ? percpu_ref_put_many.constprop.0+0x7b/0x1b0
>> [ 4812.226452]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.226461]  ? __pfx_nvme_init_identify+0x10/0x10 [nvme_core]
>> [ 4812.226514]  ? percpu_ref_put_many.constprop.0+0x80/0x1b0
>> [ 4812.226523]  ? __nvme_submit_sync_cmd+0x1e6/0x330 [nvme_core]
>> [ 4812.226576]  ? nvmf_reg_read32+0xd1/0x1e0 [nvme_fabrics]
>> [ 4812.226595]  ? __pfx_nvmf_reg_read32+0x10/0x10 [nvme_fabrics]
>> [ 4812.226610]  ? percpu_ref_put_many.constprop.0+0x80/0x1b0
>> [ 4812.226623]  ? nvme_wait_ready+0x13f/0x2d0 [nvme_core]
>> [ 4812.226677]  nvme_init_ctrl_finish+0x1c8/0x810 [nvme_core]
>> [ 4812.226732]  ? __pfx_nvme_init_ctrl_finish+0x10/0x10 [nvme_core]
>> [ 4812.226791]  ? nvme_enable_ctrl+0x3e6/0x620 [nvme_core]
>> [ 4812.226842]  ? __pfx_nvme_enable_ctrl+0x10/0x10 [nvme_core]
>> [ 4812.226893]  ? nvme_fc_connect_admin_queue.constprop.0+0xa1a/0xd20 [nvme_fc]
>> [ 4812.226914]  nvme_fc_create_association+0x880/0x1b60 [nvme_fc]
>> [ 4812.226936]  ? __pfx_debug_object_deactivate+0x10/0x10
>> [ 4812.226949]  ? __pfx_nvme_fc_create_association+0x10/0x10 [nvme_fc]
>> [ 4812.226968]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.226979]  nvme_fc_connect_ctrl_work+0x1d/0xb0 [nvme_fc]
>> [ 4812.226995]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.227002]  process_one_work+0x8cd/0x1950
>> [ 4812.227022]  ? __pfx_process_one_work+0x10/0x10
>> [ 4812.227038]  ? assign_work+0x16c/0x240
>> [ 4812.227050]  worker_thread+0x58d/0xcf0
>> [ 4812.227066]  ? __pfx_worker_thread+0x10/0x10
>> [ 4812.227077]  kthread+0x3d8/0x7a0
>> [ 4812.227088]  ? __pfx_kthread+0x10/0x10
>> [ 4812.227098]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.227107]  ? __pfx_kthread+0x10/0x10
>> [ 4812.227117]  ret_from_fork+0x406/0x510
>> [ 4812.227131]  ? __pfx_kthread+0x10/0x10
>> [ 4812.227140]  ret_from_fork_asm+0x1a/0x30
>> [ 4812.227163]  </TASK>
>> [ 4812.256970] kobject: kobject_add_internal failed for nvme-subsystem
>> with -EEXIST, don't try to register things with the same name in the
>> same directory.
>> [ 4812.257789] nvme nvme1: failed to register subsystem device.
>> [ 4812.258585] nvme nvme1: NVME-FC{1}: create_assoc failed, assoc_id
>> 5a306fd184770000 ret -17
>> [ 4812.259155] nvme nvme1: NVME-FC{1}: reset: Reconnect attempt failed (-17)
>> [ 4812.259560] nvme nvme1: NVME-FC{1}: Reconnect attempt in 2 seconds
>> [ 4812.260436] nvme nvme1: NVME-FC{1}: new ctrl: NQN
>> "nqn.2014-08.org.nvmexpress.discovery", hostnqn:
>> nqn.2014-08.org.nvmexpress:uuid:0f01fb42-9f7f-4856-b0b3-51e60b8de349
>> [ 4812.281759] (NULL device *): {0:0} Association deleted
>> [ 4812.301381] (NULL device *): {0:0} Association freed
>> [ 4812.301771] (NULL device *): Disconnect LS failed: No Association
>> [ 4812.823776] nvme nvme0: NVME-FC{0}: create association : host wwpn
>> 0x20001100aa000001  rport wwpn 0x20001100ab000001: NQN
>> "blktests-subsystem-1"
>> [ 4812.827820] (NULL device *): {0:0} Association created
>> [ 4812.828831] nvmet: Created nvm controller 1 for subsystem
>> blktests-subsystem-1 for NQN
>> nqn.2014-08.org.nvmexpress:uuid:0f01fb42-9f7f-4856-b0b3-51e60b8de349.
>> [ 4812.832443] sysfs: cannot create duplicate filename
>> '/devices/virtual/nvme-subsystem'
>> [ 4812.832925] CPU: 18 UID: 0 PID: 14937 Comm: kworker/u290:0 Tainted:
>> G        W           6.15.0+ #1 PREEMPT(voluntary)
>> [ 4812.832941] Tainted: [W]=WARN
>> [ 4812.832945] Hardware name: HP ProLiant DL380 Gen9/ProLiant DL380
>> Gen9, BIOS P89 10/05/2016
>> [ 4812.832951] Workqueue: nvme-wq nvme_fc_connect_ctrl_work [nvme_fc]
>> [ 4812.832980] Call Trace:
>> [ 4812.832985]  <TASK>
>> [ 4812.832992]  dump_stack_lvl+0xac/0xc0
>> [ 4812.833011]  sysfs_warn_dup+0x72/0x90
>> [ 4812.833026]  sysfs_create_dir_ns+0x1f2/0x260
>> [ 4812.833039]  ? __pfx_sysfs_create_dir_ns+0x10/0x10
>> [ 4812.833048]  ? __pfx_do_raw_spin_trylock+0x10/0x10
>> [ 4812.833064]  ? kobject_add_internal+0x218/0x890
>> [ 4812.833077]  ? do_raw_spin_unlock+0x55/0x1f0
>> [ 4812.833089]  kobject_add_internal+0x272/0x890
>> [ 4812.833103]  kobject_add+0x11f/0x1f0
>> [ 4812.833113]  ? __pfx_kobject_add+0x10/0x10
>> [ 4812.833124]  ? __kmalloc_cache_noprof+0x3b6/0x4e0
>> [ 4812.833137]  ? __pfx_do_raw_spin_trylock+0x10/0x10
>> [ 4812.833156]  get_device_parent+0x325/0x430
>> [ 4812.833168]  ? __pfx_klist_children_put+0x10/0x10
>> [ 4812.833183]  device_add+0x203/0x10f0
>> [ 4812.833193]  ? lockdep_init_map_type+0x51/0x270
>> [ 4812.833205]  ? __pfx_device_add+0x10/0x10
>> [ 4812.833215]  ? __init_waitqueue_head+0xcb/0x150
>> [ 4812.833232]  nvme_init_subsystem+0xa5d/0x1470 [nvme_core]
>> [ 4812.833308]  ? __pfx_nvme_identify_ctrl+0x10/0x10 [nvme_core]
>> [ 4812.833363]  ? __pfx_nvme_init_subsystem+0x10/0x10 [nvme_core]
>> [ 4812.833416]  ? blk_mq_run_hw_queue+0x35e/0x530
>> [ 4812.833427]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.833440]  nvme_init_identify+0x21c/0x2290 [nvme_core]
>> [ 4812.833495]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.833502]  ? __pfx_do_raw_spin_trylock+0x10/0x10
>> [ 4812.833514]  ? blk_mq_hw_queue_need_run+0x271/0x3a0
>> [ 4812.833523]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.833530]  ? trace_irq_enable.constprop.0+0x14a/0x1b0
>> [ 4812.833545]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.833551]  ? blk_mq_run_hw_queue+0x2e3/0x530
>> [ 4812.833559]  ? percpu_ref_put_many.constprop.0+0x7b/0x1b0
>> [ 4812.833567]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.833575]  ? __pfx_nvme_init_identify+0x10/0x10 [nvme_core]
>> [ 4812.833629]  ? percpu_ref_put_many.constprop.0+0x80/0x1b0
>> [ 4812.833638]  ? __nvme_submit_sync_cmd+0x1e6/0x330 [nvme_core]
>> [ 4812.833692]  ? nvmf_reg_read32+0xd1/0x1e0 [nvme_fabrics]
>> [ 4812.833711]  ? __pfx_nvmf_reg_read32+0x10/0x10 [nvme_fabrics]
>> [ 4812.833725]  ? percpu_ref_put_many.constprop.0+0x80/0x1b0
>> [ 4812.833739]  ? nvme_wait_ready+0x13f/0x2d0 [nvme_core]
>> [ 4812.833793]  nvme_init_ctrl_finish+0x1c8/0x810 [nvme_core]
>> [ 4812.833848]  ? __pfx_nvme_init_ctrl_finish+0x10/0x10 [nvme_core]
>> [ 4812.833906]  ? nvme_enable_ctrl+0x3e6/0x620 [nvme_core]
>> [ 4812.833958]  ? __pfx_nvme_enable_ctrl+0x10/0x10 [nvme_core]
>> [ 4812.834009]  ? nvme_fc_connect_admin_queue.constprop.0+0xa1a/0xd20 [nvme_fc]
>> [ 4812.834030]  nvme_fc_create_association+0x880/0x1b60 [nvme_fc]
>> [ 4812.834052]  ? __pfx_debug_object_deactivate+0x10/0x10
>> [ 4812.834064]  ? __pfx_nvme_fc_create_association+0x10/0x10 [nvme_fc]
>> [ 4812.834083]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.834094]  nvme_fc_connect_ctrl_work+0x1d/0xb0 [nvme_fc]
>> [ 4812.834110]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.834117]  process_one_work+0x8cd/0x1950
>> [ 4812.834137]  ? __pfx_process_one_work+0x10/0x10
>> [ 4812.834154]  ? assign_work+0x16c/0x240
>> [ 4812.834165]  worker_thread+0x58d/0xcf0
>> [ 4812.834182]  ? __pfx_worker_thread+0x10/0x10
>> [ 4812.834193]  kthread+0x3d8/0x7a0
>> [ 4812.834204]  ? __pfx_kthread+0x10/0x10
>> [ 4812.834215]  ? rcu_is_watching+0x11/0xb0
>> [ 4812.834223]  ? __pfx_kthread+0x10/0x10
>> [ 4812.834233]  ret_from_fork+0x406/0x510
>> [ 4812.834245]  ? __pfx_kthread+0x10/0x10
>> [ 4812.834255]  ret_from_fork_asm+0x1a/0x30
>> [ 4812.834277]  </TASK>
>> [ 4812.862781] kobject: kobject_add_internal failed for nvme-subsystem
>> with -EEXIST, don't try to register things with the same name in the
>> same directory.
>> [ 4812.863540] nvme nvme0: failed to register subsystem device.
>> [ 4812.864338] nvme nvme0: NVME-FC{0}: create_assoc failed, assoc_id
>> d8c1d294d6f70000 ret -17
>> [ 4812.864873] nvme nvme0: NVME-FC{0}: reset: Reconnect attempt failed (-17)
>> [ 4812.865226] nvme nvme0: NVME-FC{0}: Reconnect attempt in 1 seconds
>> [ 4812.879761] (NULL device *): {0:0} Association deleted
>> [ 4812.896413] (NULL device *): {0:0} Association freed
>> [ 4812.896821] (NULL device *): Disconnect LS failed: No Association
>> [ 4813.911770] nvme nvme0: NVME-FC{0}: create association : host wwpn
>> 0x20001100aa000001  rport wwpn 0x20001100ab000001: NQN
>> "blktests-subsystem-1"
>> [ 4813.916771] (NULL device *): {0:0} Association created
>>
>> --

Thank you for the report!

As you suggested, I could now recreate this issue on my setup.
The issue arises from an imbalance in the NVMe subsystem reference
counter while running the blktests nvme/058. As a result, the NVMe
subsystem’s kobject is not properly released, leaving a stale entry
in sysfs. When the NVMe module is subsequently reloaded, and the 
driver attempts to recreate the subsystem’s kobject in sysfs, the
presence of this stale entry causes a name collision, leading to the
observed failure.

The fix will be on its way soon...

Thanks,
--Nilay









More information about the Linux-nvme mailing list