WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]

Yi Zhang yizhan at redhat.com
Sun Mar 26 20:18:54 PDT 2017



On 03/25/2017 09:44 AM, Doug Ledford wrote:
> On Sun, 2017-03-05 at 01:41 -0500, Yi Zhang wrote:
>> Hi
>>
>> I get bellow WARNING when trying to connect nvmet with nvmecli.
>>
>> Steps I used:
>> On target:
>> 1. Use nvmetcli setup nvmet with below json
>> {
>>    "hosts": [
>>      {
>>        "nqn": "hostnqn"
>>      }
>>    ],
>>    "ports": [
>>      {
>>        "addr": {
>>          "adrfam": "ipv4",
>>          "traddr": "172.31.40.4",
> For testing ocrdma, you need to use the 172.31.45.0 network vlan.  For
> testing other RoCE cards in our lab, you can use either the 43 or 45
> vlans, but you shouldn't use the untagged interface for RoCE tests.
>
Hi Doug
Thanks for your reminder.
I've tried .45 network vlan with latest upstream, still can get below 
WARNING.

[  232.040920] nvme nvme0: new ctrl: NQN 
"nqn.2014-08.org.nvmexpress.discovery", addr 172.31.45.4:4420
[  232.088354] ------------[ cut here ]------------
[  232.111557] WARNING: CPU: 1 PID: 156 at 
drivers/infiniband/core/verbs.c:1969 __ib_drain_sq+0x16a/0x1b0 [ib_core]
[  232.157379] failed to drain send queue: -22
[  232.176836] Modules linked in: nvme_rdma nvme_fabrics nvme_core 
sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod 
ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp 
scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm 
ib_cm iw_cm ocrdma ib_core intel_rapl x86_pkg_temp_thermal 
intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul 
ipmi_ssif crc32_pclmul gpio_ich hpilo hpwdt iTCO_wdt iTCO_vendor_support 
ghash_clmulni_intel intel_cstate intel_uncore pcc_cpufreq 
intel_rapl_perf ie31200_edac sg shpchp acpi_power_meter acpi_cpufreq 
pcspkr ipmi_si ipmi_devintf edac_core ipmi_msghandler lpc_ich nfsd 
auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod 
mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt 
fb_sys_fops
[  232.511718]  ttm tg3 ahci libahci drm ptp libata crc32c_intel be2net 
i2c_core serio_raw pps_core dm_mirror dm_region_hash dm_log dm_mod
[  232.566744] CPU: 1 PID: 156 Comm: kworker/1:2 Not tainted 4.11.0-rc4 #1
[  232.600307] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 
09/01/2013
[  232.634737] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  232.666109] Call Trace:
[  232.677051]  dump_stack+0x63/0x87
[  232.691930]  __warn+0xd1/0xf0
[  232.705262]  warn_slowpath_fmt+0x5f/0x80
[  232.723238]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  232.747605]  __ib_drain_sq+0x16a/0x1b0 [ib_core]
[  232.768021]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  232.789971]  ib_drain_sq+0x25/0x30 [ib_core]
[  232.809144]  ib_drain_qp+0x12/0x30 [ib_core]
[  232.828273]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  232.855757]  nvme_rdma_destroy_admin_queue+0x60/0xb0 [nvme_rdma]
[  232.882317]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  232.908993]  __nvme_rdma_remove_ctrl+0x8c/0x90 [nvme_rdma]
[  232.933602]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  232.958178]  process_one_work+0x165/0x410
[  232.976871]  worker_thread+0x27f/0x4c0
[  232.993930]  kthread+0x101/0x140
[  233.008415]  ? rescuer_thread+0x3b0/0x3b0
[  233.026434]  ? kthread_park+0x90/0x90
[  233.042937]  ret_from_fork+0x2c/0x40
[  233.059136] ---[ end trace 2da8cf1943c3a50f ]---
[  233.105785] ------------[ cut here ]------------
[  233.128911] WARNING: CPU: 1 PID: 156 at 
drivers/infiniband/core/verbs.c:2003 __ib_drain_rq+0x15f/0x1b0 [ib_core]
[  233.178754] failed to drain recv queue: -22
[  233.197710] Modules linked in: nvme_rdma nvme_fabrics nvme_core 
sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod 
ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp 
scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm 
ib_cm iw_cm ocrdma ib_core intel_rapl x86_pkg_temp_thermal 
intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul 
ipmi_ssif crc32_pclmul gpio_ich hpilo hpwdt iTCO_wdt iTCO_vendor_support 
ghash_clmulni_intel intel_cstate intel_uncore pcc_cpufreq 
intel_rapl_perf ie31200_edac sg shpchp acpi_power_meter acpi_cpufreq 
pcspkr ipmi_si ipmi_devintf edac_core ipmi_msghandler lpc_ich nfsd 
auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod 
mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt 
fb_sys_fops
[  233.526083]  ttm tg3 ahci libahci drm ptp libata crc32c_intel be2net 
i2c_core serio_raw pps_core dm_mirror dm_region_hash dm_log dm_mod
[  233.585357] CPU: 1 PID: 156 Comm: kworker/1:2 Tainted: G W       
4.11.0-rc4 #1
[  233.623342] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 
09/01/2013
[  233.658771] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  233.691614] Call Trace:
[  233.704155]  dump_stack+0x63/0x87
[  233.720829]  __warn+0xd1/0xf0
[  233.736661]  warn_slowpath_fmt+0x5f/0x80
[  233.756792]  ? ocrdma_post_recv+0x127/0x140 [ocrdma]
[  233.781768]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  233.806196]  __ib_drain_rq+0x15f/0x1b0 [ib_core]
[  233.827474]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  233.849871]  ib_drain_rq+0x25/0x30 [ib_core]
[  233.869660]  ib_drain_qp+0x24/0x30 [ib_core]
[  233.889312]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  233.918097]  nvme_rdma_destroy_admin_queue+0x60/0xb0 [nvme_rdma]
[  233.945818]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  233.970971]  __nvme_rdma_remove_ctrl+0x8c/0x90 [nvme_rdma]
[  233.996145]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  234.021216]  process_one_work+0x165/0x410
[  234.040244]  worker_thread+0x27f/0x4c0
[  234.057532]  kthread+0x101/0x140
[  234.072476]  ? rescuer_thread+0x3b0/0x3b0
[  234.090911]  ? kthread_park+0x90/0x90
[  234.107890]  ret_from_fork+0x2c/0x40
[  234.124422] ---[ end trace 2da8cf1943c3a510 ]---
[  234.491570] nvme nvme0: creating 8 I/O queues.
[  235.099065] nvme nvme0: new ctrl: NQN "testnqn", addr 172.31.45.4:4420





More information about the Linux-nvme mailing list