WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
Yi Zhang
yizhan at redhat.com
Sat Mar 4 22:41:08 PST 2017
Hi
I get bellow WARNING when trying to connect nvmet with nvmecli.
Steps I used:
On target:
1. Use nvmetcli setup nvmet with below json
{
"hosts": [
{
"nqn": "hostnqn"
}
],
"ports": [
{
"addr": {
"adrfam": "ipv4",
"traddr": "172.31.40.4",
"treq": "not specified",
"trsvcid": "1023",
"trtype": "rdma"
},
"portid": 2,
"referrals": [],
"subsystems": [
"testnqn"
]
}
],
"subsystems": [
{
"allowed_hosts": [],
"attr": {
"allow_any_host": "1"
},
"namespaces": [
{
"device": {
"nguid": "ef90689c-6c46-d44c-89c1-4067801309a8",
"path": "/dev/nullb0"
},
"enable": 1,
"nsid": 1
}
],
"nqn": "testnqn"
}
]
}
2. Initiator side:
#nvme connect-all -t rdma -a 172.31.40.4 -s 1023
3. check the kernel log on both target/initiator side.
kernel log:
[ 242.494533] ocrdma0:Using VLAN with PFC is recommended
[ 242.520244] ocrdma0:Using VLAN 0 for this connection
[ 242.652599] ocrdma0:Using VLAN with PFC is recommended
[ 242.676365] ocrdma0:Using VLAN 0 for this connection
[ 242.700476] ocrdma0:Using VLAN with PFC is recommended
[ 242.723497] ocrdma0:Using VLAN 0 for this connection
[ 242.812331] nvme nvme0: new ctrl: NQN "nqn.2014-08.org.nvmexpress.discovery", addr 172.31.40.4:1023
[ 242.854149] ocrdma0:Using VLAN with PFC is recommended
[ 242.854149] ocrdma0:Using VLAN 0 for this connection
[ 242.854662] ------------[ cut here ]------------
[ 242.854671] WARNING: CPU: 2 PID: 158 at drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
[ 242.854671] failed to drain send queue: -22
[ 242.854671] Modules linked in: nvme_rdma nvme_fabrics nvme_core sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ocrdma ib_core ipmi_ssif intel_rapl x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul crc32_pclmul ghash_clmulni_intel intel_cstate intel_uncore ipmi_si ipmi_devintf hpilo hpwdt intel_rapl_perf gpio_ich iTCO_wdt iTCO_vendor_support pcspkr ie31200_edac shpchp sg edac_core acpi_power_meter lpc_ich ipmi_msghandler pcc_cpufreq acpi_cpufreq nfsd auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod tg3 ptp mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt
[ 242.854694] fb_sys_fops ttm drm ahci libahci libata crc32c_intel be2net fjes serio_raw i2c_core pps_core dm_mirror dm_region_hash dm_log dm_mod
[ 242.854699] CPU: 2 PID: 158 Comm: kworker/2:2 Not tainted 4.10.0 #2
[ 242.854699] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 09/01/2013
[ 242.854702] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[ 242.854702] Call Trace:
[ 242.854706] dump_stack+0x63/0x87
[ 242.854711] __warn+0xd1/0xf0
[ 242.854712] warn_slowpath_fmt+0x5f/0x80
[ 242.854715] ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[ 242.854718] __ib_drain_sq+0x182/0x1c0 [ib_core]
[ 242.854721] ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[ 242.854724] ib_drain_sq+0x25/0x30 [ib_core]
[ 242.854727] ib_drain_qp+0x12/0x30 [ib_core]
[ 242.854728] nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[ 242.854729] nvme_rdma_destroy_admin_queue+0x4d/0xc0 [nvme_rdma]
[ 242.854730] nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[ 242.854731] __nvme_rdma_remove_ctrl+0x84/0x90 [nvme_rdma]
[ 242.854732] nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[ 242.854733] process_one_work+0x165/0x410
[ 242.854734] worker_thread+0x27f/0x4c0
[ 242.854735] kthread+0x101/0x140
[ 242.854736] ? rescuer_thread+0x3b0/0x3b0
[ 242.854737] ? kthread_park+0x90/0x90
[ 242.854751] ret_from_fork+0x2c/0x40
[ 242.854752] ---[ end trace 87cf24af407fe063 ]---
[ 242.855216] ------------[ cut here ]------------
[ 242.855220] WARNING: CPU: 2 PID: 158 at drivers/infiniband/core/verbs.c:2010 __ib_drain_rq+0x177/0x1c0 [ib_core]
[ 242.855220] failed to drain recv queue: -22
[ 242.855220] Modules linked in: nvme_rdma nvme_fabrics nvme_core sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ocrdma ib_core ipmi_ssif intel_rapl x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul crc32_pclmul ghash_clmulni_intel intel_cstate intel_uncore ipmi_si ipmi_devintf hpilo hpwdt intel_rapl_perf gpio_ich iTCO_wdt iTCO_vendor_support pcspkr ie31200_edac shpchp sg edac_core acpi_power_meter lpc_ich ipmi_msghandler pcc_cpufreq acpi_cpufreq nfsd auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod tg3 ptp mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt
[ 242.855232] fb_sys_fops ttm drm ahci libahci libata crc32c_intel be2net fjes serio_raw i2c_core pps_core dm_mirror dm_region_hash dm_log dm_mod
[ 242.855235] CPU: 2 PID: 158 Comm: kworker/2:2 Tainted: G W 4.10.0 #2
[ 242.855235] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 09/01/2013
[ 242.855237] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[ 242.855237] Call Trace:
[ 242.855238] dump_stack+0x63/0x87
[ 242.855239] __warn+0xd1/0xf0
[ 242.855240] warn_slowpath_fmt+0x5f/0x80
[ 242.855242] ? ocrdma_post_recv+0x127/0x140 [ocrdma]
[ 242.855243] ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[ 242.855246] __ib_drain_rq+0x177/0x1c0 [ib_core]
[ 242.855268] ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[ 242.855271] ib_drain_rq+0x25/0x30 [ib_core]
[ 242.855273] ib_drain_qp+0x24/0x30 [ib_core]
[ 242.855274] nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[ 242.855275] nvme_rdma_destroy_admin_queue+0x4d/0xc0 [nvme_rdma]
[ 242.855276] nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[ 242.855277] __nvme_rdma_remove_ctrl+0x84/0x90 [nvme_rdma]
[ 242.855278] nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[ 242.855279] process_one_work+0x165/0x410
[ 242.855280] worker_thread+0x27f/0x4c0
[ 242.855282] kthread+0x101/0x140
[ 242.855283] ? rescuer_thread+0x3b0/0x3b0
[ 242.855283] ? kthread_park+0x90/0x90
[ 242.855284] ret_from_fork+0x2c/0x40
[ 242.855285] ---[ end trace 87cf24af407fe064 ]---
[ 244.970302] ocrdma0:Using VLAN with PFC is recommended
[ 244.970302] ocrdma0:Using VLAN 0 for this connection
[ 245.074424] ocrdma0:Using VLAN with PFC is recommended
[ 245.098195] ocrdma0:Using VLAN 0 for this connection
[ 245.122976] ocrdma0:Using VLAN with PFC is recommended
[ 245.147078] ocrdma0:Using VLAN 0 for this connection
[ 245.183682] nvme nvme0: creating 8 I/O queues.
[ 245.242477] ocrdma0:Using VLAN with PFC is recommended
[ 245.266737] ocrdma0:Using VLAN 0 for this connection
[ 245.388146] ocrdma0:Using VLAN with PFC is recommended
[ 245.411813] ocrdma0:Using VLAN 0 for this connection
[ 245.436101] ocrdma0:Using VLAN with PFC is recommended
[ 245.459634] ocrdma0:Using VLAN 0 for this connection
[ 245.485689] ocrdma0:Using VLAN with PFC is recommended
[ 245.512216] ocrdma0:Using VLAN 0 for this connection
[ 245.662331] ocrdma0:Using VLAN with PFC is recommended
[ 245.687422] ocrdma0:Using VLAN 0 for this connection
[ 245.713057] ocrdma0:Using VLAN with PFC is recommended
[ 245.738019] ocrdma0:Using VLAN 0 for this connection
[ 245.766147] ocrdma0:Using VLAN with PFC is recommended
[ 245.791930] ocrdma0:Using VLAN 0 for this connection
[ 245.824506] ocrdma0:Using VLAN with PFC is recommended
[ 245.849968] ocrdma0:Using VLAN 0 for this connection
[ 245.876203] ocrdma0:Using VLAN with PFC is recommended
[ 245.900670] ocrdma0:Using VLAN 0 for this connection
[ 245.926549] ocrdma0:Using VLAN with PFC is recommended
[ 245.949892] ocrdma0:Using VLAN 0 for this connection
[ 246.102218] ocrdma0:Using VLAN with PFC is recommended
[ 246.125597] ocrdma0:Using VLAN 0 for this connection
[ 246.150668] ocrdma0:Using VLAN with PFC is recommended
[ 246.174000] ocrdma0:Using VLAN 0 for this connection
[ 246.199945] ocrdma0:Using VLAN with PFC is recommended
[ 246.223247] ocrdma0:Using VLAN 0 for this connection
[ 246.253534] ocrdma0:Using VLAN with PFC is recommended
[ 246.278590] ocrdma0:Using VLAN 0 for this connection
[ 246.303311] ocrdma0:Using VLAN with PFC is recommended
[ 246.329326] ocrdma0:Using VLAN 0 for this connection
[ 246.358014] ocrdma0:Using VLAN with PFC is recommended
[ 246.382120] ocrdma0:Using VLAN 0 for this connection
[ 246.537506] ocrdma0:Using VLAN with PFC is recommended
[ 246.560855] ocrdma0:Using VLAN 0 for this connection
[ 246.585018] ocrdma0:Using VLAN with PFC is recommended
[ 246.608393] ocrdma0:Using VLAN 0 for this connection
[ 246.634275] ocrdma0:Using VLAN with PFC is recommended
[ 246.658760] ocrdma0:Using VLAN 0 for this connection
[ 246.779071] ocrdma0:Using VLAN with PFC is recommended
[ 246.803813] ocrdma0:Using VLAN 0 for this connection
[ 246.829115] ocrdma0:Using VLAN with PFC is recommended
[ 246.853383] ocrdma0:Using VLAN 0 for this connection
[ 246.880570] ocrdma0:Using VLAN with PFC is recommended
[ 246.905561] ocrdma0:Using VLAN 0 for this connection
[ 247.033429] ocrdma0:Using VLAN with PFC is recommended
[ 247.060779] ocrdma0:Using VLAN 0 for this connection
[ 247.087520] ocrdma0:Using VLAN with PFC is recommended
[ 247.111723] ocrdma0:Using VLAN 0 for this connection
[ 247.701156] nvme nvme0: new ctrl: NQN "testnqn", addr 172.31.40.4:1023
Best Regards,
Yi Zhang
More information about the Linux-nvme
mailing list