regression with "nvme-rdma: use new shared CQ mechanism" from v5.9-rc1
Yi Zhang
yi.zhang at redhat.com
Tue Aug 18 12:08:04 EDT 2020
Hello
With Sagi's new blktests nvme rdma[1], I found one regresson that lead nvme/004 hang.
By bisecting, I found it was introduced from [2], could anyone help check this issue, thanks.
[1] https://marc.info/?l=linux-block&m=159738590701657&w=2
# nvme_trtype=rdma ./check nvme/004 -------> hang and never finished
nvme/004 (test nvme and nvmet UUID NS descriptors)
runtime 1.647s ...
# ps aux | grep rdma_rxe
root 1657 0.0 0.0 8292 1632 pts/0 D+ 11:54 0:00 modprobe -r rdma_rxe
# cat /proc/1657/stack
[<0>] disable_device+0xa2/0x130 [ib_core]
[<0>] __ib_unregister_device+0x37/0xa0 [ib_core]
[<0>] ib_unregister_driver+0x9d/0xd0 [ib_core]
[<0>] rxe_module_exit+0x16/0x7d [rdma_rxe]
[<0>] __do_sys_delete_module.constprop.0+0x170/0x2c0
[<0>] do_syscall_64+0x33/0x40
[<0>] entry_SYSCALL_64_after_hwframe+0x44/0xa9
# dmesg
[ 273.577444] run blktests nvme/004 at 2020-08-18 11:54:22
[ 273.615610] rdma_rxe: loaded
[ 273.623675] infiniband rxe0: set active
[ 273.627962] infiniband rxe0: added eno1
[ 273.639520] infiniband rxe1: set down
[ 273.643611] infiniband rxe1: added eno2
[ 273.655617] infiniband rxe0: set down
[ 273.659713] infiniband rxe0: added eno3
[ 273.673854] infiniband rxe0: set down
[ 273.677946] infiniband rxe0: added eno4
[ 273.695954] infiniband rxe0: set active
[ 273.700236] infiniband rxe0: added enp130s0f0
[ 273.717858] infiniband rxe0: set down
[ 273.721954] infiniband rxe0: added enp130s0f1
[ 273.739460] lo speed is unknown, defaulting to 1000
[ 273.744941] lo speed is unknown, defaulting to 1000
[ 273.750413] lo speed is unknown, defaulting to 1000
[ 273.757853] infiniband rxe0: set active
[ 273.762139] infiniband rxe0: added lo
[ 273.766229] lo speed is unknown, defaulting to 1000
[ 273.771681] lo speed is unknown, defaulting to 1000
[ 273.777135] lo speed is unknown, defaulting to 1000
[ 273.992958] loop: module loaded
[ 274.024550] nvmet: adding nsid 1 to subsystem blktests-subsystem-1
[ 274.039541] nvmet_rdma: enabling port 0 (10.16.221.68:4420)
[ 274.066081] nvmet: creating controller 1 for subsystem blktests-subsystem-1 for NQN nqn.2014-08.org.nvmexpress:uuid:e5372c18c860491bb18adb3b5b025a20.
[ 274.081395] nvme nvme0: creating 32 I/O queues.
[ 274.120241] nvme nvme0: mapped 32/0/0 default/read/poll queues.
[ 274.132037] nvme nvme0: new ctrl: NQN "blktests-subsystem-1", addr 10.16.221.68:4420
[ 275.171715] nvme nvme0: Removing ctrl: NQN "blktests-subsystem-1"
[ 275.311794] rdma_rxe: not configured on eno1
[ 275.321210] rdma_rxe: not configured on eno2
[ 275.333410] rdma_rxe: not configured on eno3
[ 275.347527] rdma_rxe: not configured on eno4
[ 275.362830] rdma_rxe: not configured on enp130s0f0
[ 275.383157] rdma_rxe: not configured on enp130s0f1
[ 275.406067] rdma_rxe: not configured on lo
[ 513.938222] infiniband rocep130s0f0: set active
[ 558.930867] infiniband rocep130s0f0: set active
[ 558.948955] infiniband rocep130s0f0: set active
[ 603.930414] infiniband rocep130s0f0: set active
[2]
ca0f1a8055be nvmet-rdma: use new shared CQ mechanism
287f329e3131 nvme-rdma: use new shared CQ mechanism
Best Regards,
Yi Zhang
More information about the Linux-nvme
mailing list