kernull NULL pointer observed on initiator side after 'nvmetcli clear' on target side
Sagi Grimberg
sagi at grimberg.me
Mon Mar 13 01:09:56 PDT 2017
>> Yep... looks like we don't take into account that we can't use all the
>> queues now...
>>
>> Does this patch help:
> Still can reproduce the reconnect in 10 seconds issues with the patch,
> here is the log:
>
> [ 193.574183] nvme nvme0: new ctrl: NQN "nvme-subsystem-name", addr
> 172.31.2.3:1023
> [ 193.612039] __nvme_rdma_init_request: changing called
> [ 193.638723] __nvme_rdma_init_request: changing called
> [ 193.661767] __nvme_rdma_init_request: changing called
> [ 193.684579] __nvme_rdma_init_request: changing called
> [ 193.707327] __nvme_rdma_init_request: changing called
> [ 193.730071] __nvme_rdma_init_request: changing called
> [ 193.752896] __nvme_rdma_init_request: changing called
> [ 193.775699] __nvme_rdma_init_request: changing called
> [ 193.798813] __nvme_rdma_init_request: changing called
> [ 193.821257] __nvme_rdma_init_request: changing called
> [ 193.844090] __nvme_rdma_init_request: changing called
> [ 193.866472] __nvme_rdma_init_request: changing called
> [ 193.889375] __nvme_rdma_init_request: changing called
> [ 193.912094] __nvme_rdma_init_request: changing called
> [ 193.934942] __nvme_rdma_init_request: changing called
> [ 193.957688] __nvme_rdma_init_request: changing called
> [ 606.273376] Broke affinity for irq 16
> [ 606.291940] Broke affinity for irq 28
> [ 606.310201] Broke affinity for irq 90
> [ 606.328211] Broke affinity for irq 93
> [ 606.346263] Broke affinity for irq 97
> [ 606.364314] Broke affinity for irq 100
> [ 606.382105] Broke affinity for irq 104
> [ 606.400727] smpboot: CPU 1 is now offline
> [ 616.820505] nvme nvme0: reconnecting in 10 seconds
> [ 626.882747] blk_mq_reinit_tagset: tag is null, continue
> [ 626.914000] nvme nvme0: Connect rejected: status 8 (invalid service ID).
> [ 626.947965] nvme nvme0: rdma_resolve_addr wait failed (-104).
> [ 626.974673] nvme nvme0: Failed reconnect attempt, requeueing...
This is strange...
Is the target alive? I'm assuming it didn't crash here correct?
More information about the Linux-nvme
mailing list