kernull NULL pointer observed on initiator side after 'nvmetcli clear' on target side

Yi Zhang yizhan at redhat.com
Tue Mar 14 06:27:28 PDT 2017



On 03/13/2017 04:09 PM, Sagi Grimberg wrote:
>
>>> Yep... looks like we don't take into account that we can't use all the
>>> queues now...
>>>
>>> Does this patch help:
>> Still can reproduce the reconnect in 10 seconds issues with the patch,
>> here is the log:
>>
>> [  193.574183] nvme nvme0: new ctrl: NQN "nvme-subsystem-name", addr
>> 172.31.2.3:1023
>> [  193.612039] __nvme_rdma_init_request: changing called
>> [  193.638723] __nvme_rdma_init_request: changing called
>> [  193.661767] __nvme_rdma_init_request: changing called
>> [  193.684579] __nvme_rdma_init_request: changing called
>> [  193.707327] __nvme_rdma_init_request: changing called
>> [  193.730071] __nvme_rdma_init_request: changing called
>> [  193.752896] __nvme_rdma_init_request: changing called
>> [  193.775699] __nvme_rdma_init_request: changing called
>> [  193.798813] __nvme_rdma_init_request: changing called
>> [  193.821257] __nvme_rdma_init_request: changing called
>> [  193.844090] __nvme_rdma_init_request: changing called
>> [  193.866472] __nvme_rdma_init_request: changing called
>> [  193.889375] __nvme_rdma_init_request: changing called
>> [  193.912094] __nvme_rdma_init_request: changing called
>> [  193.934942] __nvme_rdma_init_request: changing called
>> [  193.957688] __nvme_rdma_init_request: changing called
>> [  606.273376] Broke affinity for irq 16
>> [  606.291940] Broke affinity for irq 28
>> [  606.310201] Broke affinity for irq 90
>> [  606.328211] Broke affinity for irq 93
>> [  606.346263] Broke affinity for irq 97
>> [  606.364314] Broke affinity for irq 100
>> [  606.382105] Broke affinity for irq 104
>> [  606.400727] smpboot: CPU 1 is now offline
>> [  616.820505] nvme nvme0: reconnecting in 10 seconds
>> [  626.882747] blk_mq_reinit_tagset: tag is null, continue
>> [  626.914000] nvme nvme0: Connect rejected: status 8 (invalid 
>> service ID).
>> [  626.947965] nvme nvme0: rdma_resolve_addr wait failed (-104).
>> [  626.974673] nvme nvme0: Failed reconnect attempt, requeueing...
>
> This is strange...
>
> Is the target alive? I'm assuming it didn't crash here correct?
The target was deleted by 'nvmetcli clear' command.
Then on client side, seems it doesn't know the target side was deleted 
and will always reconnecting in 10 seconds.



More information about the Linux-nvme mailing list