nvme-fabrics: crash at nvme connect-all

Sagi Grimberg sagi at lightbits.io
Thu Jun 9 06:48:20 PDT 2016


>>> Steve, did you see this before? I'm wandering if we need some sort
>>> of logic handling with resource limitation in iWARP (global mrs pool...)
>>
>> Haven't seen this.  Does 'cat /sys/kernel/debug/iw_cxgb4/blah/stats' show
>> anything interesting?  Where/why is it crashing?
>>
>
> So this is the failure:
>
> [  703.239462] rdma_rw_init_mrs: failed to allocated 128 MRs
> [  703.239498] failed to init MR pool ret= -12
> [  703.239541] nvmet_rdma: failed to create_qp ret= -12
> [  703.239582] nvmet_rdma: nvmet_rdma_alloc_queue: creating RDMA queue failed
> (-12).
>
> Not sure why it would fail.  I would think my setup would be allocating more
> given I have 16 cores on the host and target.  The debugfs "stats" file I
> mentioned above should show us something if we're running out of adapter
> resources for MR or PBL records.

Note that Marta ran both the host and the target on the same machine.
So, 8 (cores) x 128 (queue entries) x 2 (host and target) gives 2048
MRs...

What is the T5 limitation?



More information about the Linux-nvme mailing list