[PATCH RFC] nvme-rdma: support devices with queue size < 32
Marta Rybczynska
mrybczyn at kalray.eu
Tue Mar 28 04:20:22 PDT 2017
----- Mail original -----
>>> Thanks, this looks useful. But wouldn't it be better to do something
>>> like queue_size divided by 2 or 4 to get a better refill latency?
>>
>> That's an interesting question. The max number of requests is already at 3 or 4
>> times
>> of the queue size because of different message types (see Sam's original
>> message in 'NVMe RDMA driver: CX4 send queue fills up when nvme queue depth is
>> low').
>> I guess it would have inflence on configs with bigger latency.
>>
>> I would like to have Sagi's view on this as he's the one who has changed that
>> part in the iSER initiator in 6df5a128f0fde6315a44e80b30412997147f5efd
>
> Hi Marta,
>
> This looks good indeed. IIRC both for IB and iWARP we need to signal
> at least once every send-queue depth (in practice I remember that some
> devices need more than once) so maybe it'll be good to have division by
> 2.
>
> Maybe it'll be better if we do:
>
> static inline bool queue_sig_limit(struct nvme_rdma_queue *queue)
> {
> return (++queue->sig_count % (queue->queue_size / 2)) == 0;
> }
>
> And lose the hard-coded 32 entirely. Care to test that?
Hello Sigi,
I agree with you, we've found a setup where the signalling every queue
depth is not enough and we're testing the division by two that seems
to work fine till now.
In your version in case of queue length > 32 the notifications would
be sent less often that they are now. I'm wondering if it will have
impact on performance and internal card buffering (it seems that
Mellanox buffers are ~100 elements). Wouldn't it create issues?
I'd like see the magic constant removed. From what I can see we
need to have something not exceeding send buffer of the card but
also not lower than the queue depth. What do you think?
Marta
More information about the Linux-nvme
mailing list