mlx4_core 0000:07:00.0: swiotlb buffer is full and OOM observed during stress test on reset_controller
Sagi Grimberg
sagi at grimberg.me
Thu Mar 16 09:51:16 PDT 2017
>>>>> Sagi,
>>>>> The release function is placed in global workqueue. I'm not familiar
>>>>> with NVMe design and I don't know all the details, but maybe the
>>>>> proper way will
>>>>> be to create special workqueue with MEM_RECLAIM flag to ensure the
>>>>> progress?
Leon, the release work makes progress, but it is inherently slower
than the establishment work and when we are bombarded with
establishments we have no backpressure...
> I tried with 4.11.0-rc2, and still can reproduced it with less than 2000
> times.
Yi,
Can you try the below (untested) patch:
I'm not at all convinced this is the way to go because it will
slow down all the connect requests, but I'm curious to know
if it'll make the issue go away.
--
diff --git a/drivers/nvme/target/rdma.c b/drivers/nvme/target/rdma.c
index ecc4fe862561..f15fa6e6b640 100644
--- a/drivers/nvme/target/rdma.c
+++ b/drivers/nvme/target/rdma.c
@@ -1199,6 +1199,9 @@ static int nvmet_rdma_queue_connect(struct
rdma_cm_id *cm_id,
}
queue->port = cm_id->context;
+ /* Let inflight queue teardown complete */
+ flush_scheduled_work();
+
ret = nvmet_rdma_cm_accept(cm_id, queue, &event->param.conn);
if (ret)
goto release_queue;
--
Any other good ideas are welcome...
More information about the Linux-nvme
mailing list