[PATCH 10/10] nvme: implement multipath access to nvme subsystems
Bart Van Assche
Bart.VanAssche at wdc.com
Thu Aug 24 13:17:32 PDT 2017
On Thu, 2017-08-24 at 10:59 +0200, hch at lst.de wrote:
> On Wed, Aug 23, 2017 at 06:21:55PM +0000, Bart Van Assche wrote:
> > Since generic_make_request_fast() returns BLK_STS_AGAIN for a dying path:
> > can the same kind of soft lockups occur with the NVMe multipathing code as
> > with the current upstream device mapper multipathing code? See e.g.
> > "[PATCH 3/7] dm-mpath: Do not lock up a CPU with requeuing activity"
> > (https://www.redhat.com/archives/dm-devel/2017-August/msg00124.html).
>
> I suspect the code is not going to hit it because we check the controller
> state before trying to queue I/O on the lower queue. But if you point
> me to a good reproducer test case I'd like to check.
For NVMe over RDMA, how about the simulate_network_failure_loop() function in
https://github.com/bvanassche/srp-test/blob/master/lib/functions? It simulates
a network failure by writing into the reset_controller sysfs attribute.
> Also does the "single queue" case in your mail refer to the old
> request code? nvme only uses blk-mq so it would not hit that.
> But either way I think get_request should be fixed to return
> BLK_STS_IOERR if the queue is dying instead of BLK_STS_AGAIN.
The description in the patch I referred to indeed refers to the old request
code in the block layer. When I prepared that patch I had analyzed the
behavior of the old request code only.
Bart.
More information about the Linux-nvme
mailing list