NVMe fabric multipathing

Christoph Hellwig hch at infradead.org
Tue Oct 10 00:49:23 PDT 2023


On Fri, Oct 06, 2023 at 09:43:58AM +0100, Mark Syms wrote:
> > > Made unavailable for any reason. So failed switch port, failed HBA,
> > > failed SAN controller, etc. We've been "emulating" this for testing
> > > purposes by unbinding the PCI device for the HBA on the NVMe target
> > > device but I expect the same happens for any of those reasons.
> >
> > If you undbind the device there is no way NVMe can keep any knowledge
> > about the connections it has.  If you have a real path failure on
> > the underlying fabric as long as it hasn't given up on reconnects.  You
> > can set the max_reconnects value to -1 to never stop reconnecting.
> 
> That was an unbind on the remote system not on the system that we were
> pulling status from, that should, from the POV of the client host, be
> essentially the same as dropping the switch port (or indeed a
> controller in an HA Pair dropping offline) should it not? We can
> certainly try dropping a switch port by SNMP and see if the behaviour
> is different.

So how does your controller on the host disappear?  It should be trying
to reconnect, and if you set the max_reconnects value to -1 it should
never go away.




More information about the Linux-nvme mailing list