[PATCHv3] nvme-mpath: delete disk after last connection

Hannes Reinecke hare at suse.de
Thu May 6 09:42:59 BST 2021


On 5/6/21 9:43 AM, Christoph Hellwig wrote:
> On Tue, May 04, 2021 at 12:54:14PM -0700, Sagi Grimberg wrote:
>> Yes, I'm not sure I understand your comment Christoph. This addresses an
>> issue with mdraid where hot unplug+replug does not restore the device to
>> the raid group (pci and fabrics alike), where before multipath this used
>> to work.
>>
>> queue_if_no_path is a dm-multipath feature so I'm not entirely clear
>> what is the concern? mdraid on nvme (pci/fabrics) used to work a certain
>> way, with the introduction of nvme-mpath the behavior was broken (as far
>> as I understand from Hannes).
> 
> AFAIK that specific mdraid behavior is also fixed by the uevent patch
> he sent.
> 
It is most emphatically _NOT_.

These two patches are complementary.

To rephrase: with the current behaviour MD is completely hosed once one
NVMe-oF device get removed after ctrl_loss_tmo kicks in.

And _nothing_ will fix that except a system reboot.

_That_ is the issue this patch fixes.

The other patch for sending the uevent is just to tell MD that recovery
can start. But recovery _cannot_ start without this patch.

> 
> I really do not think we should change the mpath behaviors years after
> first adding it.
> 

But only because no-one ever tested MD on nvme-multipath.
It has been broken since day 1.

Cheers,

Hannes
-- 
Dr. Hannes Reinecke		        Kernel Storage Architect
hare at suse.de			               +49 911 74053 688
SUSE Software Solutions Germany GmbH, 90409 Nürnberg
GF: F. Imendörffer, HRB 36809 (AG Nürnberg)



More information about the Linux-nvme mailing list