[Bug 112121] New: Some PCIe options cause devices to be removed after syspend

Mike Lothian mike at fireburn.co.uk
Sat Feb 13 15:39:52 PST 2016


Hi

I've just tested this again, I enabled PCI Hotplug & PCIe Hotplug and
nothing - then I noticed I hadn't enabled the ACPI Hotplug driver -
once I did the issue re-appeared

I then had to use testdisk to restore my partition table :'(

I've attached the updated dmesg & my .config

Cheers

Mike

On 8 February 2016 at 13:51, Bjorn Helgaas <bhelgaas at google.com> wrote:
> [+cc linux-pci, NVMe folks,  power management folks]
>
> On Sun, Feb 7, 2016 at 11:04 AM,  <bugzilla-daemon at bugzilla.kernel.org> wrote:
>> https://bugzilla.kernel.org/show_bug.cgi?id=112121
>>
>>             Bug ID: 112121
>>            Summary: Some PCIe options cause devices to be removed after
>>                     syspend
>>            Product: Drivers
>>            Version: 2.5
>>     Kernel Version: 4.5-rc2
>>           Hardware: All
>>                 OS: Linux
>>               Tree: Mainline
>>             Status: NEW
>>           Severity: normal
>>           Priority: P1
>>          Component: PCI
>>           Assignee: drivers_pci at kernel-bugs.osdl.org
>>           Reporter: mike at fireburn.co.uk
>>         Regression: No
>>
>> Created attachment 203091
>>   --> https://bugzilla.kernel.org/attachment.cgi?id=203091&action=edit
>> Dmesg showing PCIe device removals
>>
>> I was having issues with suspend, when the machine was being resumed iommu
>> started removing devices - including my PCIe NVMe drive which contained my root
>> partition
>>
>> The problem showed up with:
>>
>> [*] PCI support
>> [*]   Support mmconfig PCI config space access
>> [*]   PCI Express Port Bus support
>> [*]     PCI Express Hotplug driver
>> [*]     Root Port Advanced Error Reporting support
>> [*]       PCI Express ECRC settings control
>> < >       PCIe AER error injector support
>> -*-     PCI Express ASPM control
>> [ ]       Debug PCI Express ASPM
>>           Default ASPM policy (BIOS default)  --->
>> [*]   Message Signaled Interrupts (MSI and MSI-X)
>> [ ]   PCI Debugging
>> [*]   Enable PCI resource re-allocation detection
>> < >   PCI Stub driver
>> [*]   Interrupts on hypertransport devices
>> [ ] PCI IOV support
>> [*] PCI PRI support
>> -*- PCI PASID support
>>     PCI host controller drivers  ----
>> < > PCCard (PCMCIA/CardBus) support  ----
>> [*] Support for PCI Hotplug  --->
>> < > RapidIO support
>>
>>
>> This is what I have now:
>>
>> [*] PCI support
>> [*]   Support mmconfig PCI config space access
>> [*]   PCI Express Port Bus support
>> [ ]     Root Port Advanced Error Reporting support
>> -*-     PCI Express ASPM control
>> [ ]       Debug PCI Express ASPM
>>           Default ASPM policy (BIOS default)  --->
>> [*]   Message Signaled Interrupts (MSI and MSI-X)
>> [*]   PCI Debugging
>> [ ]   Enable PCI resource re-allocation detection
>> < >   PCI Stub driver
>> [*]   Interrupts on hypertransport devices
>> [ ] PCI IOV support
>> [ ] PCI PRI support
>> [ ] PCI PASID support
>>     PCI host controller drivers  ----
>> < > PCCard (PCMCIA/CardBus) support  ----
>> [ ] Support for PCI Hotplug  ----
>> < > RapidIO support
>>
>> I tried disabling the iommu driver first but it had no effect
>>
>> If people are interested I could play with the above options to see which one
>> causes the issue
>
> My guess is that PCI hotplug is the important one.  It would be nice
> if dmesg contained enough information to connect nvme0n1 to a PCI
> device.  It'd be even nicer if the PCI core noted device removals or
> whatever happened here.
>
> You don't get any more details if you boot with "ignore_loglevel", do you?
>
> Mike, you didn't mark this as a regression, so I assume it's always
> been this way, and we just haven't noticed it because most people
> enable PCI hotplug (or whatever the relevant config option is).
>
> Bjorn



More information about the Linux-nvme mailing list