Re: [PATCH V3 1/2] PCI/ASPM: Add host-bridge API to override default ASPM/CLKPM link state

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Aug 29, 2025 at 12:54:20PM -0700, David Box wrote:
> On Thu, Aug 28, 2025 at 03:43:45PM -0500, Bjorn Helgaas wrote:
> > On Mon, Aug 25, 2025 at 01:35:22PM -0700, David E. Box wrote:
> > > Synthetic PCIe hierarchies, such as those created by Intel VMD, are not
> > > enumerated by firmware and do not receive BIOS-provided ASPM or CLKPM
> > > defaults. Devices in such domains may therefore run without the intended
> > > power management.
> > > 
> > > Add a host-bridge mechanism that lets controller drivers supply their own
> > > defaults. A new aspm_default_link_state field in struct pci_host_bridge is
> > > set via pci_host_set_default_pcie_link_state(). During link initialization,
> > > if this field is non-zero, ASPM and CLKPM defaults come from it instead of
> > > BIOS.
> > > 
> > > This enables drivers like VMD to align link power management with platform
> > > expectations and avoids embedding controller-specific quirks in ASPM core
> > > logic.
> > 
> > I think this kind of sidesteps the real issue.  Drivers for host
> > controllers or PCI devices should tell us about *broken* things, but
> > not about things advertised by the hardware and available for use.
> 
> I agree with the principle. The intent isn’t for VMD (or any controller) to
> override valid platform policy. It’s to handle synthetic domains where the
> platform doesn’t provide any policy path (no effective _OSC/FADT for the child
> hierarchy). In those cases, the controller is the only agent that knows the
> topology and can supply sane defaults.
> 
> I’m happy to tighten the patch to explicitly cover synthetic domains only.
> Instead of an API, we could have a boolean flag 'aspm_synthetic_domain'. When
> set by the controller, we can do:
> 
>     if (host_bridge->aspm_synthetic_domain)
>             link->aspm_default = PCIE_LINK_STATE_ALL;
> 
> This at least addresses your concern about policy decision, leaving it to the
> core to determine how these domains are handled rather than an ABI that lets
> domains set policy.
> 
> > The only documented policy controls I'm aware of for ASPM are:
> > 
> >   - FADT "PCIe ASPM Controls" bit ("if set, OS must not enable ASPM
> >     control on this platform")
> > 
> >   - _OSC negotiation for control of the PCIe Capability (OS is only
> >     allowed to write PCI_EXP_LNKCTL if platform has granted control to
> >     the OS)
> > 
> > I think what we *should* be doing is enabling ASPM when it's
> > advertised, subject to those platform policy controls and user choices
> > like CONFIG_PCIEASPM_PERFORMANCE/POWERSAVE/etc and sysfs attributes.
> > 
> > So basically I think link->aspm_default should be PCIE_LINK_STATE_ALL
> > without drivers doing anything at all.  Maybe we have to carve out
> > exceptions, e.g., "VMD hierarchies are exempt from _OSC," or "devices
> > on x86 systems before 2026 can't enable more ASPM than BIOS did," or
> > whatever.  Is there any baby step we can make in that direction?
> > 
> > This feels a little scary, so feel free to convince me it can't be
> > done :)
> 
> I understand your direction of enabling all advertised states by
> default (subject to FADT/_OSC and user settings). To explore that,
> I’ll send an RFC in parallel with this patch that proposes a baby
> step, e.g.  add instrumentation so we can see where BIOS left
> capabilities unused, and make it opt-in via a boot param so we can
> evaluate impact safely.

The instrumentation, absolutely.  We need something about what was
already enabled and when we change things.

> So this series will handle the VMD gap directly, and the RFC can
> kick off the wider discussion about defaults on ACPI-managed hosts.
> Does that sound like a reasonable approach and split?

I don't really want a parallel approach because I don't think it would
ever converge again.  BUT I think you're still OK for VMD, because I
think the default should be PCIE_LINK_STATE_ALL, and when we carve out
the exceptions that would not be in vmd.c, and it's easy to say that
there's no exception for VMD.




[Index of Archives]     [DMA Engine]     [Linux Coverity]     [Linux USB]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [Greybus]

  Powered by Linux