Warning due to "ALSA: hda: intel: More comprehensive PM runtime setup for controller driver"
Kai Vehmanen
kai.vehmanen at linux.intel.com
Fri Nov 19 16:10:18 CET 2021
Hi,
On Fri, 19 Nov 2021, Takashi Iwai wrote:
> On Thu, 18 Nov 2021 23:13:50 +0100, Heiner Kallweit wrote:
> > On 18.11.2021 22:28, Takashi Iwai wrote:
> > >> snd_hda_intel 0000:00:1f.3: Unbalanced pm_runtime_enable!
> > >>
> > >> Not sure how this patch was tested because the warning is obvious.
> > >> The patch doesn't consider what the PCI sub-system does with regard to
> > >> RPM. Have a look at pci_pm_init().
[...]
> > >
> > > Thanks for the notice. It's been through Intel CI and tests on a few
> > > local machines, maybe we haven't checked carefully those errors but
> > > only concentrated on the other issues, as it seems.
ack on this. This did not go through our full CI, but rather I tested
with the local setup I used to bisect this problem that was reported with
linux-next. We specifically got the "Unbalanced pm_runtime" warning on the
first iteration of the patch, but it was clean in the submitted version.
But yeah, a wider test round would have caught this, sorry about that.
> > Having a look at the commit message of "ALSA: hda: fix general protection
> > fault in azx_runtime_idle" the following sounds weird:
> >
> > - pci-driver.c:pm_runtime_put_sync() leads to a call
> > to rpm_idle() which again calls azx_runtime_idle()
> >
> > rpm_idle() is only called if usage_count is 1 when entering
[...]
> > This doesn't seem to happen. Maybe for whatever reason
> > pm_runtime_get_noresume() isn't called in azx_free(), or azx_free()
> > isn't called from remove().
> > I think you should trace the call chain from the PCI core calling
> > remove() to pm_runtime_get_noresume() getting called or not.
>
> Neither of them, supposedly. Now I took a deeper look at the code
> around it and dug into the git log, and found that the likely problem
> was the recent PCI core code refactoring (removal of pci->driver, etc)
> that have been already reverted; that was why linux-next-20211109 was
> broken and linux-next-20211110 worked. With the leftover pci->driver,
> the stale runtime PM callback was called at the pm_runtime_put_sync()
> call in pci_device_remove().
That probably explains it. I specifically saw runtime idle callback
_after_ the PCI remove driver callback (azx_remove()) was run to
completion. And this happened within execution of pci_device_remove(). But
alas I could not hit this problem with post 20211110 linux-next.
> In anyway, I'll drop the invalid calls of pm_runtime_enable() /
> disable() & co. Maybe keeping pm_runtime_forbid() and
> pm_runtime_dont_use_autosuspend() at remove still makes some sense as
> a counter-part for the probe calls, though.
Hmm, I was about to port this change to the SOF driver as well. But if the
outcome is that driver can safely assume RPM callbacks are _not_ called
after remove, then maybe we can keep the SOF implementation of
remove as is.
Br, Kai
More information about the Alsa-devel
mailing list