From: Daniel Baluta <daniel.baluta@oss.nxp.com>
To: rjw@rjwysocki.net, len.brown@intel.com,
ranjani.sridharan@linux.intel.com
Cc: daniel.baluta@oss.nxp.com, aisheng.dong@nxp.com,
ulf.hansson@linaro.org, festevam@gmail.com,
alsa-devel@alsa-project.org, linux-pm@vger.kernel.org,
gregkh@linuxfoundation.org, s.hauer@pengutronix.de,
khilman@kernel.org, linux-kernel@vger.kernel.org,
daniel.baluta@nxp.com, pierre-louis.bossart@linux.intel.com,
paul.olaru@nxp.com, linux-imx@nxp.com, kernel@pengutronix.de,
shawnguo@kernel.org, shengjiu.wang@nxp.com,
linux-arm-kernel@lists.infradead.org
Subject: [RFC PATCH v2 2/2] ASoC: SOF: Use multi PM domains helpers
Date: Wed, 4 Mar 2020 14:19:43 +0200 [thread overview]
Message-ID: <20200304121943.28989-3-daniel.baluta@oss.nxp.com> (raw)
In-Reply-To: <20200304121943.28989-1-daniel.baluta@oss.nxp.com>
From: Daniel Baluta <daniel.baluta@nxp.com>
Use dev_multi_pm_attach / dev_multi_pm_detach instead of the hardcoded
version.
Signed-off-by: Daniel Baluta <daniel.baluta@nxp.com>
---
sound/soc/sof/imx/imx8.c | 60 ++++++----------------------------------
1 file changed, 9 insertions(+), 51 deletions(-)
diff --git a/sound/soc/sof/imx/imx8.c b/sound/soc/sof/imx/imx8.c
index b692752b2178..2e7635b697cf 100644
--- a/sound/soc/sof/imx/imx8.c
+++ b/sound/soc/sof/imx/imx8.c
@@ -51,10 +51,7 @@ struct imx8_priv {
struct imx_sc_ipc *sc_ipc;
/* Power domain handling */
- int num_domains;
- struct device **pd_dev;
- struct device_link **link;
-
+ struct dev_multi_pm_domain_data *mpd;
};
static void imx8_get_reply(struct snd_sof_dev *sdev)
@@ -207,7 +204,6 @@ static int imx8_probe(struct snd_sof_dev *sdev)
struct resource res;
u32 base, size;
int ret = 0;
- int i;
priv = devm_kzalloc(&pdev->dev, sizeof(*priv), GFP_KERNEL);
if (!priv)
@@ -218,45 +214,15 @@ static int imx8_probe(struct snd_sof_dev *sdev)
priv->sdev = sdev;
/* power up device associated power domains */
- priv->num_domains = of_count_phandle_with_args(np, "power-domains",
- "#power-domain-cells");
- if (priv->num_domains < 0) {
- dev_err(sdev->dev, "no power-domains property in %pOF\n", np);
- return priv->num_domains;
- }
-
- priv->pd_dev = devm_kmalloc_array(&pdev->dev, priv->num_domains,
- sizeof(*priv->pd_dev), GFP_KERNEL);
- if (!priv->pd_dev)
- return -ENOMEM;
-
- priv->link = devm_kmalloc_array(&pdev->dev, priv->num_domains,
- sizeof(*priv->link), GFP_KERNEL);
- if (!priv->link)
- return -ENOMEM;
-
- for (i = 0; i < priv->num_domains; i++) {
- priv->pd_dev[i] = dev_pm_domain_attach_by_id(&pdev->dev, i);
- if (IS_ERR(priv->pd_dev[i])) {
- ret = PTR_ERR(priv->pd_dev[i]);
- goto exit_unroll_pm;
- }
- priv->link[i] = device_link_add(&pdev->dev, priv->pd_dev[i],
- DL_FLAG_STATELESS |
- DL_FLAG_PM_RUNTIME |
- DL_FLAG_RPM_ACTIVE);
- if (!priv->link[i]) {
- ret = -ENOMEM;
- dev_pm_domain_detach(priv->pd_dev[i], false);
- goto exit_unroll_pm;
- }
- }
+ priv->mpd = dev_multi_pm_attach(&pdev->dev);
+ if (IS_ERR(priv->mpd))
+ return PTR_ERR(priv->mpd);
ret = imx_scu_get_handle(&priv->sc_ipc);
if (ret) {
dev_err(sdev->dev, "Cannot obtain SCU handle (err = %d)\n",
ret);
- goto exit_unroll_pm;
+ goto exit_detach_pm;
}
priv->ipc_dev = platform_device_register_data(sdev->dev, "imx-dsp",
@@ -264,7 +230,7 @@ static int imx8_probe(struct snd_sof_dev *sdev)
pdev, sizeof(*pdev));
if (IS_ERR(priv->ipc_dev)) {
ret = PTR_ERR(priv->ipc_dev);
- goto exit_unroll_pm;
+ goto exit_detach_pm;
}
priv->dsp_ipc = dev_get_drvdata(&priv->ipc_dev->dev);
@@ -328,26 +294,18 @@ static int imx8_probe(struct snd_sof_dev *sdev)
exit_pdev_unregister:
platform_device_unregister(priv->ipc_dev);
-exit_unroll_pm:
- while (--i >= 0) {
- device_link_del(priv->link[i]);
- dev_pm_domain_detach(priv->pd_dev[i], false);
- }
-
+exit_detach_pm:
+ dev_multi_pm_detach(priv->mpd);
return ret;
}
static int imx8_remove(struct snd_sof_dev *sdev)
{
struct imx8_priv *priv = (struct imx8_priv *)sdev->private;
- int i;
platform_device_unregister(priv->ipc_dev);
- for (i = 0; i < priv->num_domains; i++) {
- device_link_del(priv->link[i]);
- dev_pm_domain_detach(priv->pd_dev[i], false);
- }
+ dev_multi_pm_detach(priv->mpd);
return 0;
}
--
2.17.1
next prev parent reply other threads:[~2020-03-04 12:20 UTC|newest]
Thread overview: 7+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-03-04 12:19 [RFC PATCH v2 0/2] Introduce multi PM domains helpers Daniel Baluta
2020-03-04 12:19 ` [RFC PATCH v2 1/2] PM / domains: " Daniel Baluta
2020-04-21 14:00 ` Ulf Hansson
2020-04-21 14:18 ` Daniel Baluta
2020-03-04 12:19 ` Daniel Baluta [this message]
2020-03-04 17:37 ` [RFC PATCH v2 0/2] " Ranjani Sridharan
2020-03-30 8:38 ` Daniel Baluta
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200304121943.28989-3-daniel.baluta@oss.nxp.com \
--to=daniel.baluta@oss.nxp.com \
--cc=aisheng.dong@nxp.com \
--cc=alsa-devel@alsa-project.org \
--cc=daniel.baluta@nxp.com \
--cc=festevam@gmail.com \
--cc=gregkh@linuxfoundation.org \
--cc=kernel@pengutronix.de \
--cc=khilman@kernel.org \
--cc=len.brown@intel.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-imx@nxp.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-pm@vger.kernel.org \
--cc=paul.olaru@nxp.com \
--cc=pierre-louis.bossart@linux.intel.com \
--cc=ranjani.sridharan@linux.intel.com \
--cc=rjw@rjwysocki.net \
--cc=s.hauer@pengutronix.de \
--cc=shawnguo@kernel.org \
--cc=shengjiu.wang@nxp.com \
--cc=ulf.hansson@linaro.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).