All of lore.kernel.org
 help / color / mirror / Atom feed
From: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
To: Abhinav Kumar <quic_abhinavk@quicinc.com>,
	Bjorn Andersson <bjorn.andersson@linaro.org>,
	Rob Clark <robdclark@gmail.com>, Sean Paul <sean@poorly.run>
Cc: David Airlie <airlied@linux.ie>,
	linux-arm-msm@vger.kernel.org, dri-devel@lists.freedesktop.org,
	Stephen Boyd <swboyd@chromium.org>,
	freedreno@lists.freedesktop.org
Subject: Re: [PATCH v5 6/6] drm/msm/dpu: move VBIF blocks handling to dpu_rm
Date: Mon, 14 Feb 2022 23:56:28 +0300	[thread overview]
Message-ID: <a2c001ee-c257-5d25-e0a9-dda7d1dcdead@linaro.org> (raw)
In-Reply-To: <e38947b6-461b-ac26-a3f2-d72b439330d4@quicinc.com>

On 14/02/2022 22:53, Abhinav Kumar wrote:
> 
> 
> On 1/21/2022 1:06 PM, Dmitry Baryshkov wrote:
>> Move handling of VBIF blocks into dpu_rm. This serves the purpose of
>> unification of handling of all hardware blocks inside the DPU driver.
>> This removes hand-coded loops in dpu_vbif (which look for necessary VBIF
>> instance by looping through the dpu_kms->hw_vbif and comparing
>> vbif_idx).
>>
>> Signed-off-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
> 
> I have a slightly different idea about this. Let me know what you think.
>  
> VBIF is a bus interface for the dpu to fetch from. I am not sure if 
> pulling it in the RM is right because its not a dedicated HW block like
> the others in the RM.

It's not a hardware block, but a it's still a hardware resource 
(hardware instance). It is described in the hw catalog. Thus I suggested 
moving it to dpu_rm.

As you have seen, from my previous iterations of this patchset, I tried 
  to move things out of dpu_rm. After some hacking, I saw that having 
alloc/free loops in several places seems like a worse idea. So I moved 
dpu_hw_intf back to dpu_rm and then moved dpu_hw_vbif to dpu_rm too.

> 
> But, I agree with your problem statement of hand-coded loops.
> 
> So instead, why dont you just have a helper in the dpu_vbif.c to get
> you the vbif hw for the passed index like, maybe something like this?
> 
> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
> @@ -11,6 +11,19 @@
>   #include "dpu_hw_vbif.h"
>   #include "dpu_trace.h"
> 
> +static dpu_hw_vbif *dpu_vbif_get_hw(struct dpu_kms *dpu_kms, u32 vbif_idx)
> +{
> +       int i;
> +
> +       for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
> +               if (dpu_kms->hw_vbif[i] &&
> +                               dpu_kms->hw_vbif[i]->idx == vbif_idx)
> +                       vbif = dpu_kms->hw_vbif[i];
> +       }
> +
> +       return vbif;
> +}
> +

You see, this code still bears an idea of looping through hw_vbif 
entries looking for the correct one (we can directly access hw_vbif[idx 
- VBIF_0] instead).

And also the alloc/destroy loops are very similar to rm ones, but are 
working against the array in dpu_kms.

One of the previous iterations had neearly the same idea as yours patch 
proposes, but I later abandoned this idea.

I'm trying to place common code nearby, so that there is a less chance 
of an error.

>   /**
>    * _dpu_vbif_wait_for_xin_halt - wait for the xin to halt
>    * @vbif:      Pointer to hardware vbif driver
> @@ -156,11 +169,7 @@ void dpu_vbif_set_ot_limit(struct dpu_kms *dpu_kms,
> 
>          mdp = dpu_kms->hw_mdp;
> 
> -       for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
> -               if (dpu_kms->hw_vbif[i] &&
> -                               dpu_kms->hw_vbif[i]->idx == 
> params->vbif_idx)
> -                       vbif = dpu_kms->hw_vbif[i];
> -       }
> +       vbif = dpu_vbif_get_hw(dpu_kms, params->vbif_idx);
> 
>          if (!vbif || !mdp) {
>                  DRM_DEBUG_ATOMIC("invalid arguments vbif %d mdp %d\n",
> @@ -216,13 +225,7 @@ void dpu_vbif_set_qos_remap(struct dpu_kms *dpu_kms,
>          }
>          mdp = dpu_kms->hw_mdp;
> 
> -       for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
> -               if (dpu_kms->hw_vbif[i] &&
> -                               dpu_kms->hw_vbif[i]->idx == 
> params->vbif_idx) {
> -                       vbif = dpu_kms->hw_vbif[i];
> -                       break;
> -               }
> -       }
> +       vbif = dpu_vbif_get_hw(params->vbif_idx);
> 
> 
> 
>> ---
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h |  1 +
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c     | 28 +--------------------
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h     |  1 -
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c      | 19 ++++++++++++++
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h      | 12 +++++++++
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c    | 26 ++++++-------------
>>   6 files changed, 40 insertions(+), 47 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h
>> index 6417aa28d32c..895e86dabcb6 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h
>> @@ -8,6 +8,7 @@
>>   #include "dpu_hw_catalog.h"
>>   #include "dpu_hw_mdss.h"
>>   #include "dpu_hw_util.h"
>> +#include "dpu_hw_blk.h"
>>   struct dpu_hw_vbif;
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
>> index 47fe11a84a77..4a1983d8561b 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
>> @@ -782,8 +782,6 @@ static long dpu_kms_round_pixclk(struct msm_kms 
>> *kms, unsigned long rate,
>>   static void _dpu_kms_hw_destroy(struct dpu_kms *dpu_kms)
>>   {
>> -    int i;
>> -
>>       if (dpu_kms->hw_intr)
>>           dpu_hw_intr_destroy(dpu_kms->hw_intr);
>>       dpu_kms->hw_intr = NULL;
>> @@ -791,15 +789,6 @@ static void _dpu_kms_hw_destroy(struct dpu_kms 
>> *dpu_kms)
>>       /* safe to call these more than once during shutdown */
>>       _dpu_kms_mmu_destroy(dpu_kms);
>> -    if (dpu_kms->catalog) {
>> -        for (i = 0; i < dpu_kms->catalog->vbif_count; i++) {
>> -            u32 vbif_idx = dpu_kms->catalog->vbif[i].id;
>> -
>> -            if ((vbif_idx < VBIF_MAX) && dpu_kms->hw_vbif[vbif_idx])
>> -                dpu_hw_vbif_destroy(dpu_kms->hw_vbif[vbif_idx]);
>> -        }
>> -    }
>> -
>>       if (dpu_kms->rm_init)
>>           dpu_rm_destroy(&dpu_kms->rm);
>>       dpu_kms->rm_init = false;
>> @@ -1027,7 +1016,7 @@ static int dpu_kms_hw_init(struct msm_kms *kms)
>>   {
>>       struct dpu_kms *dpu_kms;
>>       struct drm_device *dev;
>> -    int i, rc = -EINVAL;
>> +    int rc = -EINVAL;
>>       if (!kms) {
>>           DPU_ERROR("invalid kms\n");
>> @@ -1116,21 +1105,6 @@ static int dpu_kms_hw_init(struct msm_kms *kms)
>>           goto power_error;
>>       }
>> -    for (i = 0; i < dpu_kms->catalog->vbif_count; i++) {
>> -        u32 vbif_idx = dpu_kms->catalog->vbif[i].id;
>> -
>> -        dpu_kms->hw_vbif[i] = dpu_hw_vbif_init(vbif_idx,
>> -                dpu_kms->vbif[vbif_idx], dpu_kms->catalog);
>> -        if (IS_ERR_OR_NULL(dpu_kms->hw_vbif[vbif_idx])) {
>> -            rc = PTR_ERR(dpu_kms->hw_vbif[vbif_idx]);
>> -            if (!dpu_kms->hw_vbif[vbif_idx])
>> -                rc = -EINVAL;
>> -            DPU_ERROR("failed to init vbif %d: %d\n", vbif_idx, rc);
>> -            dpu_kms->hw_vbif[vbif_idx] = NULL;
>> -            goto power_error;
>> -        }
>> -    }
>> -
>>       rc = dpu_core_perf_init(&dpu_kms->perf, dev, dpu_kms->catalog,
>>               _dpu_kms_get_clk(dpu_kms, "core"));
>>       if (rc) {
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h
>> index 3f518c809e33..b96c901483ae 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h
>> @@ -105,7 +105,6 @@ struct dpu_kms {
>>       struct dpu_rm rm;
>>       bool rm_init;
>> -    struct dpu_hw_vbif *hw_vbif[VBIF_MAX];
>>       struct dpu_hw_mdp *hw_mdp;
>>       bool has_danger_ctrl;
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
>> index 7497538adae1..6d49666c4e77 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
>> @@ -76,6 +76,8 @@ int dpu_rm_destroy(struct dpu_rm *rm)
>>       }
>>       for (i = 0; i < ARRAY_SIZE(rm->hw_intf); i++)
>>           dpu_hw_intf_destroy(rm->hw_intf[i]);
>> +    for (i = 0; i < ARRAY_SIZE(rm->hw_vbif); i++)
>> +        dpu_hw_vbif_destroy(rm->hw_vbif[i]);
>>       return 0;
>>   }
>> @@ -210,6 +212,23 @@ int dpu_rm_init(struct dpu_rm *rm,
>>           rm->dspp_blks[dspp->id - DSPP_0] = &hw->base;
>>       }
>> +    for (i = 0; i < cat->vbif_count; i++) {
>> +        struct dpu_hw_vbif *hw;
>> +        const struct dpu_vbif_cfg *vbif = &cat->vbif[i];
>> +
>> +        if (vbif->id < VBIF_0 || vbif->id >= VBIF_MAX) {
>> +            DPU_ERROR("skip vbif %d with invalid id\n", vbif->id);
>> +            continue;
>> +        }
>> +        hw = dpu_hw_vbif_init(vbif->id, mmio, cat);
>> +        if (IS_ERR(hw)) {
>> +            rc = PTR_ERR(hw);
>> +            DPU_ERROR("failed vbif object creation: err %d\n", rc);
>> +            goto fail;
>> +        }
>> +        rm->hw_vbif[vbif->id - VBIF_0] = hw;
>> +    }
>> +
>>       return 0;
>>   fail:
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
>> index 9b13200a050a..a15977bdceeb 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
>> @@ -20,6 +20,7 @@ struct dpu_global_state;
>>    * @ctl_blks: array of ctl hardware resources
>>    * @hw_intf: array of intf hardware resources
>>    * @dspp_blks: array of dspp hardware resources
>> + * @hw_vbif: array of vbif hardware resources
>>    */
>>   struct dpu_rm {
>>       struct dpu_hw_blk *pingpong_blks[PINGPONG_MAX - PINGPONG_0];
>> @@ -28,6 +29,7 @@ struct dpu_rm {
>>       struct dpu_hw_intf *hw_intf[INTF_MAX - INTF_0];
>>       struct dpu_hw_blk *dspp_blks[DSPP_MAX - DSPP_0];
>>       struct dpu_hw_blk *merge_3d_blks[MERGE_3D_MAX - MERGE_3D_0];
>> +    struct dpu_hw_vbif *hw_vbif[VBIF_MAX - VBIF_0];
>>   };
>>   /**
>> @@ -95,5 +97,15 @@ static inline struct dpu_hw_intf 
>> *dpu_rm_get_intf(struct dpu_rm *rm, enum dpu_in
>>       return rm->hw_intf[intf_idx - INTF_0];
>>   }
>> +/**
>> + * dpu_rm_get_vbif - Return a struct dpu_hw_vbif instance given it's 
>> index.
>> + * @rm: DPU Resource Manager handle
>> + * @vbif_idx: VBIF's index
>> + */
>> +static inline struct dpu_hw_vbif *dpu_rm_get_vbif(struct dpu_rm *rm, 
>> enum dpu_vbif vbif_idx)
>> +{
>> +    return rm->hw_vbif[vbif_idx - VBIF_0];
>> +}
>> +
>>   #endif /* __DPU_RM_H__ */
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
>> index 21d20373eb8b..00ac2aa81651 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
>> @@ -152,15 +152,10 @@ void dpu_vbif_set_ot_limit(struct dpu_kms *dpu_kms,
>>       struct dpu_hw_mdp *mdp;
>>       bool forced_on = false;
>>       u32 ot_lim;
>> -    int ret, i;
>> +    int ret;
>>       mdp = dpu_kms->hw_mdp;
>> -
>> -    for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
>> -        if (dpu_kms->hw_vbif[i] &&
>> -                dpu_kms->hw_vbif[i]->idx == params->vbif_idx)
>> -            vbif = dpu_kms->hw_vbif[i];
>> -    }
>> +    vbif = dpu_rm_get_vbif(&dpu_kms->rm, params->vbif_idx);
>>       if (!vbif || !mdp) {
>>           DRM_DEBUG_ATOMIC("invalid arguments vbif %d mdp %d\n",
>> @@ -216,14 +211,7 @@ void dpu_vbif_set_qos_remap(struct dpu_kms *dpu_kms,
>>       }
>>       mdp = dpu_kms->hw_mdp;
>> -    for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
>> -        if (dpu_kms->hw_vbif[i] &&
>> -                dpu_kms->hw_vbif[i]->idx == params->vbif_idx) {
>> -            vbif = dpu_kms->hw_vbif[i];
>> -            break;
>> -        }
>> -    }
>> -
>> +    vbif = dpu_rm_get_vbif(&dpu_kms->rm, params->vbif_idx);
>>       if (!vbif || !vbif->cap) {
>>           DPU_ERROR("invalid vbif %d\n", params->vbif_idx);
>>           return;
>> @@ -261,8 +249,8 @@ void dpu_vbif_clear_errors(struct dpu_kms *dpu_kms)
>>       struct dpu_hw_vbif *vbif;
>>       u32 i, pnd, src;
>> -    for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
>> -        vbif = dpu_kms->hw_vbif[i];
>> +    for (i = VBIF_0; i < VBIF_MAX; i++) {
>> +        vbif = dpu_rm_get_vbif(&dpu_kms->rm, i);
>>           if (vbif && vbif->ops.clear_errors) {
>>               vbif->ops.clear_errors(vbif, &pnd, &src);
>>               if (pnd || src) {
>> @@ -278,8 +266,8 @@ void dpu_vbif_init_memtypes(struct dpu_kms *dpu_kms)
>>       struct dpu_hw_vbif *vbif;
>>       int i, j;
>> -    for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
>> -        vbif = dpu_kms->hw_vbif[i];
>> +    for (i = VBIF_0; i < VBIF_MAX; i++) {
>> +        vbif = dpu_rm_get_vbif(&dpu_kms->rm, i);
>>           if (vbif && vbif->cap && vbif->ops.set_mem_type) {
>>               for (j = 0; j < vbif->cap->memtype_count; j++)
>>                   vbif->ops.set_mem_type(


-- 
With best wishes
Dmitry

WARNING: multiple messages have this Message-ID (diff)
From: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
To: Abhinav Kumar <quic_abhinavk@quicinc.com>,
	Bjorn Andersson <bjorn.andersson@linaro.org>,
	Rob Clark <robdclark@gmail.com>, Sean Paul <sean@poorly.run>
Cc: Stephen Boyd <swboyd@chromium.org>,
	David Airlie <airlied@linux.ie>, Daniel Vetter <daniel@ffwll.ch>,
	linux-arm-msm@vger.kernel.org, dri-devel@lists.freedesktop.org,
	freedreno@lists.freedesktop.org
Subject: Re: [PATCH v5 6/6] drm/msm/dpu: move VBIF blocks handling to dpu_rm
Date: Mon, 14 Feb 2022 23:56:28 +0300	[thread overview]
Message-ID: <a2c001ee-c257-5d25-e0a9-dda7d1dcdead@linaro.org> (raw)
In-Reply-To: <e38947b6-461b-ac26-a3f2-d72b439330d4@quicinc.com>

On 14/02/2022 22:53, Abhinav Kumar wrote:
> 
> 
> On 1/21/2022 1:06 PM, Dmitry Baryshkov wrote:
>> Move handling of VBIF blocks into dpu_rm. This serves the purpose of
>> unification of handling of all hardware blocks inside the DPU driver.
>> This removes hand-coded loops in dpu_vbif (which look for necessary VBIF
>> instance by looping through the dpu_kms->hw_vbif and comparing
>> vbif_idx).
>>
>> Signed-off-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
> 
> I have a slightly different idea about this. Let me know what you think.
>  
> VBIF is a bus interface for the dpu to fetch from. I am not sure if 
> pulling it in the RM is right because its not a dedicated HW block like
> the others in the RM.

It's not a hardware block, but a it's still a hardware resource 
(hardware instance). It is described in the hw catalog. Thus I suggested 
moving it to dpu_rm.

As you have seen, from my previous iterations of this patchset, I tried 
  to move things out of dpu_rm. After some hacking, I saw that having 
alloc/free loops in several places seems like a worse idea. So I moved 
dpu_hw_intf back to dpu_rm and then moved dpu_hw_vbif to dpu_rm too.

> 
> But, I agree with your problem statement of hand-coded loops.
> 
> So instead, why dont you just have a helper in the dpu_vbif.c to get
> you the vbif hw for the passed index like, maybe something like this?
> 
> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
> @@ -11,6 +11,19 @@
>   #include "dpu_hw_vbif.h"
>   #include "dpu_trace.h"
> 
> +static dpu_hw_vbif *dpu_vbif_get_hw(struct dpu_kms *dpu_kms, u32 vbif_idx)
> +{
> +       int i;
> +
> +       for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
> +               if (dpu_kms->hw_vbif[i] &&
> +                               dpu_kms->hw_vbif[i]->idx == vbif_idx)
> +                       vbif = dpu_kms->hw_vbif[i];
> +       }
> +
> +       return vbif;
> +}
> +

You see, this code still bears an idea of looping through hw_vbif 
entries looking for the correct one (we can directly access hw_vbif[idx 
- VBIF_0] instead).

And also the alloc/destroy loops are very similar to rm ones, but are 
working against the array in dpu_kms.

One of the previous iterations had neearly the same idea as yours patch 
proposes, but I later abandoned this idea.

I'm trying to place common code nearby, so that there is a less chance 
of an error.

>   /**
>    * _dpu_vbif_wait_for_xin_halt - wait for the xin to halt
>    * @vbif:      Pointer to hardware vbif driver
> @@ -156,11 +169,7 @@ void dpu_vbif_set_ot_limit(struct dpu_kms *dpu_kms,
> 
>          mdp = dpu_kms->hw_mdp;
> 
> -       for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
> -               if (dpu_kms->hw_vbif[i] &&
> -                               dpu_kms->hw_vbif[i]->idx == 
> params->vbif_idx)
> -                       vbif = dpu_kms->hw_vbif[i];
> -       }
> +       vbif = dpu_vbif_get_hw(dpu_kms, params->vbif_idx);
> 
>          if (!vbif || !mdp) {
>                  DRM_DEBUG_ATOMIC("invalid arguments vbif %d mdp %d\n",
> @@ -216,13 +225,7 @@ void dpu_vbif_set_qos_remap(struct dpu_kms *dpu_kms,
>          }
>          mdp = dpu_kms->hw_mdp;
> 
> -       for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
> -               if (dpu_kms->hw_vbif[i] &&
> -                               dpu_kms->hw_vbif[i]->idx == 
> params->vbif_idx) {
> -                       vbif = dpu_kms->hw_vbif[i];
> -                       break;
> -               }
> -       }
> +       vbif = dpu_vbif_get_hw(params->vbif_idx);
> 
> 
> 
>> ---
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h |  1 +
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c     | 28 +--------------------
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h     |  1 -
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c      | 19 ++++++++++++++
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h      | 12 +++++++++
>>   drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c    | 26 ++++++-------------
>>   6 files changed, 40 insertions(+), 47 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h
>> index 6417aa28d32c..895e86dabcb6 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h
>> @@ -8,6 +8,7 @@
>>   #include "dpu_hw_catalog.h"
>>   #include "dpu_hw_mdss.h"
>>   #include "dpu_hw_util.h"
>> +#include "dpu_hw_blk.h"
>>   struct dpu_hw_vbif;
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
>> index 47fe11a84a77..4a1983d8561b 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
>> @@ -782,8 +782,6 @@ static long dpu_kms_round_pixclk(struct msm_kms 
>> *kms, unsigned long rate,
>>   static void _dpu_kms_hw_destroy(struct dpu_kms *dpu_kms)
>>   {
>> -    int i;
>> -
>>       if (dpu_kms->hw_intr)
>>           dpu_hw_intr_destroy(dpu_kms->hw_intr);
>>       dpu_kms->hw_intr = NULL;
>> @@ -791,15 +789,6 @@ static void _dpu_kms_hw_destroy(struct dpu_kms 
>> *dpu_kms)
>>       /* safe to call these more than once during shutdown */
>>       _dpu_kms_mmu_destroy(dpu_kms);
>> -    if (dpu_kms->catalog) {
>> -        for (i = 0; i < dpu_kms->catalog->vbif_count; i++) {
>> -            u32 vbif_idx = dpu_kms->catalog->vbif[i].id;
>> -
>> -            if ((vbif_idx < VBIF_MAX) && dpu_kms->hw_vbif[vbif_idx])
>> -                dpu_hw_vbif_destroy(dpu_kms->hw_vbif[vbif_idx]);
>> -        }
>> -    }
>> -
>>       if (dpu_kms->rm_init)
>>           dpu_rm_destroy(&dpu_kms->rm);
>>       dpu_kms->rm_init = false;
>> @@ -1027,7 +1016,7 @@ static int dpu_kms_hw_init(struct msm_kms *kms)
>>   {
>>       struct dpu_kms *dpu_kms;
>>       struct drm_device *dev;
>> -    int i, rc = -EINVAL;
>> +    int rc = -EINVAL;
>>       if (!kms) {
>>           DPU_ERROR("invalid kms\n");
>> @@ -1116,21 +1105,6 @@ static int dpu_kms_hw_init(struct msm_kms *kms)
>>           goto power_error;
>>       }
>> -    for (i = 0; i < dpu_kms->catalog->vbif_count; i++) {
>> -        u32 vbif_idx = dpu_kms->catalog->vbif[i].id;
>> -
>> -        dpu_kms->hw_vbif[i] = dpu_hw_vbif_init(vbif_idx,
>> -                dpu_kms->vbif[vbif_idx], dpu_kms->catalog);
>> -        if (IS_ERR_OR_NULL(dpu_kms->hw_vbif[vbif_idx])) {
>> -            rc = PTR_ERR(dpu_kms->hw_vbif[vbif_idx]);
>> -            if (!dpu_kms->hw_vbif[vbif_idx])
>> -                rc = -EINVAL;
>> -            DPU_ERROR("failed to init vbif %d: %d\n", vbif_idx, rc);
>> -            dpu_kms->hw_vbif[vbif_idx] = NULL;
>> -            goto power_error;
>> -        }
>> -    }
>> -
>>       rc = dpu_core_perf_init(&dpu_kms->perf, dev, dpu_kms->catalog,
>>               _dpu_kms_get_clk(dpu_kms, "core"));
>>       if (rc) {
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h
>> index 3f518c809e33..b96c901483ae 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h
>> @@ -105,7 +105,6 @@ struct dpu_kms {
>>       struct dpu_rm rm;
>>       bool rm_init;
>> -    struct dpu_hw_vbif *hw_vbif[VBIF_MAX];
>>       struct dpu_hw_mdp *hw_mdp;
>>       bool has_danger_ctrl;
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
>> index 7497538adae1..6d49666c4e77 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
>> @@ -76,6 +76,8 @@ int dpu_rm_destroy(struct dpu_rm *rm)
>>       }
>>       for (i = 0; i < ARRAY_SIZE(rm->hw_intf); i++)
>>           dpu_hw_intf_destroy(rm->hw_intf[i]);
>> +    for (i = 0; i < ARRAY_SIZE(rm->hw_vbif); i++)
>> +        dpu_hw_vbif_destroy(rm->hw_vbif[i]);
>>       return 0;
>>   }
>> @@ -210,6 +212,23 @@ int dpu_rm_init(struct dpu_rm *rm,
>>           rm->dspp_blks[dspp->id - DSPP_0] = &hw->base;
>>       }
>> +    for (i = 0; i < cat->vbif_count; i++) {
>> +        struct dpu_hw_vbif *hw;
>> +        const struct dpu_vbif_cfg *vbif = &cat->vbif[i];
>> +
>> +        if (vbif->id < VBIF_0 || vbif->id >= VBIF_MAX) {
>> +            DPU_ERROR("skip vbif %d with invalid id\n", vbif->id);
>> +            continue;
>> +        }
>> +        hw = dpu_hw_vbif_init(vbif->id, mmio, cat);
>> +        if (IS_ERR(hw)) {
>> +            rc = PTR_ERR(hw);
>> +            DPU_ERROR("failed vbif object creation: err %d\n", rc);
>> +            goto fail;
>> +        }
>> +        rm->hw_vbif[vbif->id - VBIF_0] = hw;
>> +    }
>> +
>>       return 0;
>>   fail:
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
>> index 9b13200a050a..a15977bdceeb 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
>> @@ -20,6 +20,7 @@ struct dpu_global_state;
>>    * @ctl_blks: array of ctl hardware resources
>>    * @hw_intf: array of intf hardware resources
>>    * @dspp_blks: array of dspp hardware resources
>> + * @hw_vbif: array of vbif hardware resources
>>    */
>>   struct dpu_rm {
>>       struct dpu_hw_blk *pingpong_blks[PINGPONG_MAX - PINGPONG_0];
>> @@ -28,6 +29,7 @@ struct dpu_rm {
>>       struct dpu_hw_intf *hw_intf[INTF_MAX - INTF_0];
>>       struct dpu_hw_blk *dspp_blks[DSPP_MAX - DSPP_0];
>>       struct dpu_hw_blk *merge_3d_blks[MERGE_3D_MAX - MERGE_3D_0];
>> +    struct dpu_hw_vbif *hw_vbif[VBIF_MAX - VBIF_0];
>>   };
>>   /**
>> @@ -95,5 +97,15 @@ static inline struct dpu_hw_intf 
>> *dpu_rm_get_intf(struct dpu_rm *rm, enum dpu_in
>>       return rm->hw_intf[intf_idx - INTF_0];
>>   }
>> +/**
>> + * dpu_rm_get_vbif - Return a struct dpu_hw_vbif instance given it's 
>> index.
>> + * @rm: DPU Resource Manager handle
>> + * @vbif_idx: VBIF's index
>> + */
>> +static inline struct dpu_hw_vbif *dpu_rm_get_vbif(struct dpu_rm *rm, 
>> enum dpu_vbif vbif_idx)
>> +{
>> +    return rm->hw_vbif[vbif_idx - VBIF_0];
>> +}
>> +
>>   #endif /* __DPU_RM_H__ */
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c 
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
>> index 21d20373eb8b..00ac2aa81651 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
>> @@ -152,15 +152,10 @@ void dpu_vbif_set_ot_limit(struct dpu_kms *dpu_kms,
>>       struct dpu_hw_mdp *mdp;
>>       bool forced_on = false;
>>       u32 ot_lim;
>> -    int ret, i;
>> +    int ret;
>>       mdp = dpu_kms->hw_mdp;
>> -
>> -    for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
>> -        if (dpu_kms->hw_vbif[i] &&
>> -                dpu_kms->hw_vbif[i]->idx == params->vbif_idx)
>> -            vbif = dpu_kms->hw_vbif[i];
>> -    }
>> +    vbif = dpu_rm_get_vbif(&dpu_kms->rm, params->vbif_idx);
>>       if (!vbif || !mdp) {
>>           DRM_DEBUG_ATOMIC("invalid arguments vbif %d mdp %d\n",
>> @@ -216,14 +211,7 @@ void dpu_vbif_set_qos_remap(struct dpu_kms *dpu_kms,
>>       }
>>       mdp = dpu_kms->hw_mdp;
>> -    for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
>> -        if (dpu_kms->hw_vbif[i] &&
>> -                dpu_kms->hw_vbif[i]->idx == params->vbif_idx) {
>> -            vbif = dpu_kms->hw_vbif[i];
>> -            break;
>> -        }
>> -    }
>> -
>> +    vbif = dpu_rm_get_vbif(&dpu_kms->rm, params->vbif_idx);
>>       if (!vbif || !vbif->cap) {
>>           DPU_ERROR("invalid vbif %d\n", params->vbif_idx);
>>           return;
>> @@ -261,8 +249,8 @@ void dpu_vbif_clear_errors(struct dpu_kms *dpu_kms)
>>       struct dpu_hw_vbif *vbif;
>>       u32 i, pnd, src;
>> -    for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
>> -        vbif = dpu_kms->hw_vbif[i];
>> +    for (i = VBIF_0; i < VBIF_MAX; i++) {
>> +        vbif = dpu_rm_get_vbif(&dpu_kms->rm, i);
>>           if (vbif && vbif->ops.clear_errors) {
>>               vbif->ops.clear_errors(vbif, &pnd, &src);
>>               if (pnd || src) {
>> @@ -278,8 +266,8 @@ void dpu_vbif_init_memtypes(struct dpu_kms *dpu_kms)
>>       struct dpu_hw_vbif *vbif;
>>       int i, j;
>> -    for (i = 0; i < ARRAY_SIZE(dpu_kms->hw_vbif); i++) {
>> -        vbif = dpu_kms->hw_vbif[i];
>> +    for (i = VBIF_0; i < VBIF_MAX; i++) {
>> +        vbif = dpu_rm_get_vbif(&dpu_kms->rm, i);
>>           if (vbif && vbif->cap && vbif->ops.set_mem_type) {
>>               for (j = 0; j < vbif->cap->memtype_count; j++)
>>                   vbif->ops.set_mem_type(


-- 
With best wishes
Dmitry

  reply	other threads:[~2022-02-14 20:56 UTC|newest]

Thread overview: 46+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-01-21 21:06 [PATCH v5 0/6] drm/msm/dpu: simplify RM code Dmitry Baryshkov
2022-01-21 21:06 ` Dmitry Baryshkov
2022-01-21 21:06 ` [PATCH v5 1/6] drm/msm/dpu: drop unused lm_max_width from RM Dmitry Baryshkov
2022-01-21 21:06   ` Dmitry Baryshkov
2022-02-10  0:08   ` [Freedreno] " Abhinav Kumar
2022-02-10  0:08     ` Abhinav Kumar
2022-01-21 21:06 ` [PATCH v5 2/6] drm/msm/dpu: add DSPP blocks teardown Dmitry Baryshkov
2022-01-21 21:06   ` Dmitry Baryshkov
2022-02-10  0:09   ` Abhinav Kumar
2022-02-10  0:09     ` Abhinav Kumar
2022-01-21 21:06 ` [PATCH v5 3/6] drm/msm/dpu: get INTF blocks directly rather than through RM Dmitry Baryshkov
2022-01-21 21:06   ` Dmitry Baryshkov
2022-02-10  0:25   ` Abhinav Kumar
2022-02-10  0:25     ` Abhinav Kumar
2022-02-10  9:32     ` Dmitry Baryshkov
2022-02-10  9:32       ` Dmitry Baryshkov
2022-02-10 23:31       ` [Freedreno] " Abhinav Kumar
2022-02-10 23:31         ` Abhinav Kumar
2022-02-11 13:47         ` Dmitry Baryshkov
2022-02-11 13:47           ` Dmitry Baryshkov
2022-02-12  1:08           ` Abhinav Kumar
2022-02-12  1:08             ` Abhinav Kumar
2022-01-21 21:06 ` [PATCH v5 4/6] drm/msm/dpu: stop embedding dpu_hw_blk into dpu_hw_intf Dmitry Baryshkov
2022-01-21 21:06   ` Dmitry Baryshkov
2022-02-14 19:08   ` Abhinav Kumar
2022-02-14 19:08     ` Abhinav Kumar
2022-01-21 21:06 ` [PATCH v5 5/6] drm/msm/dpu: fix error handling in dpu_rm_init Dmitry Baryshkov
2022-01-21 21:06   ` Dmitry Baryshkov
2022-02-14 19:15   ` Abhinav Kumar
2022-02-14 19:15     ` Abhinav Kumar
2022-02-14 20:43     ` Dmitry Baryshkov
2022-02-14 20:43       ` Dmitry Baryshkov
2022-02-14 21:25       ` Abhinav Kumar
2022-02-14 21:25         ` Abhinav Kumar
2022-01-21 21:06 ` [PATCH v5 6/6] drm/msm/dpu: move VBIF blocks handling to dpu_rm Dmitry Baryshkov
2022-01-21 21:06   ` Dmitry Baryshkov
2022-02-14 19:53   ` Abhinav Kumar
2022-02-14 19:53     ` Abhinav Kumar
2022-02-14 20:56     ` Dmitry Baryshkov [this message]
2022-02-14 20:56       ` Dmitry Baryshkov
2022-02-14 22:04       ` [Freedreno] " Abhinav Kumar
2022-02-14 22:04         ` Abhinav Kumar
2022-02-14 22:39         ` Dmitry Baryshkov
2022-02-14 22:39           ` Dmitry Baryshkov
2022-02-14 23:22           ` Abhinav Kumar
2022-02-14 23:22             ` Abhinav Kumar

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=a2c001ee-c257-5d25-e0a9-dda7d1dcdead@linaro.org \
    --to=dmitry.baryshkov@linaro.org \
    --cc=airlied@linux.ie \
    --cc=bjorn.andersson@linaro.org \
    --cc=dri-devel@lists.freedesktop.org \
    --cc=freedreno@lists.freedesktop.org \
    --cc=linux-arm-msm@vger.kernel.org \
    --cc=quic_abhinavk@quicinc.com \
    --cc=robdclark@gmail.com \
    --cc=sean@poorly.run \
    --cc=swboyd@chromium.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.