On Wed, 10 Aug 2022, Jani Nikula <jani.nik...@linux.intel.com> wrote:
> On Tue, 02 Aug 2022, Matt Roper <matthew.d.ro...@intel.com> wrote:
>> On Wed, Jul 27, 2022 at 06:34:10PM -0700, Radhakrishna Sripada wrote:
>>> Since Xe LPD+, Memory latency data are in LATENCY_LPX_LPY registers
>>> instead of GT driver mailbox.
>>> 
>>> Bspec: 64608
>>> 
>>> Cc: Matt Roper <matthew.d.ro...@intel.com>
>>> Original Author: Caz Yokoyama
>>> Signed-off-by: Radhakrishna Sripada <radhakrishna.srip...@intel.com>
>>> ---
>>>  drivers/gpu/drm/i915/i915_reg.h |   7 +++
>>>  drivers/gpu/drm/i915/intel_pm.c | 105 +++++++++++++++++++-------------
>>>  2 files changed, 71 insertions(+), 41 deletions(-)
>>> 
>>> diff --git a/drivers/gpu/drm/i915/i915_reg.h 
>>> b/drivers/gpu/drm/i915/i915_reg.h
>>> index 6087d40eed70..23b50d671550 100644
>>> --- a/drivers/gpu/drm/i915/i915_reg.h
>>> +++ b/drivers/gpu/drm/i915/i915_reg.h
>>> @@ -8754,4 +8754,11 @@ enum skl_power_gate {
>>>  #define GEN12_STATE_ACK_DEBUG              _MMIO(0x20BC)
>>>  
>>>  #define MTL_MEDIA_GSI_BASE         0x380000
>>> +
>>> +#define MTL_LATENCY_LP0_LP1                _MMIO(0x45780)
>>> +#define MTL_LATENCY_LP2_LP3                _MMIO(0x45784)
>>> +#define MTL_LATENCY_LP4_LP5                _MMIO(0x45788)
>>> +#define  MTL_LATENCY_LEVEL0_2_4_MASK       REG_GENMASK(12, 0)
>>> +#define  MTL_LATENCY_LEVEL1_3_5_MASK       REG_GENMASK(28, 16)
>>> +
>>>  #endif /* _I915_REG_H_ */
>>> diff --git a/drivers/gpu/drm/i915/intel_pm.c 
>>> b/drivers/gpu/drm/i915/intel_pm.c
>>> index ef7553b494ea..fac565d23d57 100644
>>> --- a/drivers/gpu/drm/i915/intel_pm.c
>>> +++ b/drivers/gpu/drm/i915/intel_pm.c
>>> @@ -2861,16 +2861,75 @@ static void ilk_compute_wm_level(const struct 
>>> drm_i915_private *dev_priv,
>>>     result->enable = true;
>>>  }
>>>  
>>> +static void
>>> +adjust_wm_latency(u16 wm[], int max_level, int read_latency,
>>> +             bool wm_lv_0_adjust_needed)
>>
>> The refactoring to separate the adjustment from the readout should
>> probably be a separate patch before you add the MTL-specific changes on
>> top.
>
> Agreed.

And to elaborate, this kind of stuff should happen upstream months
before anyone's even heard of the new platform!

BR,
Jani.

>
>>
>>
>> Matt
>>
>>> +{
>>> +   int i, level;
>>> +
>>> +   /*
>>> +    * If a level n (n > 1) has a 0us latency, all levels m (m >= n)
>>> +    * need to be disabled. We make sure to sanitize the values out
>>> +    * of the punit to satisfy this requirement.
>>> +    */
>>> +   for (level = 1; level <= max_level; level++) {
>>> +           if (wm[level] == 0) {
>>> +                   for (i = level + 1; i <= max_level; i++)
>>> +                           wm[i] = 0;
>>> +
>>> +                   max_level = level - 1;
>>> +                   break;
>>> +           }
>>> +   }
>>> +
>>> +   /*
>>> +    * WaWmMemoryReadLatency
>>> +    *
>>> +    * punit doesn't take into account the read latency so we need
>>> +    * to add proper adjustement to each valid level we retrieve
>>> +    * from the punit when level 0 response data is 0us.
>>> +    */
>>> +   if (wm[0] == 0) {
>>> +           for (level = 0; level <= max_level; level++)
>>> +                   wm[level] += read_latency;
>>> +   }
>>> +
>>> +   /*
>>> +    * WA Level-0 adjustment for 16GB DIMMs: SKL+
>>> +    * If we could not get dimm info enable this WA to prevent from
>>> +    * any underrun. If not able to get Dimm info assume 16GB dimm
>>> +    * to avoid any underrun.
>>> +    */
>>> +   if (wm_lv_0_adjust_needed)
>>> +           wm[0] += 1;
>>> +}
>>> +
>>>  static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
>>>                               u16 wm[])
>>>  {
>>>     struct intel_uncore *uncore = &dev_priv->uncore;
>>> +   int max_level = ilk_wm_max_level(dev_priv);
>>>  
>>> -   if (DISPLAY_VER(dev_priv) >= 9) {
>>> +   if (DISPLAY_VER(dev_priv) >= 14) {
>>>             u32 val;
>>> -           int ret, i;
>>> -           int level, max_level = ilk_wm_max_level(dev_priv);
>>> +
>>> +           val = intel_uncore_read(uncore, MTL_LATENCY_LP0_LP1);
>>> +           wm[0] = REG_FIELD_GET(MTL_LATENCY_LEVEL0_2_4_MASK, val);
>>> +           wm[1] = REG_FIELD_GET(MTL_LATENCY_LEVEL1_3_5_MASK, val);
>>> +           val = intel_uncore_read(uncore, MTL_LATENCY_LP2_LP3);
>>> +           wm[2] = REG_FIELD_GET(MTL_LATENCY_LEVEL0_2_4_MASK, val);
>>> +           wm[3] = REG_FIELD_GET(MTL_LATENCY_LEVEL1_3_5_MASK, val);
>>> +           val = intel_uncore_read(uncore, MTL_LATENCY_LP4_LP5);
>>> +           wm[4] = REG_FIELD_GET(MTL_LATENCY_LEVEL0_2_4_MASK, val);
>>> +           wm[5] = REG_FIELD_GET(MTL_LATENCY_LEVEL1_3_5_MASK, val);
>>> +
>>> +           adjust_wm_latency(wm, max_level, 6,
>>> +                             dev_priv->dram_info.wm_lv_0_adjust_needed);
>>> +   } else if (DISPLAY_VER(dev_priv) >= 9) {
>>> +           int read_latency = DISPLAY_VER(dev_priv) >= 12 ? 3 : 2;
>>>             int mult = IS_DG2(dev_priv) ? 2 : 1;
>>> +           u32 val;
>>> +           int ret;
>>>  
>>>             /* read the first set of memory latencies[0:3] */
>>>             val = 0; /* data0 to be programmed to 0 for first set */
>>> @@ -2909,44 +2968,8 @@ static void intel_read_wm_latency(struct 
>>> drm_i915_private *dev_priv,
>>>             wm[7] = ((val >> GEN9_MEM_LATENCY_LEVEL_3_7_SHIFT) &
>>>                             GEN9_MEM_LATENCY_LEVEL_MASK) * mult;
>>>  
>>> -           /*
>>> -            * If a level n (n > 1) has a 0us latency, all levels m (m >= n)
>>> -            * need to be disabled. We make sure to sanitize the values out
>>> -            * of the punit to satisfy this requirement.
>>> -            */
>>> -           for (level = 1; level <= max_level; level++) {
>>> -                   if (wm[level] == 0) {
>>> -                           for (i = level + 1; i <= max_level; i++)
>>> -                                   wm[i] = 0;
>>> -
>>> -                           max_level = level - 1;
>>> -
>>> -                           break;
>>> -                   }
>>> -           }
>>> -
>>> -           /*
>>> -            * WaWmMemoryReadLatency
>>> -            *
>>> -            * punit doesn't take into account the read latency so we need
>>> -            * to add proper adjustement to each valid level we retrieve
>>> -            * from the punit when level 0 response data is 0us.
>>> -            */
>>> -           if (wm[0] == 0) {
>>> -                   u8 adjust = DISPLAY_VER(dev_priv) >= 12 ? 3 : 2;
>>> -
>>> -                   for (level = 0; level <= max_level; level++)
>>> -                           wm[level] += adjust;
>>> -           }
>>> -
>>> -           /*
>>> -            * WA Level-0 adjustment for 16GB DIMMs: SKL+
>>> -            * If we could not get dimm info enable this WA to prevent from
>>> -            * any underrun. If not able to get Dimm info assume 16GB dimm
>>> -            * to avoid any underrun.
>>> -            */
>>> -           if (dev_priv->dram_info.wm_lv_0_adjust_needed)
>>> -                   wm[0] += 1;
>>> +           adjust_wm_latency(wm, max_level, read_latency,
>>> +                             dev_priv->dram_info.wm_lv_0_adjust_needed);
>>>     } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
>>>             u64 sskpd = intel_uncore_read64(uncore, MCH_SSKPD);
>>>  
>>> -- 
>>> 2.25.1
>>> 

-- 
Jani Nikula, Intel Open Source Graphics Center

Reply via email to