Commit 6de12da1 authored by John Harrison's avatar John Harrison

drm/i915/guc: Provide mmio list to be saved/restored on engine reset

The driver must provide GuC with a list of mmio registers
that should be saved/restored during a GuC-based engine reset.
Unfortunately, the list must be dynamically allocated as its size is
variable. That means the driver must generate the list twice - once to
work out the size and a second time to actually save it.

v2:
 (Alan / CI)
  - GEN7_GT_MODE -> GEN6_GT_MODE to fix WA selftest failure
Signed-off-by: default avatarJohn Harrison <John.C.Harrison@Intel.com>
Signed-off-by: default avatarFernando Pacheco <fernando.pacheco@intel.com>
Signed-off-by: default avatarMatthew Brost <matthew.brost@intel.com>
Cc: Daniele Ceraolo Spurio <daniele.ceraolospurio@intel.com>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Reviewed-by: default avatarMatthew Brost <matthew.brost@intel.com>
Signed-off-by: default avatarJohn Harrison <John.C.Harrison@Intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20210727002348.97202-16-matthew.brost@intel.com
parent 933864af
...@@ -150,13 +150,14 @@ static void _wa_add(struct i915_wa_list *wal, const struct i915_wa *wa) ...@@ -150,13 +150,14 @@ static void _wa_add(struct i915_wa_list *wal, const struct i915_wa *wa)
} }
static void wa_add(struct i915_wa_list *wal, i915_reg_t reg, static void wa_add(struct i915_wa_list *wal, i915_reg_t reg,
u32 clear, u32 set, u32 read_mask) u32 clear, u32 set, u32 read_mask, bool masked_reg)
{ {
struct i915_wa wa = { struct i915_wa wa = {
.reg = reg, .reg = reg,
.clr = clear, .clr = clear,
.set = set, .set = set,
.read = read_mask, .read = read_mask,
.masked_reg = masked_reg,
}; };
_wa_add(wal, &wa); _wa_add(wal, &wa);
...@@ -165,7 +166,7 @@ static void wa_add(struct i915_wa_list *wal, i915_reg_t reg, ...@@ -165,7 +166,7 @@ static void wa_add(struct i915_wa_list *wal, i915_reg_t reg,
static void static void
wa_write_clr_set(struct i915_wa_list *wal, i915_reg_t reg, u32 clear, u32 set) wa_write_clr_set(struct i915_wa_list *wal, i915_reg_t reg, u32 clear, u32 set)
{ {
wa_add(wal, reg, clear, set, clear); wa_add(wal, reg, clear, set, clear, false);
} }
static void static void
...@@ -200,20 +201,20 @@ wa_write_clr(struct i915_wa_list *wal, i915_reg_t reg, u32 clr) ...@@ -200,20 +201,20 @@ wa_write_clr(struct i915_wa_list *wal, i915_reg_t reg, u32 clr)
static void static void
wa_masked_en(struct i915_wa_list *wal, i915_reg_t reg, u32 val) wa_masked_en(struct i915_wa_list *wal, i915_reg_t reg, u32 val)
{ {
wa_add(wal, reg, 0, _MASKED_BIT_ENABLE(val), val); wa_add(wal, reg, 0, _MASKED_BIT_ENABLE(val), val, true);
} }
static void static void
wa_masked_dis(struct i915_wa_list *wal, i915_reg_t reg, u32 val) wa_masked_dis(struct i915_wa_list *wal, i915_reg_t reg, u32 val)
{ {
wa_add(wal, reg, 0, _MASKED_BIT_DISABLE(val), val); wa_add(wal, reg, 0, _MASKED_BIT_DISABLE(val), val, true);
} }
static void static void
wa_masked_field_set(struct i915_wa_list *wal, i915_reg_t reg, wa_masked_field_set(struct i915_wa_list *wal, i915_reg_t reg,
u32 mask, u32 val) u32 mask, u32 val)
{ {
wa_add(wal, reg, 0, _MASKED_FIELD(mask, val), mask); wa_add(wal, reg, 0, _MASKED_FIELD(mask, val), mask, true);
} }
static void gen6_ctx_workarounds_init(struct intel_engine_cs *engine, static void gen6_ctx_workarounds_init(struct intel_engine_cs *engine,
...@@ -533,10 +534,10 @@ static void icl_ctx_workarounds_init(struct intel_engine_cs *engine, ...@@ -533,10 +534,10 @@ static void icl_ctx_workarounds_init(struct intel_engine_cs *engine,
wa_masked_en(wal, ICL_HDC_MODE, HDC_FORCE_NON_COHERENT); wa_masked_en(wal, ICL_HDC_MODE, HDC_FORCE_NON_COHERENT);
/* WaEnableFloatBlendOptimization:icl */ /* WaEnableFloatBlendOptimization:icl */
wa_write_clr_set(wal, wa_add(wal, GEN10_CACHE_MODE_SS, 0,
GEN10_CACHE_MODE_SS, _MASKED_BIT_ENABLE(FLOAT_BLEND_OPTIMIZATION_ENABLE),
0, /* write-only, so skip validation */ 0 /* write-only, so skip validation */,
_MASKED_BIT_ENABLE(FLOAT_BLEND_OPTIMIZATION_ENABLE)); true);
/* WaDisableGPGPUMidThreadPreemption:icl */ /* WaDisableGPGPUMidThreadPreemption:icl */
wa_masked_field_set(wal, GEN8_CS_CHICKEN1, wa_masked_field_set(wal, GEN8_CS_CHICKEN1,
...@@ -581,7 +582,7 @@ static void gen12_ctx_gt_tuning_init(struct intel_engine_cs *engine, ...@@ -581,7 +582,7 @@ static void gen12_ctx_gt_tuning_init(struct intel_engine_cs *engine,
FF_MODE2, FF_MODE2,
FF_MODE2_TDS_TIMER_MASK, FF_MODE2_TDS_TIMER_MASK,
FF_MODE2_TDS_TIMER_128, FF_MODE2_TDS_TIMER_128,
0); 0, false);
} }
static void gen12_ctx_workarounds_init(struct intel_engine_cs *engine, static void gen12_ctx_workarounds_init(struct intel_engine_cs *engine,
...@@ -619,7 +620,7 @@ static void gen12_ctx_workarounds_init(struct intel_engine_cs *engine, ...@@ -619,7 +620,7 @@ static void gen12_ctx_workarounds_init(struct intel_engine_cs *engine,
FF_MODE2, FF_MODE2,
FF_MODE2_GS_TIMER_MASK, FF_MODE2_GS_TIMER_MASK,
FF_MODE2_GS_TIMER_224, FF_MODE2_GS_TIMER_224,
0); 0, false);
/* /*
* Wa_14012131227:dg1 * Wa_14012131227:dg1
...@@ -795,7 +796,7 @@ hsw_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal) ...@@ -795,7 +796,7 @@ hsw_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
wa_add(wal, wa_add(wal,
HSW_ROW_CHICKEN3, 0, HSW_ROW_CHICKEN3, 0,
_MASKED_BIT_ENABLE(HSW_ROW_CHICKEN3_L3_GLOBAL_ATOMICS_DISABLE), _MASKED_BIT_ENABLE(HSW_ROW_CHICKEN3_L3_GLOBAL_ATOMICS_DISABLE),
0 /* XXX does this reg exist? */); 0 /* XXX does this reg exist? */, true);
/* WaVSRefCountFullforceMissDisable:hsw */ /* WaVSRefCountFullforceMissDisable:hsw */
wa_write_clr(wal, GEN7_FF_THREAD_MODE, GEN7_FF_VS_REF_CNT_FFME); wa_write_clr(wal, GEN7_FF_THREAD_MODE, GEN7_FF_VS_REF_CNT_FFME);
...@@ -1824,10 +1825,10 @@ rcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal) ...@@ -1824,10 +1825,10 @@ rcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal)
* disable bit, which we don't touch here, but it's good * disable bit, which we don't touch here, but it's good
* to keep in mind (see 3DSTATE_PS and 3DSTATE_WM). * to keep in mind (see 3DSTATE_PS and 3DSTATE_WM).
*/ */
wa_add(wal, GEN7_GT_MODE, 0, wa_masked_field_set(wal,
_MASKED_FIELD(GEN6_WIZ_HASHING_MASK, GEN7_GT_MODE,
GEN6_WIZ_HASHING_16x4), GEN6_WIZ_HASHING_MASK,
GEN6_WIZ_HASHING_16x4); GEN6_WIZ_HASHING_16x4);
} }
if (IS_GRAPHICS_VER(i915, 6, 7)) if (IS_GRAPHICS_VER(i915, 6, 7))
...@@ -1877,10 +1878,10 @@ rcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal) ...@@ -1877,10 +1878,10 @@ rcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal)
* disable bit, which we don't touch here, but it's good * disable bit, which we don't touch here, but it's good
* to keep in mind (see 3DSTATE_PS and 3DSTATE_WM). * to keep in mind (see 3DSTATE_PS and 3DSTATE_WM).
*/ */
wa_add(wal, wa_masked_field_set(wal,
GEN6_GT_MODE, 0, GEN6_GT_MODE,
_MASKED_FIELD(GEN6_WIZ_HASHING_MASK, GEN6_WIZ_HASHING_16x4), GEN6_WIZ_HASHING_MASK,
GEN6_WIZ_HASHING_16x4); GEN6_WIZ_HASHING_16x4);
/* WaDisable_RenderCache_OperationalFlush:snb */ /* WaDisable_RenderCache_OperationalFlush:snb */
wa_masked_dis(wal, CACHE_MODE_0, RC_OP_FLUSH_ENABLE); wa_masked_dis(wal, CACHE_MODE_0, RC_OP_FLUSH_ENABLE);
...@@ -1901,7 +1902,7 @@ rcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal) ...@@ -1901,7 +1902,7 @@ rcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal)
wa_add(wal, MI_MODE, wa_add(wal, MI_MODE,
0, _MASKED_BIT_ENABLE(VS_TIMER_DISPATCH), 0, _MASKED_BIT_ENABLE(VS_TIMER_DISPATCH),
/* XXX bit doesn't stick on Broadwater */ /* XXX bit doesn't stick on Broadwater */
IS_I965G(i915) ? 0 : VS_TIMER_DISPATCH); IS_I965G(i915) ? 0 : VS_TIMER_DISPATCH, true);
if (GRAPHICS_VER(i915) == 4) if (GRAPHICS_VER(i915) == 4)
/* /*
...@@ -1916,7 +1917,8 @@ rcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal) ...@@ -1916,7 +1917,8 @@ rcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal)
*/ */
wa_add(wal, ECOSKPD, wa_add(wal, ECOSKPD,
0, _MASKED_BIT_ENABLE(ECO_CONSTANT_BUFFER_SR_DISABLE), 0, _MASKED_BIT_ENABLE(ECO_CONSTANT_BUFFER_SR_DISABLE),
0 /* XXX bit doesn't stick on Broadwater */); 0 /* XXX bit doesn't stick on Broadwater */,
true);
} }
static void static void
......
...@@ -15,6 +15,7 @@ struct i915_wa { ...@@ -15,6 +15,7 @@ struct i915_wa {
u32 clr; u32 clr;
u32 set; u32 set;
u32 read; u32 read;
bool masked_reg;
}; };
struct i915_wa_list { struct i915_wa_list {
......
...@@ -59,6 +59,7 @@ struct intel_guc { ...@@ -59,6 +59,7 @@ struct intel_guc {
struct i915_vma *ads_vma; struct i915_vma *ads_vma;
struct __guc_ads_blob *ads_blob; struct __guc_ads_blob *ads_blob;
u32 ads_regset_size;
struct i915_vma *lrc_desc_pool; struct i915_vma *lrc_desc_pool;
void *lrc_desc_pool_vaddr; void *lrc_desc_pool_vaddr;
......
...@@ -3,6 +3,8 @@ ...@@ -3,6 +3,8 @@
* Copyright © 2014-2019 Intel Corporation * Copyright © 2014-2019 Intel Corporation
*/ */
#include <linux/bsearch.h>
#include "gt/intel_gt.h" #include "gt/intel_gt.h"
#include "gt/intel_lrc.h" #include "gt/intel_lrc.h"
#include "intel_guc_ads.h" #include "intel_guc_ads.h"
...@@ -23,7 +25,12 @@ ...@@ -23,7 +25,12 @@
* | guc_policies | * | guc_policies |
* +---------------------------------------+ * +---------------------------------------+
* | guc_gt_system_info | * | guc_gt_system_info |
* +---------------------------------------+ * +---------------------------------------+ <== static
* | guc_mmio_reg[countA] (engine 0.0) |
* | guc_mmio_reg[countB] (engine 0.1) |
* | guc_mmio_reg[countC] (engine 1.0) |
* | ... |
* +---------------------------------------+ <== dynamic
* | padding | * | padding |
* +---------------------------------------+ <== 4K aligned * +---------------------------------------+ <== 4K aligned
* | private data | * | private data |
...@@ -35,16 +42,33 @@ struct __guc_ads_blob { ...@@ -35,16 +42,33 @@ struct __guc_ads_blob {
struct guc_ads ads; struct guc_ads ads;
struct guc_policies policies; struct guc_policies policies;
struct guc_gt_system_info system_info; struct guc_gt_system_info system_info;
/* From here on, location is dynamic! Refer to above diagram. */
struct guc_mmio_reg regset[0];
} __packed; } __packed;
static u32 guc_ads_regset_size(struct intel_guc *guc)
{
GEM_BUG_ON(!guc->ads_regset_size);
return guc->ads_regset_size;
}
static u32 guc_ads_private_data_size(struct intel_guc *guc) static u32 guc_ads_private_data_size(struct intel_guc *guc)
{ {
return PAGE_ALIGN(guc->fw.private_data_size); return PAGE_ALIGN(guc->fw.private_data_size);
} }
static u32 guc_ads_regset_offset(struct intel_guc *guc)
{
return offsetof(struct __guc_ads_blob, regset);
}
static u32 guc_ads_private_data_offset(struct intel_guc *guc) static u32 guc_ads_private_data_offset(struct intel_guc *guc)
{ {
return PAGE_ALIGN(sizeof(struct __guc_ads_blob)); u32 offset;
offset = guc_ads_regset_offset(guc) +
guc_ads_regset_size(guc);
return PAGE_ALIGN(offset);
} }
static u32 guc_ads_blob_size(struct intel_guc *guc) static u32 guc_ads_blob_size(struct intel_guc *guc)
...@@ -83,6 +107,165 @@ static void guc_mapping_table_init(struct intel_gt *gt, ...@@ -83,6 +107,165 @@ static void guc_mapping_table_init(struct intel_gt *gt,
} }
} }
/*
* The save/restore register list must be pre-calculated to a temporary
* buffer of driver defined size before it can be generated in place
* inside the ADS.
*/
#define MAX_MMIO_REGS 128 /* Arbitrary size, increase as needed */
struct temp_regset {
struct guc_mmio_reg *registers;
u32 used;
u32 size;
};
static int guc_mmio_reg_cmp(const void *a, const void *b)
{
const struct guc_mmio_reg *ra = a;
const struct guc_mmio_reg *rb = b;
return (int)ra->offset - (int)rb->offset;
}
static void guc_mmio_reg_add(struct temp_regset *regset,
u32 offset, u32 flags)
{
u32 count = regset->used;
struct guc_mmio_reg reg = {
.offset = offset,
.flags = flags,
};
struct guc_mmio_reg *slot;
GEM_BUG_ON(count >= regset->size);
/*
* The mmio list is built using separate lists within the driver.
* It's possible that at some point we may attempt to add the same
* register more than once. Do not consider this an error; silently
* move on if the register is already in the list.
*/
if (bsearch(&reg, regset->registers, count,
sizeof(reg), guc_mmio_reg_cmp))
return;
slot = &regset->registers[count];
regset->used++;
*slot = reg;
while (slot-- > regset->registers) {
GEM_BUG_ON(slot[0].offset == slot[1].offset);
if (slot[1].offset > slot[0].offset)
break;
swap(slot[1], slot[0]);
}
}
#define GUC_MMIO_REG_ADD(regset, reg, masked) \
guc_mmio_reg_add(regset, \
i915_mmio_reg_offset((reg)), \
(masked) ? GUC_REGSET_MASKED : 0)
static void guc_mmio_regset_init(struct temp_regset *regset,
struct intel_engine_cs *engine)
{
const u32 base = engine->mmio_base;
struct i915_wa_list *wal = &engine->wa_list;
struct i915_wa *wa;
unsigned int i;
regset->used = 0;
GUC_MMIO_REG_ADD(regset, RING_MODE_GEN7(base), true);
GUC_MMIO_REG_ADD(regset, RING_HWS_PGA(base), false);
GUC_MMIO_REG_ADD(regset, RING_IMR(base), false);
for (i = 0, wa = wal->list; i < wal->count; i++, wa++)
GUC_MMIO_REG_ADD(regset, wa->reg, wa->masked_reg);
/* Be extra paranoid and include all whitelist registers. */
for (i = 0; i < RING_MAX_NONPRIV_SLOTS; i++)
GUC_MMIO_REG_ADD(regset,
RING_FORCE_TO_NONPRIV(base, i),
false);
/* add in local MOCS registers */
for (i = 0; i < GEN9_LNCFCMOCS_REG_COUNT; i++)
GUC_MMIO_REG_ADD(regset, GEN9_LNCFCMOCS(i), false);
}
static int guc_mmio_reg_state_query(struct intel_guc *guc)
{
struct intel_gt *gt = guc_to_gt(guc);
struct intel_engine_cs *engine;
enum intel_engine_id id;
struct temp_regset temp_set;
u32 total;
/*
* Need to actually build the list in order to filter out
* duplicates and other such data dependent constructions.
*/
temp_set.size = MAX_MMIO_REGS;
temp_set.registers = kmalloc_array(temp_set.size,
sizeof(*temp_set.registers),
GFP_KERNEL);
if (!temp_set.registers)
return -ENOMEM;
total = 0;
for_each_engine(engine, gt, id) {
guc_mmio_regset_init(&temp_set, engine);
total += temp_set.used;
}
kfree(temp_set.registers);
return total * sizeof(struct guc_mmio_reg);
}
static void guc_mmio_reg_state_init(struct intel_guc *guc,
struct __guc_ads_blob *blob)
{
struct intel_gt *gt = guc_to_gt(guc);
struct intel_engine_cs *engine;
enum intel_engine_id id;
struct temp_regset temp_set;
struct guc_mmio_reg_set *ads_reg_set;
u32 addr_ggtt, offset;
u8 guc_class;
offset = guc_ads_regset_offset(guc);
addr_ggtt = intel_guc_ggtt_offset(guc, guc->ads_vma) + offset;
temp_set.registers = (struct guc_mmio_reg *)(((u8 *)blob) + offset);
temp_set.size = guc->ads_regset_size / sizeof(temp_set.registers[0]);
for_each_engine(engine, gt, id) {
/* Class index is checked in class converter */
GEM_BUG_ON(engine->instance >= GUC_MAX_INSTANCES_PER_CLASS);
guc_class = engine_class_to_guc_class(engine->class);
ads_reg_set = &blob->ads.reg_state_list[guc_class][engine->instance];
guc_mmio_regset_init(&temp_set, engine);
if (!temp_set.used) {
ads_reg_set->address = 0;
ads_reg_set->count = 0;
continue;
}
ads_reg_set->address = addr_ggtt;
ads_reg_set->count = temp_set.used;
temp_set.size -= temp_set.used;
temp_set.registers += temp_set.used;
addr_ggtt += temp_set.used * sizeof(struct guc_mmio_reg);
}
GEM_BUG_ON(temp_set.size);
}
/* /*
* The first 80 dwords of the register state context, containing the * The first 80 dwords of the register state context, containing the
* execlists and ppgtt registers. * execlists and ppgtt registers.
...@@ -121,8 +304,7 @@ static void __guc_ads_init(struct intel_guc *guc) ...@@ -121,8 +304,7 @@ static void __guc_ads_init(struct intel_guc *guc)
*/ */
blob->ads.golden_context_lrca[guc_class] = 0; blob->ads.golden_context_lrca[guc_class] = 0;
blob->ads.eng_state_size[guc_class] = blob->ads.eng_state_size[guc_class] =
intel_engine_context_size(guc_to_gt(guc), intel_engine_context_size(gt, engine_class) -
engine_class) -
skipped_size; skipped_size;
} }
...@@ -153,6 +335,9 @@ static void __guc_ads_init(struct intel_guc *guc) ...@@ -153,6 +335,9 @@ static void __guc_ads_init(struct intel_guc *guc)
blob->ads.scheduler_policies = base + ptr_offset(blob, policies); blob->ads.scheduler_policies = base + ptr_offset(blob, policies);
blob->ads.gt_system_info = base + ptr_offset(blob, system_info); blob->ads.gt_system_info = base + ptr_offset(blob, system_info);
/* MMIO save/restore list */
guc_mmio_reg_state_init(guc, blob);
/* Private Data */ /* Private Data */
blob->ads.private_data = base + guc_ads_private_data_offset(guc); blob->ads.private_data = base + guc_ads_private_data_offset(guc);
...@@ -173,6 +358,12 @@ int intel_guc_ads_create(struct intel_guc *guc) ...@@ -173,6 +358,12 @@ int intel_guc_ads_create(struct intel_guc *guc)
GEM_BUG_ON(guc->ads_vma); GEM_BUG_ON(guc->ads_vma);
/* Need to calculate the reg state size dynamically: */
ret = guc_mmio_reg_state_query(guc);
if (ret < 0)
return ret;
guc->ads_regset_size = ret;
size = guc_ads_blob_size(guc); size = guc_ads_blob_size(guc);
ret = intel_guc_allocate_and_map_vma(guc, size, &guc->ads_vma, ret = intel_guc_allocate_and_map_vma(guc, size, &guc->ads_vma,
......
...@@ -12316,6 +12316,7 @@ enum skl_power_gate { ...@@ -12316,6 +12316,7 @@ enum skl_power_gate {
/* MOCS (Memory Object Control State) registers */ /* MOCS (Memory Object Control State) registers */
#define GEN9_LNCFCMOCS(i) _MMIO(0xb020 + (i) * 4) /* L3 Cache Control */ #define GEN9_LNCFCMOCS(i) _MMIO(0xb020 + (i) * 4) /* L3 Cache Control */
#define GEN9_LNCFCMOCS_REG_COUNT 32
#define __GEN9_RCS0_MOCS0 0xc800 #define __GEN9_RCS0_MOCS0 0xc800
#define GEN9_GFX_MOCS(i) _MMIO(__GEN9_RCS0_MOCS0 + (i) * 4) #define GEN9_GFX_MOCS(i) _MMIO(__GEN9_RCS0_MOCS0 + (i) * 4)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment