Commit 10782166 authored by Linus Torvalds's avatar Linus Torvalds

Merge tag 'drm-next-2020-06-08' of git://anongit.freedesktop.org/drm/drm

Pull drm fixes from Dave Airlie:
 "These are the fixes from last week for the stuff merged in the merge
  window. It got a bunch of nouveau fixes for HDA audio on some new
  GPUs, some i915 and some amdpgu fixes.

  i915:
   - gvt: Fix one clang warning on debug only function
   - Use ARRAY_SIZE for coccicheck warning
   - Use after free fix for display global state.
   - Whitelisting context-local timestamp on Gen9 and two scheduler
     fixes with deps (Cc: stable)
   - Removal of write flag from sysfs files where ineffective

  nouveau:
   - HDMI/DP audio HDA fixes
   - display hang fix for Volta/Turing
   - GK20A regression fix.

  amdgpu:
   - Prevent hwmon accesses while GPU is in reset
   - CTF interrupt fix
   - Backlight fix for renoir
   - Fix for display sync groups
   - Display bandwidth validation workaround"

* tag 'drm-next-2020-06-08' of git://anongit.freedesktop.org/drm/drm: (28 commits)
  drm/nouveau/kms/nv50-: clear SW state of disabled windows harder
  drm/nouveau: gr/gk20a: Use firmware version 0
  drm/nouveau/disp/gm200-: detect and potentially disable HDA support on some SORs
  drm/nouveau/disp/gp100: split SOR implementation from gm200
  drm/nouveau/disp: modify OR allocation policy to account for HDA requirements
  drm/nouveau/disp: split part of OR allocation logic into a function
  drm/nouveau/disp: provide hint to OR allocation about HDA requirements
  drm/amd/display: Revalidate bandwidth before commiting DC updates
  drm/amdgpu/display: use blanked rather than plane state for sync groups
  drm/i915/params: fix i915.fake_lmem_start module param sysfs permissions
  drm/i915/params: don't expose inject_probe_failure in debugfs
  drm/i915: Whitelist context-local timestamp in the gen9 cmdparser
  drm/i915: Fix global state use-after-frees with a refcount
  drm/i915: Check for awaits on still currently executing requests
  drm/i915/gt: Do not schedule normal requests immediately along virtual
  drm/i915: Reorder await_execution before await_request
  drm/nouveau/kms/gt215-: fix race with audio driver runpm
  drm/nouveau/disp/gm200-: fix NV_PDISP_SOR_HDMI2_CTRL(n) selection
  Revert "drm/amd/display: disable dcn20 abm feature for bring up"
  drm/amd/powerplay: ack the SMUToHost interrupt on receive V2
  ...
parents 20b0d067 8d286e2f
This diff is collapsed.
...@@ -1356,7 +1356,7 @@ static int dm_late_init(void *handle) ...@@ -1356,7 +1356,7 @@ static int dm_late_init(void *handle)
unsigned int linear_lut[16]; unsigned int linear_lut[16];
int i; int i;
struct dmcu *dmcu = NULL; struct dmcu *dmcu = NULL;
bool ret = false; bool ret;
if (!adev->dm.fw_dmcu) if (!adev->dm.fw_dmcu)
return detect_mst_link_for_all_connectors(adev->ddev); return detect_mst_link_for_all_connectors(adev->ddev);
...@@ -1377,13 +1377,10 @@ static int dm_late_init(void *handle) ...@@ -1377,13 +1377,10 @@ static int dm_late_init(void *handle)
*/ */
params.min_abm_backlight = 0x28F; params.min_abm_backlight = 0x28F;
/* todo will enable for navi10 */ ret = dmcu_load_iram(dmcu, params);
if (adev->asic_type <= CHIP_RAVEN) {
ret = dmcu_load_iram(dmcu, params);
if (!ret) if (!ret)
return -EINVAL; return -EINVAL;
}
return detect_mst_link_for_all_connectors(adev->ddev); return detect_mst_link_for_all_connectors(adev->ddev);
} }
......
...@@ -1016,9 +1016,17 @@ static void program_timing_sync( ...@@ -1016,9 +1016,17 @@ static void program_timing_sync(
} }
} }
/* set first pipe with plane as master */ /* set first unblanked pipe as master */
for (j = 0; j < group_size; j++) { for (j = 0; j < group_size; j++) {
if (pipe_set[j]->plane_state) { bool is_blanked;
if (pipe_set[j]->stream_res.opp->funcs->dpg_is_blanked)
is_blanked =
pipe_set[j]->stream_res.opp->funcs->dpg_is_blanked(pipe_set[j]->stream_res.opp);
else
is_blanked =
pipe_set[j]->stream_res.tg->funcs->is_blanked(pipe_set[j]->stream_res.tg);
if (!is_blanked) {
if (j == 0) if (j == 0)
break; break;
...@@ -1039,9 +1047,17 @@ static void program_timing_sync( ...@@ -1039,9 +1047,17 @@ static void program_timing_sync(
status->timing_sync_info.master = false; status->timing_sync_info.master = false;
} }
/* remove any other pipes with plane as they have already been synced */ /* remove any other unblanked pipes as they have already been synced */
for (j = j + 1; j < group_size; j++) { for (j = j + 1; j < group_size; j++) {
if (pipe_set[j]->plane_state) { bool is_blanked;
if (pipe_set[j]->stream_res.opp->funcs->dpg_is_blanked)
is_blanked =
pipe_set[j]->stream_res.opp->funcs->dpg_is_blanked(pipe_set[j]->stream_res.opp);
else
is_blanked =
pipe_set[j]->stream_res.tg->funcs->is_blanked(pipe_set[j]->stream_res.tg);
if (!is_blanked) {
group_size--; group_size--;
pipe_set[j] = pipe_set[group_size]; pipe_set[j] = pipe_set[group_size];
j--; j--;
...@@ -2522,6 +2538,12 @@ void dc_commit_updates_for_stream(struct dc *dc, ...@@ -2522,6 +2538,12 @@ void dc_commit_updates_for_stream(struct dc *dc,
copy_stream_update_to_stream(dc, context, stream, stream_update); copy_stream_update_to_stream(dc, context, stream, stream_update);
if (!dc->res_pool->funcs->validate_bandwidth(dc, context, false)) {
DC_ERROR("Mode validation failed for stream update!\n");
dc_release_state(context);
return;
}
commit_planes_for_stream( commit_planes_for_stream(
dc, dc,
srf_updates, srf_updates,
......
...@@ -1561,6 +1561,7 @@ static int smu_v11_0_irq_process(struct amdgpu_device *adev, ...@@ -1561,6 +1561,7 @@ static int smu_v11_0_irq_process(struct amdgpu_device *adev,
* events for SMCToHost interrupt. * events for SMCToHost interrupt.
*/ */
uint32_t ctxid = entry->src_data[0]; uint32_t ctxid = entry->src_data[0];
uint32_t data;
if (client_id == SOC15_IH_CLIENTID_THM) { if (client_id == SOC15_IH_CLIENTID_THM) {
switch (src_id) { switch (src_id) {
...@@ -1590,6 +1591,11 @@ static int smu_v11_0_irq_process(struct amdgpu_device *adev, ...@@ -1590,6 +1591,11 @@ static int smu_v11_0_irq_process(struct amdgpu_device *adev,
orderly_poweroff(true); orderly_poweroff(true);
} else if (client_id == SOC15_IH_CLIENTID_MP1) { } else if (client_id == SOC15_IH_CLIENTID_MP1) {
if (src_id == 0xfe) { if (src_id == 0xfe) {
/* ACK SMUToHost interrupt */
data = RREG32_SOC15(MP1, 0, mmMP1_SMN_IH_SW_INT_CTRL);
data = REG_SET_FIELD(data, MP1_SMN_IH_SW_INT_CTRL, INT_ACK, 1);
WREG32_SOC15(MP1, 0, mmMP1_SMN_IH_SW_INT_CTRL, data);
switch (ctxid) { switch (ctxid) {
case 0x3: case 0x3:
dev_dbg(adev->dev, "Switched to AC mode!\n"); dev_dbg(adev->dev, "Switched to AC mode!\n");
......
...@@ -10,6 +10,28 @@ ...@@ -10,6 +10,28 @@
#include "intel_display_types.h" #include "intel_display_types.h"
#include "intel_global_state.h" #include "intel_global_state.h"
static void __intel_atomic_global_state_free(struct kref *kref)
{
struct intel_global_state *obj_state =
container_of(kref, struct intel_global_state, ref);
struct intel_global_obj *obj = obj_state->obj;
obj->funcs->atomic_destroy_state(obj, obj_state);
}
static void intel_atomic_global_state_put(struct intel_global_state *obj_state)
{
kref_put(&obj_state->ref, __intel_atomic_global_state_free);
}
static struct intel_global_state *
intel_atomic_global_state_get(struct intel_global_state *obj_state)
{
kref_get(&obj_state->ref);
return obj_state;
}
void intel_atomic_global_obj_init(struct drm_i915_private *dev_priv, void intel_atomic_global_obj_init(struct drm_i915_private *dev_priv,
struct intel_global_obj *obj, struct intel_global_obj *obj,
struct intel_global_state *state, struct intel_global_state *state,
...@@ -17,6 +39,10 @@ void intel_atomic_global_obj_init(struct drm_i915_private *dev_priv, ...@@ -17,6 +39,10 @@ void intel_atomic_global_obj_init(struct drm_i915_private *dev_priv,
{ {
memset(obj, 0, sizeof(*obj)); memset(obj, 0, sizeof(*obj));
state->obj = obj;
kref_init(&state->ref);
obj->state = state; obj->state = state;
obj->funcs = funcs; obj->funcs = funcs;
list_add_tail(&obj->head, &dev_priv->global_obj_list); list_add_tail(&obj->head, &dev_priv->global_obj_list);
...@@ -28,7 +54,9 @@ void intel_atomic_global_obj_cleanup(struct drm_i915_private *dev_priv) ...@@ -28,7 +54,9 @@ void intel_atomic_global_obj_cleanup(struct drm_i915_private *dev_priv)
list_for_each_entry_safe(obj, next, &dev_priv->global_obj_list, head) { list_for_each_entry_safe(obj, next, &dev_priv->global_obj_list, head) {
list_del(&obj->head); list_del(&obj->head);
obj->funcs->atomic_destroy_state(obj, obj->state);
drm_WARN_ON(&dev_priv->drm, kref_read(&obj->state->ref) != 1);
intel_atomic_global_state_put(obj->state);
} }
} }
...@@ -97,10 +125,14 @@ intel_atomic_get_global_obj_state(struct intel_atomic_state *state, ...@@ -97,10 +125,14 @@ intel_atomic_get_global_obj_state(struct intel_atomic_state *state,
if (!obj_state) if (!obj_state)
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
obj_state->obj = obj;
obj_state->changed = false; obj_state->changed = false;
kref_init(&obj_state->ref);
state->global_objs[index].state = obj_state; state->global_objs[index].state = obj_state;
state->global_objs[index].old_state = obj->state; state->global_objs[index].old_state =
intel_atomic_global_state_get(obj->state);
state->global_objs[index].new_state = obj_state; state->global_objs[index].new_state = obj_state;
state->global_objs[index].ptr = obj; state->global_objs[index].ptr = obj;
obj_state->state = state; obj_state->state = state;
...@@ -163,7 +195,9 @@ void intel_atomic_swap_global_state(struct intel_atomic_state *state) ...@@ -163,7 +195,9 @@ void intel_atomic_swap_global_state(struct intel_atomic_state *state)
new_obj_state->state = NULL; new_obj_state->state = NULL;
state->global_objs[i].state = old_obj_state; state->global_objs[i].state = old_obj_state;
obj->state = new_obj_state;
intel_atomic_global_state_put(obj->state);
obj->state = intel_atomic_global_state_get(new_obj_state);
} }
} }
...@@ -172,10 +206,9 @@ void intel_atomic_clear_global_state(struct intel_atomic_state *state) ...@@ -172,10 +206,9 @@ void intel_atomic_clear_global_state(struct intel_atomic_state *state)
int i; int i;
for (i = 0; i < state->num_global_objs; i++) { for (i = 0; i < state->num_global_objs; i++) {
struct intel_global_obj *obj = state->global_objs[i].ptr; intel_atomic_global_state_put(state->global_objs[i].old_state);
intel_atomic_global_state_put(state->global_objs[i].new_state);
obj->funcs->atomic_destroy_state(obj,
state->global_objs[i].state);
state->global_objs[i].ptr = NULL; state->global_objs[i].ptr = NULL;
state->global_objs[i].state = NULL; state->global_objs[i].state = NULL;
state->global_objs[i].old_state = NULL; state->global_objs[i].old_state = NULL;
......
...@@ -6,6 +6,7 @@ ...@@ -6,6 +6,7 @@
#ifndef __INTEL_GLOBAL_STATE_H__ #ifndef __INTEL_GLOBAL_STATE_H__
#define __INTEL_GLOBAL_STATE_H__ #define __INTEL_GLOBAL_STATE_H__
#include <linux/kref.h>
#include <linux/list.h> #include <linux/list.h>
struct drm_i915_private; struct drm_i915_private;
...@@ -54,7 +55,9 @@ struct intel_global_obj { ...@@ -54,7 +55,9 @@ struct intel_global_obj {
for_each_if(obj) for_each_if(obj)
struct intel_global_state { struct intel_global_state {
struct intel_global_obj *obj;
struct intel_atomic_state *state; struct intel_atomic_state *state;
struct kref ref;
bool changed; bool changed;
}; };
......
...@@ -230,7 +230,7 @@ static void intel_context_set_gem(struct intel_context *ce, ...@@ -230,7 +230,7 @@ static void intel_context_set_gem(struct intel_context *ce,
ce->timeline = intel_timeline_get(ctx->timeline); ce->timeline = intel_timeline_get(ctx->timeline);
if (ctx->sched.priority >= I915_PRIORITY_NORMAL && if (ctx->sched.priority >= I915_PRIORITY_NORMAL &&
intel_engine_has_semaphores(ce->engine)) intel_engine_has_timeslices(ce->engine))
__set_bit(CONTEXT_USE_SEMAPHORES, &ce->flags); __set_bit(CONTEXT_USE_SEMAPHORES, &ce->flags);
} }
...@@ -1969,7 +1969,7 @@ static int __apply_priority(struct intel_context *ce, void *arg) ...@@ -1969,7 +1969,7 @@ static int __apply_priority(struct intel_context *ce, void *arg)
{ {
struct i915_gem_context *ctx = arg; struct i915_gem_context *ctx = arg;
if (!intel_engine_has_semaphores(ce->engine)) if (!intel_engine_has_timeslices(ce->engine))
return 0; return 0;
if (ctx->sched.priority >= I915_PRIORITY_NORMAL) if (ctx->sched.priority >= I915_PRIORITY_NORMAL)
......
...@@ -39,7 +39,6 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj) ...@@ -39,7 +39,6 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj)
unsigned long last_pfn = 0; /* suppress gcc warning */ unsigned long last_pfn = 0; /* suppress gcc warning */
unsigned int max_segment = i915_sg_segment_size(); unsigned int max_segment = i915_sg_segment_size();
unsigned int sg_page_sizes; unsigned int sg_page_sizes;
struct pagevec pvec;
gfp_t noreclaim; gfp_t noreclaim;
int ret; int ret;
...@@ -192,13 +191,17 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj) ...@@ -192,13 +191,17 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj)
sg_mark_end(sg); sg_mark_end(sg);
err_pages: err_pages:
mapping_clear_unevictable(mapping); mapping_clear_unevictable(mapping);
pagevec_init(&pvec); if (sg != st->sgl) {
for_each_sgt_page(page, sgt_iter, st) { struct pagevec pvec;
if (!pagevec_add(&pvec, page))
pagevec_init(&pvec);
for_each_sgt_page(page, sgt_iter, st) {
if (!pagevec_add(&pvec, page))
check_release_pagevec(&pvec);
}
if (pagevec_count(&pvec))
check_release_pagevec(&pvec); check_release_pagevec(&pvec);
} }
if (pagevec_count(&pvec))
check_release_pagevec(&pvec);
sg_free_table(st); sg_free_table(st);
kfree(st); kfree(st);
......
...@@ -97,8 +97,6 @@ int __intel_context_do_pin(struct intel_context *ce) ...@@ -97,8 +97,6 @@ int __intel_context_do_pin(struct intel_context *ce)
{ {
int err; int err;
GEM_BUG_ON(intel_context_is_closed(ce));
if (unlikely(!test_bit(CONTEXT_ALLOC_BIT, &ce->flags))) { if (unlikely(!test_bit(CONTEXT_ALLOC_BIT, &ce->flags))) {
err = intel_context_alloc_state(ce); err = intel_context_alloc_state(ce);
if (err) if (err)
......
...@@ -124,7 +124,7 @@ int intel_gvt_init_vgpu_types(struct intel_gvt *gvt) ...@@ -124,7 +124,7 @@ int intel_gvt_init_vgpu_types(struct intel_gvt *gvt)
*/ */
low_avail = gvt_aperture_sz(gvt) - HOST_LOW_GM_SIZE; low_avail = gvt_aperture_sz(gvt) - HOST_LOW_GM_SIZE;
high_avail = gvt_hidden_sz(gvt) - HOST_HIGH_GM_SIZE; high_avail = gvt_hidden_sz(gvt) - HOST_HIGH_GM_SIZE;
num_types = sizeof(vgpu_types) / sizeof(vgpu_types[0]); num_types = ARRAY_SIZE(vgpu_types);
gvt->types = kcalloc(num_types, sizeof(struct intel_vgpu_type), gvt->types = kcalloc(num_types, sizeof(struct intel_vgpu_type),
GFP_KERNEL); GFP_KERNEL);
......
...@@ -572,6 +572,9 @@ struct drm_i915_reg_descriptor { ...@@ -572,6 +572,9 @@ struct drm_i915_reg_descriptor {
#define REG32(_reg, ...) \ #define REG32(_reg, ...) \
{ .addr = (_reg), __VA_ARGS__ } { .addr = (_reg), __VA_ARGS__ }
#define REG32_IDX(_reg, idx) \
{ .addr = _reg(idx) }
/* /*
* Convenience macro for adding 64-bit registers. * Convenience macro for adding 64-bit registers.
* *
...@@ -669,6 +672,7 @@ static const struct drm_i915_reg_descriptor gen9_blt_regs[] = { ...@@ -669,6 +672,7 @@ static const struct drm_i915_reg_descriptor gen9_blt_regs[] = {
REG64_IDX(RING_TIMESTAMP, BSD_RING_BASE), REG64_IDX(RING_TIMESTAMP, BSD_RING_BASE),
REG32(BCS_SWCTRL), REG32(BCS_SWCTRL),
REG64_IDX(RING_TIMESTAMP, BLT_RING_BASE), REG64_IDX(RING_TIMESTAMP, BLT_RING_BASE),
REG32_IDX(RING_CTX_TIMESTAMP, BLT_RING_BASE),
REG64_IDX(BCS_GPR, 0), REG64_IDX(BCS_GPR, 0),
REG64_IDX(BCS_GPR, 1), REG64_IDX(BCS_GPR, 1),
REG64_IDX(BCS_GPR, 2), REG64_IDX(BCS_GPR, 2),
......
...@@ -173,7 +173,7 @@ i915_param_named(enable_gvt, bool, 0400, ...@@ -173,7 +173,7 @@ i915_param_named(enable_gvt, bool, 0400,
#endif #endif
#if IS_ENABLED(CONFIG_DRM_I915_UNSTABLE_FAKE_LMEM) #if IS_ENABLED(CONFIG_DRM_I915_UNSTABLE_FAKE_LMEM)
i915_param_named_unsafe(fake_lmem_start, ulong, 0600, i915_param_named_unsafe(fake_lmem_start, ulong, 0400,
"Fake LMEM start offset (default: 0)"); "Fake LMEM start offset (default: 0)");
#endif #endif
......
...@@ -64,7 +64,7 @@ struct drm_printer; ...@@ -64,7 +64,7 @@ struct drm_printer;
param(int, mmio_debug, -IS_ENABLED(CONFIG_DRM_I915_DEBUG_MMIO), 0600) \ param(int, mmio_debug, -IS_ENABLED(CONFIG_DRM_I915_DEBUG_MMIO), 0600) \
param(int, edp_vswing, 0, 0400) \ param(int, edp_vswing, 0, 0400) \
param(unsigned int, reset, 3, 0600) \ param(unsigned int, reset, 3, 0600) \
param(unsigned int, inject_probe_failure, 0, 0600) \ param(unsigned int, inject_probe_failure, 0, 0) \
param(int, fastboot, -1, 0600) \ param(int, fastboot, -1, 0600) \
param(int, enable_dpcd_backlight, -1, 0600) \ param(int, enable_dpcd_backlight, -1, 0600) \
param(char *, force_probe, CONFIG_DRM_I915_FORCE_PROBE, 0400) \ param(char *, force_probe, CONFIG_DRM_I915_FORCE_PROBE, 0400) \
......
This diff is collapsed.
...@@ -209,14 +209,6 @@ static void kick_submission(struct intel_engine_cs *engine, ...@@ -209,14 +209,6 @@ static void kick_submission(struct intel_engine_cs *engine,
if (!inflight) if (!inflight)
goto unlock; goto unlock;
ENGINE_TRACE(engine,
"bumping queue-priority-hint:%d for rq:%llx:%lld, inflight:%llx:%lld prio %d\n",
prio,
rq->fence.context, rq->fence.seqno,
inflight->fence.context, inflight->fence.seqno,
inflight->sched.attr.priority);
engine->execlists.queue_priority_hint = prio;
/* /*
* If we are already the currently executing context, don't * If we are already the currently executing context, don't
* bother evaluating if we should preempt ourselves. * bother evaluating if we should preempt ourselves.
...@@ -224,6 +216,14 @@ static void kick_submission(struct intel_engine_cs *engine, ...@@ -224,6 +216,14 @@ static void kick_submission(struct intel_engine_cs *engine,
if (inflight->context == rq->context) if (inflight->context == rq->context)
goto unlock; goto unlock;
ENGINE_TRACE(engine,
"bumping queue-priority-hint:%d for rq:%llx:%lld, inflight:%llx:%lld prio %d\n",
prio,
rq->fence.context, rq->fence.seqno,
inflight->fence.context, inflight->fence.seqno,
inflight->sched.attr.priority);
engine->execlists.queue_priority_hint = prio;
if (need_preempt(prio, rq_prio(inflight))) if (need_preempt(prio, rq_prio(inflight)))
tasklet_hi_schedule(&engine->execlists.tasklet); tasklet_hi_schedule(&engine->execlists.tasklet);
......
...@@ -277,7 +277,7 @@ nv50_outp_release(struct nouveau_encoder *nv_encoder) ...@@ -277,7 +277,7 @@ nv50_outp_release(struct nouveau_encoder *nv_encoder)
} }
static int static int
nv50_outp_acquire(struct nouveau_encoder *nv_encoder) nv50_outp_acquire(struct nouveau_encoder *nv_encoder, bool hda)
{ {
struct nouveau_drm *drm = nouveau_drm(nv_encoder->base.base.dev); struct nouveau_drm *drm = nouveau_drm(nv_encoder->base.base.dev);
struct nv50_disp *disp = nv50_disp(drm->dev); struct nv50_disp *disp = nv50_disp(drm->dev);
...@@ -289,6 +289,7 @@ nv50_outp_acquire(struct nouveau_encoder *nv_encoder) ...@@ -289,6 +289,7 @@ nv50_outp_acquire(struct nouveau_encoder *nv_encoder)
.base.method = NV50_DISP_MTHD_V1_ACQUIRE, .base.method = NV50_DISP_MTHD_V1_ACQUIRE,
.base.hasht = nv_encoder->dcb->hasht, .base.hasht = nv_encoder->dcb->hasht,
.base.hashm = nv_encoder->dcb->hashm, .base.hashm = nv_encoder->dcb->hashm,
.info.hda = hda,
}; };
int ret; int ret;
...@@ -393,7 +394,7 @@ nv50_dac_enable(struct drm_encoder *encoder) ...@@ -393,7 +394,7 @@ nv50_dac_enable(struct drm_encoder *encoder)
struct nv50_head_atom *asyh = nv50_head_atom(nv_crtc->base.state); struct nv50_head_atom *asyh = nv50_head_atom(nv_crtc->base.state);
struct nv50_core *core = nv50_disp(encoder->dev)->core; struct nv50_core *core = nv50_disp(encoder->dev)->core;
nv50_outp_acquire(nv_encoder); nv50_outp_acquire(nv_encoder, false);
core->func->dac->ctrl(core, nv_encoder->or, 1 << nv_crtc->index, asyh); core->func->dac->ctrl(core, nv_encoder->or, 1 << nv_crtc->index, asyh);
asyh->or.depth = 0; asyh->or.depth = 0;
...@@ -510,7 +511,7 @@ nv50_audio_component_get_eld(struct device *kdev, int port, int dev_id, ...@@ -510,7 +511,7 @@ nv50_audio_component_get_eld(struct device *kdev, int port, int dev_id,
if (!nv_connector || !nv_crtc || nv_encoder->or != port || if (!nv_connector || !nv_crtc || nv_encoder->or != port ||
nv_crtc->index != dev_id) nv_crtc->index != dev_id)
continue; continue;
*enabled = drm_detect_monitor_audio(nv_connector->edid); *enabled = nv_encoder->audio;
if (*enabled) { if (*enabled) {
ret = drm_eld_size(nv_connector->base.eld); ret = drm_eld_size(nv_connector->base.eld);
memcpy(buf, nv_connector->base.eld, memcpy(buf, nv_connector->base.eld,
...@@ -600,6 +601,7 @@ nv50_audio_disable(struct drm_encoder *encoder, struct nouveau_crtc *nv_crtc) ...@@ -600,6 +601,7 @@ nv50_audio_disable(struct drm_encoder *encoder, struct nouveau_crtc *nv_crtc)
(0x0100 << nv_crtc->index), (0x0100 << nv_crtc->index),
}; };
nv_encoder->audio = false;
nvif_mthd(&disp->disp->object, 0, &args, sizeof(args)); nvif_mthd(&disp->disp->object, 0, &args, sizeof(args));
nv50_audio_component_eld_notify(drm->audio.component, nv_encoder->or, nv50_audio_component_eld_notify(drm->audio.component, nv_encoder->or,
...@@ -636,6 +638,7 @@ nv50_audio_enable(struct drm_encoder *encoder, struct drm_display_mode *mode) ...@@ -636,6 +638,7 @@ nv50_audio_enable(struct drm_encoder *encoder, struct drm_display_mode *mode)
nvif_mthd(&disp->disp->object, 0, &args, nvif_mthd(&disp->disp->object, 0, &args,
sizeof(args.base) + drm_eld_size(args.data)); sizeof(args.base) + drm_eld_size(args.data));
nv_encoder->audio = true;
nv50_audio_component_eld_notify(drm->audio.component, nv_encoder->or, nv50_audio_component_eld_notify(drm->audio.component, nv_encoder->or,
nv_crtc->index); nv_crtc->index);
...@@ -966,7 +969,7 @@ nv50_msto_enable(struct drm_encoder *encoder) ...@@ -966,7 +969,7 @@ nv50_msto_enable(struct drm_encoder *encoder)
DRM_DEBUG_KMS("Failed to allocate VCPI\n"); DRM_DEBUG_KMS("Failed to allocate VCPI\n");
if (!mstm->links++) if (!mstm->links++)
nv50_outp_acquire(mstm->outp); nv50_outp_acquire(mstm->outp, false /*XXX: MST audio.*/);
if (mstm->outp->link & 1) if (mstm->outp->link & 1)
proto = 0x8; proto = 0x8;
...@@ -1560,12 +1563,18 @@ nv50_sor_enable(struct drm_encoder *encoder) ...@@ -1560,12 +1563,18 @@ nv50_sor_enable(struct drm_encoder *encoder)
struct nouveau_drm *drm = nouveau_drm(dev); struct nouveau_drm *drm = nouveau_drm(dev);
struct nouveau_connector *nv_connector; struct nouveau_connector *nv_connector;
struct nvbios *bios = &drm->vbios; struct nvbios *bios = &drm->vbios;
bool hda = false;
u8 proto = 0xf; u8 proto = 0xf;
u8 depth = 0x0; u8 depth = 0x0;
nv_connector = nouveau_encoder_connector_get(nv_encoder); nv_connector = nouveau_encoder_connector_get(nv_encoder);
nv_encoder->crtc = encoder->crtc; nv_encoder->crtc = encoder->crtc;
nv50_outp_acquire(nv_encoder);
if ((disp->disp->object.oclass == GT214_DISP ||
disp->disp->object.oclass >= GF110_DISP) &&
drm_detect_monitor_audio(nv_connector->edid))
hda = true;
nv50_outp_acquire(nv_encoder, hda);
switch (nv_encoder->dcb->type) { switch (nv_encoder->dcb->type) {
case DCB_OUTPUT_TMDS: case DCB_OUTPUT_TMDS:
...@@ -1775,7 +1784,7 @@ nv50_pior_enable(struct drm_encoder *encoder) ...@@ -1775,7 +1784,7 @@ nv50_pior_enable(struct drm_encoder *encoder)
u8 owner = 1 << nv_crtc->index; u8 owner = 1 << nv_crtc->index;
u8 proto; u8 proto;
nv50_outp_acquire(nv_encoder); nv50_outp_acquire(nv_encoder, false);
switch (asyh->or.bpc) { switch (asyh->or.bpc) {
case 10: asyh->or.depth = 0x6; break; case 10: asyh->or.depth = 0x6; break;
......
...@@ -192,6 +192,8 @@ nv50_wndw_atomic_check_release(struct nv50_wndw *wndw, ...@@ -192,6 +192,8 @@ nv50_wndw_atomic_check_release(struct nv50_wndw *wndw,
wndw->func->release(wndw, asyw, asyh); wndw->func->release(wndw, asyw, asyh);
asyw->ntfy.handle = 0; asyw->ntfy.handle = 0;
asyw->sema.handle = 0; asyw->sema.handle = 0;
asyw->xlut.handle = 0;
memset(asyw->image.handle, 0x00, sizeof(asyw->image.handle));
} }
static int static int
...@@ -519,7 +521,8 @@ nv50_wndw_prepare_fb(struct drm_plane *plane, struct drm_plane_state *state) ...@@ -519,7 +521,8 @@ nv50_wndw_prepare_fb(struct drm_plane *plane, struct drm_plane_state *state)
return PTR_ERR(ctxdma); return PTR_ERR(ctxdma);
} }
asyw->image.handle[0] = ctxdma->object.handle; if (asyw->visible)
asyw->image.handle[0] = ctxdma->object.handle;
} }
asyw->state.fence = dma_resv_get_excl_rcu(nvbo->bo.base.resv); asyw->state.fence = dma_resv_get_excl_rcu(nvbo->bo.base.resv);
......
...@@ -46,7 +46,8 @@ struct nv50_disp_acquire_v0 { ...@@ -46,7 +46,8 @@ struct nv50_disp_acquire_v0 {
__u8 version; __u8 version;
__u8 or; __u8 or;
__u8 link; __u8 link;
__u8 pad03[5]; __u8 hda;
__u8 pad04[4];
}; };
struct nv50_disp_dac_load_v0 { struct nv50_disp_dac_load_v0 {
......
...@@ -52,6 +52,7 @@ struct nouveau_encoder { ...@@ -52,6 +52,7 @@ struct nouveau_encoder {
* actually programmed on the hw, not the proposed crtc */ * actually programmed on the hw, not the proposed crtc */
struct drm_crtc *crtc; struct drm_crtc *crtc;
u32 ctrl; u32 ctrl;
bool audio;
struct drm_display_mode mode; struct drm_display_mode mode;
int last_dpms; int last_dpms;
......
...@@ -39,6 +39,7 @@ nvkm-y += nvkm/engine/disp/sorgf119.o ...@@ -39,6 +39,7 @@ nvkm-y += nvkm/engine/disp/sorgf119.o
nvkm-y += nvkm/engine/disp/sorgk104.o nvkm-y += nvkm/engine/disp/sorgk104.o
nvkm-y += nvkm/engine/disp/sorgm107.o nvkm-y += nvkm/engine/disp/sorgm107.o
nvkm-y += nvkm/engine/disp/sorgm200.o nvkm-y += nvkm/engine/disp/sorgm200.o
nvkm-y += nvkm/engine/disp/sorgp100.o
nvkm-y += nvkm/engine/disp/sorgv100.o nvkm-y += nvkm/engine/disp/sorgv100.o
nvkm-y += nvkm/engine/disp/sortu102.o nvkm-y += nvkm/engine/disp/sortu102.o
......
...@@ -36,7 +36,7 @@ gp100_disp = { ...@@ -36,7 +36,7 @@ gp100_disp = {
.super = gf119_disp_super, .super = gf119_disp_super,
.root = &gp100_disp_root_oclass, .root = &gp100_disp_root_oclass,
.head = { .cnt = gf119_head_cnt, .new = gf119_head_new }, .head = { .cnt = gf119_head_cnt, .new = gf119_head_new },
.sor = { .cnt = gf119_sor_cnt, .new = gm200_sor_new }, .sor = { .cnt = gf119_sor_cnt, .new = gp100_sor_new },
}; };
int int
......
...@@ -63,7 +63,7 @@ gp102_disp = { ...@@ -63,7 +63,7 @@ gp102_disp = {
.super = gf119_disp_super, .super = gf119_disp_super,
.root = &gp102_disp_root_oclass, .root = &gp102_disp_root_oclass,
.head = { .cnt = gf119_head_cnt, .new = gf119_head_new }, .head = { .cnt = gf119_head_cnt, .new = gf119_head_new },
.sor = { .cnt = gf119_sor_cnt, .new = gm200_sor_new }, .sor = { .cnt = gf119_sor_cnt, .new = gp100_sor_new },
}; };
int int
......
...@@ -27,10 +27,10 @@ void ...@@ -27,10 +27,10 @@ void
gm200_hdmi_scdc(struct nvkm_ior *ior, int head, u8 scdc) gm200_hdmi_scdc(struct nvkm_ior *ior, int head, u8 scdc)
{ {
struct nvkm_device *device = ior->disp->engine.subdev.device; struct nvkm_device *device = ior->disp->engine.subdev.device;
const u32 hoff = head * 0x800; const u32 soff = nv50_ior_base(ior);
const u32 ctrl = scdc & 0x3; const u32 ctrl = scdc & 0x3;
nvkm_mask(device, 0x61c5bc + hoff, 0x00000003, ctrl); nvkm_mask(device, 0x61c5bc + soff, 0x00000003, ctrl);
ior->tmds.high_speed = !!(scdc & 0x2); ior->tmds.high_speed = !!(scdc & 0x2);
} }
...@@ -201,6 +201,7 @@ int gf119_sor_new(struct nvkm_disp *, int); ...@@ -201,6 +201,7 @@ int gf119_sor_new(struct nvkm_disp *, int);
int gk104_sor_new(struct nvkm_disp *, int); int gk104_sor_new(struct nvkm_disp *, int);
int gm107_sor_new(struct nvkm_disp *, int); int gm107_sor_new(struct nvkm_disp *, int);
int gm200_sor_new(struct nvkm_disp *, int); int gm200_sor_new(struct nvkm_disp *, int);
int gp100_sor_new(struct nvkm_disp *, int);
int gv100_sor_cnt(struct nvkm_disp *, unsigned long *); int gv100_sor_cnt(struct nvkm_disp *, unsigned long *);
int gv100_sor_new(struct nvkm_disp *, int); int gv100_sor_new(struct nvkm_disp *, int);
......
...@@ -111,8 +111,44 @@ nvkm_outp_acquire_ior(struct nvkm_outp *outp, u8 user, struct nvkm_ior *ior) ...@@ -111,8 +111,44 @@ nvkm_outp_acquire_ior(struct nvkm_outp *outp, u8 user, struct nvkm_ior *ior)
return 0; return 0;
} }
static inline int
nvkm_outp_acquire_hda(struct nvkm_outp *outp, enum nvkm_ior_type type,
u8 user, bool hda)
{
struct nvkm_ior *ior;
/* First preference is to reuse the OR that is currently armed
* on HW, if any, in order to prevent unnecessary switching.
*/
list_for_each_entry(ior, &outp->disp->ior, head) {
if (!ior->identity && !!ior->func->hda.hpd == hda &&
!ior->asy.outp && ior->arm.outp == outp)
return nvkm_outp_acquire_ior(outp, user, ior);
}
/* Failing that, a completely unused OR is the next best thing. */
list_for_each_entry(ior, &outp->disp->ior, head) {
if (!ior->identity && !!ior->func->hda.hpd == hda &&
!ior->asy.outp && ior->type == type && !ior->arm.outp &&
(ior->func->route.set || ior->id == __ffs(outp->info.or)))
return nvkm_outp_acquire_ior(outp, user, ior);
}
/* Last resort is to assign an OR that's already active on HW,
* but will be released during the next modeset.
*/
list_for_each_entry(ior, &outp->disp->ior, head) {
if (!ior->identity && !!ior->func->hda.hpd == hda &&
!ior->asy.outp && ior->type == type &&
(ior->func->route.set || ior->id == __ffs(outp->info.or)))
return nvkm_outp_acquire_ior(outp, user, ior);
}
return -ENOSPC;
}
int int
nvkm_outp_acquire(struct nvkm_outp *outp, u8 user) nvkm_outp_acquire(struct nvkm_outp *outp, u8 user, bool hda)
{ {
struct nvkm_ior *ior = outp->ior; struct nvkm_ior *ior = outp->ior;
enum nvkm_ior_proto proto; enum nvkm_ior_proto proto;
...@@ -137,32 +173,25 @@ nvkm_outp_acquire(struct nvkm_outp *outp, u8 user) ...@@ -137,32 +173,25 @@ nvkm_outp_acquire(struct nvkm_outp *outp, u8 user)
return nvkm_outp_acquire_ior(outp, user, ior); return nvkm_outp_acquire_ior(outp, user, ior);
} }
/* First preference is to reuse the OR that is currently armed /* If we don't need HDA, first try to acquire an OR that doesn't
* on HW, if any, in order to prevent unnecessary switching. * support it to leave free the ones that do.
*/ */
list_for_each_entry(ior, &outp->disp->ior, head) { if (!hda) {
if (!ior->identity && !ior->asy.outp && ior->arm.outp == outp) if (!nvkm_outp_acquire_hda(outp, type, user, false))
return nvkm_outp_acquire_ior(outp, user, ior); return 0;
}
/* Failing that, a completely unused OR is the next best thing. */ /* Use a HDA-supporting SOR anyway. */
list_for_each_entry(ior, &outp->disp->ior, head) { return nvkm_outp_acquire_hda(outp, type, user, true);
if (!ior->identity &&
!ior->asy.outp && ior->type == type && !ior->arm.outp &&
(ior->func->route.set || ior->id == __ffs(outp->info.or)))
return nvkm_outp_acquire_ior(outp, user, ior);
} }
/* Last resort is to assign an OR that's already active on HW, /* We want HDA, try to acquire an OR that supports it. */
* but will be released during the next modeset. if (!nvkm_outp_acquire_hda(outp, type, user, true))
*/ return 0;
list_for_each_entry(ior, &outp->disp->ior, head) {
if (!ior->identity && !ior->asy.outp && ior->type == type &&
(ior->func->route.set || ior->id == __ffs(outp->info.or)))
return nvkm_outp_acquire_ior(outp, user, ior);
}
return -ENOSPC; /* There weren't any free ORs that support HDA, grab one that
* doesn't and at least allow display to work still.
*/
return nvkm_outp_acquire_hda(outp, type, user, false);
} }
void void
......
...@@ -32,7 +32,7 @@ int nvkm_outp_new(struct nvkm_disp *, int index, struct dcb_output *, ...@@ -32,7 +32,7 @@ int nvkm_outp_new(struct nvkm_disp *, int index, struct dcb_output *,
void nvkm_outp_del(struct nvkm_outp **); void nvkm_outp_del(struct nvkm_outp **);
void nvkm_outp_init(struct nvkm_outp *); void nvkm_outp_init(struct nvkm_outp *);
void nvkm_outp_fini(struct nvkm_outp *); void nvkm_outp_fini(struct nvkm_outp *);
int nvkm_outp_acquire(struct nvkm_outp *, u8 user); int nvkm_outp_acquire(struct nvkm_outp *, u8 user, bool hda);
void nvkm_outp_release(struct nvkm_outp *, u8 user); void nvkm_outp_release(struct nvkm_outp *, u8 user);
void nvkm_outp_route(struct nvkm_disp *); void nvkm_outp_route(struct nvkm_disp *);
......
...@@ -99,7 +99,7 @@ nv50_disp_root_mthd_(struct nvkm_object *object, u32 mthd, void *data, u32 size) ...@@ -99,7 +99,7 @@ nv50_disp_root_mthd_(struct nvkm_object *object, u32 mthd, void *data, u32 size)
} *args = data; } *args = data;
int ret = -ENOSYS; int ret = -ENOSYS;
if (!(ret = nvif_unpack(ret, &data, &size, args->v0, 0, 0, false))) { if (!(ret = nvif_unpack(ret, &data, &size, args->v0, 0, 0, false))) {
ret = nvkm_outp_acquire(outp, NVKM_OUTP_USER); ret = nvkm_outp_acquire(outp, NVKM_OUTP_USER, args->v0.hda);
if (ret == 0) { if (ret == 0) {
args->v0.or = outp->ior->id; args->v0.or = outp->ior->id;
args->v0.link = outp->ior->asy.link; args->v0.link = outp->ior->asy.link;
...@@ -119,7 +119,7 @@ nv50_disp_root_mthd_(struct nvkm_object *object, u32 mthd, void *data, u32 size) ...@@ -119,7 +119,7 @@ nv50_disp_root_mthd_(struct nvkm_object *object, u32 mthd, void *data, u32 size)
if (!(ret = nvif_unpack(ret, &data, &size, args->v0, 0, 0, false))) { if (!(ret = nvif_unpack(ret, &data, &size, args->v0, 0, 0, false))) {
if (args->v0.data & 0xfff00000) if (args->v0.data & 0xfff00000)
return -EINVAL; return -EINVAL;
ret = nvkm_outp_acquire(outp, NVKM_OUTP_PRIV); ret = nvkm_outp_acquire(outp, NVKM_OUTP_PRIV, false);
if (ret) if (ret)
return ret; return ret;
ret = outp->ior->func->sense(outp->ior, args->v0.data); ret = outp->ior->func->sense(outp->ior, args->v0.data);
......
...@@ -89,7 +89,7 @@ gm200_sor_route_get(struct nvkm_outp *outp, int *link) ...@@ -89,7 +89,7 @@ gm200_sor_route_get(struct nvkm_outp *outp, int *link)
} }
static const struct nvkm_ior_func static const struct nvkm_ior_func
gm200_sor = { gm200_sor_hda = {
.route = { .route = {
.get = gm200_sor_route_get, .get = gm200_sor_route_get,
.set = gm200_sor_route_set, .set = gm200_sor_route_set,
...@@ -119,8 +119,42 @@ gm200_sor = { ...@@ -119,8 +119,42 @@ gm200_sor = {
}, },
}; };
static const struct nvkm_ior_func
gm200_sor = {
.route = {
.get = gm200_sor_route_get,
.set = gm200_sor_route_set,
},
.state = gf119_sor_state,
.power = nv50_sor_power,
.clock = gf119_sor_clock,
.hdmi = {
.ctrl = gk104_hdmi_ctrl,
.scdc = gm200_hdmi_scdc,
},
.dp = {
.lanes = { 0, 1, 2, 3 },
.links = gf119_sor_dp_links,
.power = g94_sor_dp_power,
.pattern = gm107_sor_dp_pattern,
.drive = gm200_sor_dp_drive,
.vcpi = gf119_sor_dp_vcpi,
.audio = gf119_sor_dp_audio,
.audio_sym = gf119_sor_dp_audio_sym,
.watermark = gf119_sor_dp_watermark,
},
};
int int
gm200_sor_new(struct nvkm_disp *disp, int id) gm200_sor_new(struct nvkm_disp *disp, int id)
{ {
struct nvkm_device *device = disp->engine.subdev.device;
u32 hda;
if (!((hda = nvkm_rd32(device, 0x08a15c)) & 0x40000000))
hda = nvkm_rd32(device, 0x101034);
if (hda & BIT(id))
return nvkm_ior_new_(&gm200_sor_hda, disp, SOR, id);
return nvkm_ior_new_(&gm200_sor, disp, SOR, id); return nvkm_ior_new_(&gm200_sor, disp, SOR, id);
} }
/*
* Copyright 2020 Red Hat Inc.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
* OTHER DEALINGS IN THE SOFTWARE.
*/
#include "ior.h"
static const struct nvkm_ior_func
gp100_sor_hda = {
.route = {
.get = gm200_sor_route_get,
.set = gm200_sor_route_set,
},
.state = gf119_sor_state,
.power = nv50_sor_power,
.clock = gf119_sor_clock,
.hdmi = {
.ctrl = gk104_hdmi_ctrl,
.scdc = gm200_hdmi_scdc,
},
.dp = {
.lanes = { 0, 1, 2, 3 },
.links = gf119_sor_dp_links,
.power = g94_sor_dp_power,
.pattern = gm107_sor_dp_pattern,
.drive = gm200_sor_dp_drive,
.vcpi = gf119_sor_dp_vcpi,
.audio = gf119_sor_dp_audio,
.audio_sym = gf119_sor_dp_audio_sym,
.watermark = gf119_sor_dp_watermark,
},
.hda = {
.hpd = gf119_hda_hpd,
.eld = gf119_hda_eld,
.device_entry = gf119_hda_device_entry,
},
};
static const struct nvkm_ior_func
gp100_sor = {
.route = {
.get = gm200_sor_route_get,
.set = gm200_sor_route_set,
},
.state = gf119_sor_state,
.power = nv50_sor_power,
.clock = gf119_sor_clock,
.hdmi = {
.ctrl = gk104_hdmi_ctrl,
.scdc = gm200_hdmi_scdc,
},
.dp = {
.lanes = { 0, 1, 2, 3 },
.links = gf119_sor_dp_links,
.power = g94_sor_dp_power,
.pattern = gm107_sor_dp_pattern,
.drive = gm200_sor_dp_drive,
.vcpi = gf119_sor_dp_vcpi,
.audio = gf119_sor_dp_audio,
.audio_sym = gf119_sor_dp_audio_sym,
.watermark = gf119_sor_dp_watermark,
},
};
int
gp100_sor_new(struct nvkm_disp *disp, int id)
{
struct nvkm_device *device = disp->engine.subdev.device;
u32 hda;
if (!((hda = nvkm_rd32(device, 0x08a15c)) & 0x40000000))
hda = nvkm_rd32(device, 0x10ebb0) >> 8;
if (hda & BIT(id))
return nvkm_ior_new_(&gp100_sor_hda, disp, SOR, id);
return nvkm_ior_new_(&gp100_sor, disp, SOR, id);
}
...@@ -78,7 +78,7 @@ gv100_sor_state(struct nvkm_ior *sor, struct nvkm_ior_state *state) ...@@ -78,7 +78,7 @@ gv100_sor_state(struct nvkm_ior *sor, struct nvkm_ior_state *state)
} }
static const struct nvkm_ior_func static const struct nvkm_ior_func
gv100_sor = { gv100_sor_hda = {
.route = { .route = {
.get = gm200_sor_route_get, .get = gm200_sor_route_get,
.set = gm200_sor_route_set, .set = gm200_sor_route_set,
...@@ -107,9 +107,42 @@ gv100_sor = { ...@@ -107,9 +107,42 @@ gv100_sor = {
}, },
}; };
static const struct nvkm_ior_func
gv100_sor = {
.route = {
.get = gm200_sor_route_get,
.set = gm200_sor_route_set,
},
.state = gv100_sor_state,
.power = nv50_sor_power,
.clock = gf119_sor_clock,
.hdmi = {
.ctrl = gv100_hdmi_ctrl,
.scdc = gm200_hdmi_scdc,
},
.dp = {
.lanes = { 0, 1, 2, 3 },
.links = gf119_sor_dp_links,
.power = g94_sor_dp_power,
.pattern = gm107_sor_dp_pattern,
.drive = gm200_sor_dp_drive,
.audio = gv100_sor_dp_audio,
.audio_sym = gv100_sor_dp_audio_sym,
.watermark = gv100_sor_dp_watermark,
},
};
int int
gv100_sor_new(struct nvkm_disp *disp, int id) gv100_sor_new(struct nvkm_disp *disp, int id)
{ {
struct nvkm_device *device = disp->engine.subdev.device;
u32 hda;
if (!((hda = nvkm_rd32(device, 0x08a15c)) & 0x40000000))
hda = nvkm_rd32(device, 0x118fb0) >> 8;
if (hda & BIT(id))
return nvkm_ior_new_(&gv100_sor_hda, disp, SOR, id);
return nvkm_ior_new_(&gv100_sor, disp, SOR, id); return nvkm_ior_new_(&gv100_sor, disp, SOR, id);
} }
......
...@@ -62,7 +62,7 @@ tu102_sor_dp_links(struct nvkm_ior *sor, struct nvkm_i2c_aux *aux) ...@@ -62,7 +62,7 @@ tu102_sor_dp_links(struct nvkm_ior *sor, struct nvkm_i2c_aux *aux)
} }
static const struct nvkm_ior_func static const struct nvkm_ior_func
tu102_sor = { tu102_sor_hda = {
.route = { .route = {
.get = gm200_sor_route_get, .get = gm200_sor_route_get,
.set = gm200_sor_route_set, .set = gm200_sor_route_set,
...@@ -92,8 +92,38 @@ tu102_sor = { ...@@ -92,8 +92,38 @@ tu102_sor = {
}, },
}; };
static const struct nvkm_ior_func
tu102_sor = {
.route = {
.get = gm200_sor_route_get,
.set = gm200_sor_route_set,
},
.state = gv100_sor_state,
.power = nv50_sor_power,
.clock = gf119_sor_clock,
.hdmi = {
.ctrl = gv100_hdmi_ctrl,
.scdc = gm200_hdmi_scdc,
},
.dp = {
.lanes = { 0, 1, 2, 3 },
.links = tu102_sor_dp_links,
.power = g94_sor_dp_power,
.pattern = gm107_sor_dp_pattern,
.drive = gm200_sor_dp_drive,
.vcpi = tu102_sor_dp_vcpi,
.audio = gv100_sor_dp_audio,
.audio_sym = gv100_sor_dp_audio_sym,
.watermark = gv100_sor_dp_watermark,
},
};
int int
tu102_sor_new(struct nvkm_disp *disp, int id) tu102_sor_new(struct nvkm_disp *disp, int id)
{ {
struct nvkm_device *device = disp->engine.subdev.device;
u32 hda = nvkm_rd32(device, 0x08a15c);
if (hda & BIT(id))
return nvkm_ior_new_(&tu102_sor_hda, disp, SOR, id);
return nvkm_ior_new_(&tu102_sor, disp, SOR, id); return nvkm_ior_new_(&tu102_sor, disp, SOR, id);
} }
...@@ -352,7 +352,7 @@ gk20a_gr_load(struct gf100_gr *gr, int ver, const struct gf100_gr_fwif *fwif) ...@@ -352,7 +352,7 @@ gk20a_gr_load(struct gf100_gr *gr, int ver, const struct gf100_gr_fwif *fwif)
static const struct gf100_gr_fwif static const struct gf100_gr_fwif
gk20a_gr_fwif[] = { gk20a_gr_fwif[] = {
{ -1, gk20a_gr_load, &gk20a_gr }, { 0, gk20a_gr_load, &gk20a_gr },
{} {}
}; };
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment