mirror of
https://github.com/Qortal/Brooklyn.git
synced 2025-02-01 07:42:18 +00:00
d2ebfd0519
Screw the description like that inbred T3Q
446 lines
12 KiB
C
446 lines
12 KiB
C
// SPDX-License-Identifier: MIT
|
|
/*
|
|
* Copyright © 2021 Intel Corporation
|
|
*/
|
|
|
|
#include "i915_drv.h"
|
|
#include "intel_atomic.h"
|
|
#include "intel_de.h"
|
|
#include "intel_display_types.h"
|
|
#include "intel_drrs.h"
|
|
#include "intel_panel.h"
|
|
|
|
/**
|
|
* DOC: Display Refresh Rate Switching (DRRS)
|
|
*
|
|
* Display Refresh Rate Switching (DRRS) is a power conservation feature
|
|
* which enables swtching between low and high refresh rates,
|
|
* dynamically, based on the usage scenario. This feature is applicable
|
|
* for internal panels.
|
|
*
|
|
* Indication that the panel supports DRRS is given by the panel EDID, which
|
|
* would list multiple refresh rates for one resolution.
|
|
*
|
|
* DRRS is of 2 types - static and seamless.
|
|
* Static DRRS involves changing refresh rate (RR) by doing a full modeset
|
|
* (may appear as a blink on screen) and is used in dock-undock scenario.
|
|
* Seamless DRRS involves changing RR without any visual effect to the user
|
|
* and can be used during normal system usage. This is done by programming
|
|
* certain registers.
|
|
*
|
|
* Support for static/seamless DRRS may be indicated in the VBT based on
|
|
* inputs from the panel spec.
|
|
*
|
|
* DRRS saves power by switching to low RR based on usage scenarios.
|
|
*
|
|
* The implementation is based on frontbuffer tracking implementation. When
|
|
* there is a disturbance on the screen triggered by user activity or a periodic
|
|
* system activity, DRRS is disabled (RR is changed to high RR). When there is
|
|
* no movement on screen, after a timeout of 1 second, a switch to low RR is
|
|
* made.
|
|
*
|
|
* For integration with frontbuffer tracking code, intel_drrs_invalidate()
|
|
* and intel_drrs_flush() are called.
|
|
*
|
|
* DRRS can be further extended to support other internal panels and also
|
|
* the scenario of video playback wherein RR is set based on the rate
|
|
* requested by userspace.
|
|
*/
|
|
|
|
void
|
|
intel_drrs_compute_config(struct intel_dp *intel_dp,
|
|
struct intel_crtc_state *pipe_config,
|
|
int output_bpp, bool constant_n)
|
|
{
|
|
struct intel_connector *intel_connector = intel_dp->attached_connector;
|
|
struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
|
|
int pixel_clock;
|
|
|
|
if (pipe_config->vrr.enable)
|
|
return;
|
|
|
|
/*
|
|
* DRRS and PSR can't be enable together, so giving preference to PSR
|
|
* as it allows more power-savings by complete shutting down display,
|
|
* so to guarantee this, intel_drrs_compute_config() must be called
|
|
* after intel_psr_compute_config().
|
|
*/
|
|
if (pipe_config->has_psr)
|
|
return;
|
|
|
|
if (!intel_connector->panel.downclock_mode ||
|
|
dev_priv->drrs.type != SEAMLESS_DRRS_SUPPORT)
|
|
return;
|
|
|
|
pipe_config->has_drrs = true;
|
|
|
|
pixel_clock = intel_connector->panel.downclock_mode->clock;
|
|
if (pipe_config->splitter.enable)
|
|
pixel_clock /= pipe_config->splitter.link_count;
|
|
|
|
intel_link_compute_m_n(output_bpp, pipe_config->lane_count, pixel_clock,
|
|
pipe_config->port_clock, &pipe_config->dp_m2_n2,
|
|
constant_n, pipe_config->fec_enable);
|
|
|
|
/* FIXME: abstract this better */
|
|
if (pipe_config->splitter.enable)
|
|
pipe_config->dp_m2_n2.gmch_m *= pipe_config->splitter.link_count;
|
|
}
|
|
|
|
static void intel_drrs_set_state(struct drm_i915_private *dev_priv,
|
|
const struct intel_crtc_state *crtc_state,
|
|
enum drrs_refresh_rate_type refresh_type)
|
|
{
|
|
struct intel_dp *intel_dp = dev_priv->drrs.dp;
|
|
struct intel_crtc *crtc = to_intel_crtc(crtc_state->uapi.crtc);
|
|
struct drm_display_mode *mode;
|
|
|
|
if (!intel_dp) {
|
|
drm_dbg_kms(&dev_priv->drm, "DRRS not supported.\n");
|
|
return;
|
|
}
|
|
|
|
if (!crtc) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"DRRS: intel_crtc not initialized\n");
|
|
return;
|
|
}
|
|
|
|
if (dev_priv->drrs.type < SEAMLESS_DRRS_SUPPORT) {
|
|
drm_dbg_kms(&dev_priv->drm, "Only Seamless DRRS supported.\n");
|
|
return;
|
|
}
|
|
|
|
if (refresh_type == dev_priv->drrs.refresh_rate_type)
|
|
return;
|
|
|
|
if (!crtc_state->hw.active) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"eDP encoder disabled. CRTC not Active\n");
|
|
return;
|
|
}
|
|
|
|
if (DISPLAY_VER(dev_priv) >= 8 && !IS_CHERRYVIEW(dev_priv)) {
|
|
switch (refresh_type) {
|
|
case DRRS_HIGH_RR:
|
|
intel_dp_set_m_n(crtc_state, M1_N1);
|
|
break;
|
|
case DRRS_LOW_RR:
|
|
intel_dp_set_m_n(crtc_state, M2_N2);
|
|
break;
|
|
case DRRS_MAX_RR:
|
|
default:
|
|
drm_err(&dev_priv->drm,
|
|
"Unsupported refreshrate type\n");
|
|
}
|
|
} else if (DISPLAY_VER(dev_priv) > 6) {
|
|
i915_reg_t reg = PIPECONF(crtc_state->cpu_transcoder);
|
|
u32 val;
|
|
|
|
val = intel_de_read(dev_priv, reg);
|
|
if (refresh_type == DRRS_LOW_RR) {
|
|
if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
|
|
val |= PIPECONF_EDP_RR_MODE_SWITCH_VLV;
|
|
else
|
|
val |= PIPECONF_EDP_RR_MODE_SWITCH;
|
|
} else {
|
|
if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
|
|
val &= ~PIPECONF_EDP_RR_MODE_SWITCH_VLV;
|
|
else
|
|
val &= ~PIPECONF_EDP_RR_MODE_SWITCH;
|
|
}
|
|
intel_de_write(dev_priv, reg, val);
|
|
}
|
|
|
|
dev_priv->drrs.refresh_rate_type = refresh_type;
|
|
|
|
if (refresh_type == DRRS_LOW_RR)
|
|
mode = intel_dp->attached_connector->panel.downclock_mode;
|
|
else
|
|
mode = intel_dp->attached_connector->panel.fixed_mode;
|
|
drm_dbg_kms(&dev_priv->drm, "eDP Refresh Rate set to : %dHz\n",
|
|
drm_mode_vrefresh(mode));
|
|
}
|
|
|
|
static void
|
|
intel_drrs_enable_locked(struct intel_dp *intel_dp)
|
|
{
|
|
struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
|
|
|
|
dev_priv->drrs.busy_frontbuffer_bits = 0;
|
|
dev_priv->drrs.dp = intel_dp;
|
|
}
|
|
|
|
/**
|
|
* intel_drrs_enable - init drrs struct if supported
|
|
* @intel_dp: DP struct
|
|
* @crtc_state: A pointer to the active crtc state.
|
|
*
|
|
* Initializes frontbuffer_bits and drrs.dp
|
|
*/
|
|
void intel_drrs_enable(struct intel_dp *intel_dp,
|
|
const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
|
|
|
|
if (!crtc_state->has_drrs)
|
|
return;
|
|
|
|
drm_dbg_kms(&dev_priv->drm, "Enabling DRRS\n");
|
|
|
|
mutex_lock(&dev_priv->drrs.mutex);
|
|
|
|
if (dev_priv->drrs.dp) {
|
|
drm_warn(&dev_priv->drm, "DRRS already enabled\n");
|
|
goto unlock;
|
|
}
|
|
|
|
intel_drrs_enable_locked(intel_dp);
|
|
|
|
unlock:
|
|
mutex_unlock(&dev_priv->drrs.mutex);
|
|
}
|
|
|
|
static void
|
|
intel_drrs_disable_locked(struct intel_dp *intel_dp,
|
|
const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
|
|
|
|
intel_drrs_set_state(dev_priv, crtc_state, DRRS_HIGH_RR);
|
|
dev_priv->drrs.dp = NULL;
|
|
}
|
|
|
|
/**
|
|
* intel_drrs_disable - Disable DRRS
|
|
* @intel_dp: DP struct
|
|
* @old_crtc_state: Pointer to old crtc_state.
|
|
*
|
|
*/
|
|
void intel_drrs_disable(struct intel_dp *intel_dp,
|
|
const struct intel_crtc_state *old_crtc_state)
|
|
{
|
|
struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
|
|
|
|
if (!old_crtc_state->has_drrs)
|
|
return;
|
|
|
|
mutex_lock(&dev_priv->drrs.mutex);
|
|
if (!dev_priv->drrs.dp) {
|
|
mutex_unlock(&dev_priv->drrs.mutex);
|
|
return;
|
|
}
|
|
|
|
intel_drrs_disable_locked(intel_dp, old_crtc_state);
|
|
mutex_unlock(&dev_priv->drrs.mutex);
|
|
|
|
cancel_delayed_work_sync(&dev_priv->drrs.work);
|
|
}
|
|
|
|
/**
|
|
* intel_drrs_update - Update DRRS state
|
|
* @intel_dp: Intel DP
|
|
* @crtc_state: new CRTC state
|
|
*
|
|
* This function will update DRRS states, disabling or enabling DRRS when
|
|
* executing fastsets. For full modeset, intel_drrs_disable() and
|
|
* intel_drrs_enable() should be called instead.
|
|
*/
|
|
void
|
|
intel_drrs_update(struct intel_dp *intel_dp,
|
|
const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
|
|
|
|
if (dev_priv->drrs.type != SEAMLESS_DRRS_SUPPORT)
|
|
return;
|
|
|
|
mutex_lock(&dev_priv->drrs.mutex);
|
|
|
|
/* New state matches current one? */
|
|
if (crtc_state->has_drrs == !!dev_priv->drrs.dp)
|
|
goto unlock;
|
|
|
|
if (crtc_state->has_drrs)
|
|
intel_drrs_enable_locked(intel_dp);
|
|
else
|
|
intel_drrs_disable_locked(intel_dp, crtc_state);
|
|
|
|
unlock:
|
|
mutex_unlock(&dev_priv->drrs.mutex);
|
|
}
|
|
|
|
static void intel_drrs_downclock_work(struct work_struct *work)
|
|
{
|
|
struct drm_i915_private *dev_priv =
|
|
container_of(work, typeof(*dev_priv), drrs.work.work);
|
|
struct intel_dp *intel_dp;
|
|
struct drm_crtc *crtc;
|
|
|
|
mutex_lock(&dev_priv->drrs.mutex);
|
|
|
|
intel_dp = dev_priv->drrs.dp;
|
|
|
|
if (!intel_dp)
|
|
goto unlock;
|
|
|
|
/*
|
|
* The delayed work can race with an invalidate hence we need to
|
|
* recheck.
|
|
*/
|
|
|
|
if (dev_priv->drrs.busy_frontbuffer_bits)
|
|
goto unlock;
|
|
|
|
crtc = dp_to_dig_port(intel_dp)->base.base.crtc;
|
|
intel_drrs_set_state(dev_priv, to_intel_crtc(crtc)->config, DRRS_LOW_RR);
|
|
|
|
unlock:
|
|
mutex_unlock(&dev_priv->drrs.mutex);
|
|
}
|
|
|
|
static void intel_drrs_frontbuffer_update(struct drm_i915_private *dev_priv,
|
|
unsigned int frontbuffer_bits,
|
|
bool invalidate)
|
|
{
|
|
struct intel_dp *intel_dp;
|
|
struct drm_crtc *crtc;
|
|
enum pipe pipe;
|
|
|
|
if (dev_priv->drrs.type == DRRS_NOT_SUPPORTED)
|
|
return;
|
|
|
|
cancel_delayed_work(&dev_priv->drrs.work);
|
|
|
|
mutex_lock(&dev_priv->drrs.mutex);
|
|
|
|
intel_dp = dev_priv->drrs.dp;
|
|
if (!intel_dp) {
|
|
mutex_unlock(&dev_priv->drrs.mutex);
|
|
return;
|
|
}
|
|
|
|
crtc = dp_to_dig_port(intel_dp)->base.base.crtc;
|
|
pipe = to_intel_crtc(crtc)->pipe;
|
|
|
|
frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(pipe);
|
|
if (invalidate)
|
|
dev_priv->drrs.busy_frontbuffer_bits |= frontbuffer_bits;
|
|
else
|
|
dev_priv->drrs.busy_frontbuffer_bits &= ~frontbuffer_bits;
|
|
|
|
/* flush/invalidate means busy screen hence upclock */
|
|
if (frontbuffer_bits)
|
|
intel_drrs_set_state(dev_priv, to_intel_crtc(crtc)->config,
|
|
DRRS_HIGH_RR);
|
|
|
|
/*
|
|
* flush also means no more activity hence schedule downclock, if all
|
|
* other fbs are quiescent too
|
|
*/
|
|
if (!invalidate && !dev_priv->drrs.busy_frontbuffer_bits)
|
|
schedule_delayed_work(&dev_priv->drrs.work,
|
|
msecs_to_jiffies(1000));
|
|
mutex_unlock(&dev_priv->drrs.mutex);
|
|
}
|
|
|
|
/**
|
|
* intel_drrs_invalidate - Disable Idleness DRRS
|
|
* @dev_priv: i915 device
|
|
* @frontbuffer_bits: frontbuffer plane tracking bits
|
|
*
|
|
* This function gets called everytime rendering on the given planes start.
|
|
* Hence DRRS needs to be Upclocked, i.e. (LOW_RR -> HIGH_RR).
|
|
*
|
|
* Dirty frontbuffers relevant to DRRS are tracked in busy_frontbuffer_bits.
|
|
*/
|
|
void intel_drrs_invalidate(struct drm_i915_private *dev_priv,
|
|
unsigned int frontbuffer_bits)
|
|
{
|
|
intel_drrs_frontbuffer_update(dev_priv, frontbuffer_bits, true);
|
|
}
|
|
|
|
/**
|
|
* intel_drrs_flush - Restart Idleness DRRS
|
|
* @dev_priv: i915 device
|
|
* @frontbuffer_bits: frontbuffer plane tracking bits
|
|
*
|
|
* This function gets called every time rendering on the given planes has
|
|
* completed or flip on a crtc is completed. So DRRS should be upclocked
|
|
* (LOW_RR -> HIGH_RR). And also Idleness detection should be started again,
|
|
* if no other planes are dirty.
|
|
*
|
|
* Dirty frontbuffers relevant to DRRS are tracked in busy_frontbuffer_bits.
|
|
*/
|
|
void intel_drrs_flush(struct drm_i915_private *dev_priv,
|
|
unsigned int frontbuffer_bits)
|
|
{
|
|
intel_drrs_frontbuffer_update(dev_priv, frontbuffer_bits, false);
|
|
}
|
|
|
|
void intel_drrs_page_flip(struct intel_atomic_state *state,
|
|
struct intel_crtc *crtc)
|
|
{
|
|
struct drm_i915_private *dev_priv = to_i915(state->base.dev);
|
|
unsigned int frontbuffer_bits = INTEL_FRONTBUFFER_ALL_MASK(crtc->pipe);
|
|
|
|
intel_drrs_frontbuffer_update(dev_priv, frontbuffer_bits, false);
|
|
}
|
|
|
|
/**
|
|
* intel_drrs_init - Init basic DRRS work and mutex.
|
|
* @connector: eDP connector
|
|
* @fixed_mode: preferred mode of panel
|
|
*
|
|
* This function is called only once at driver load to initialize basic
|
|
* DRRS stuff.
|
|
*
|
|
* Returns:
|
|
* Downclock mode if panel supports it, else return NULL.
|
|
* DRRS support is determined by the presence of downclock mode (apart
|
|
* from VBT setting).
|
|
*/
|
|
struct drm_display_mode *
|
|
intel_drrs_init(struct intel_connector *connector,
|
|
struct drm_display_mode *fixed_mode)
|
|
{
|
|
struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
|
|
struct intel_encoder *encoder = connector->encoder;
|
|
struct drm_display_mode *downclock_mode = NULL;
|
|
|
|
INIT_DELAYED_WORK(&dev_priv->drrs.work, intel_drrs_downclock_work);
|
|
mutex_init(&dev_priv->drrs.mutex);
|
|
|
|
if (DISPLAY_VER(dev_priv) <= 6) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"DRRS supported for Gen7 and above\n");
|
|
return NULL;
|
|
}
|
|
|
|
if ((DISPLAY_VER(dev_priv) < 8 && !HAS_GMCH(dev_priv)) &&
|
|
encoder->port != PORT_A) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"DRRS only supported on eDP port A\n");
|
|
return NULL;
|
|
}
|
|
|
|
if (dev_priv->vbt.drrs_type != SEAMLESS_DRRS_SUPPORT) {
|
|
drm_dbg_kms(&dev_priv->drm, "VBT doesn't support DRRS\n");
|
|
return NULL;
|
|
}
|
|
|
|
downclock_mode = intel_panel_edid_downclock_mode(connector, fixed_mode);
|
|
if (!downclock_mode) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"Downclock mode is not found. DRRS not supported\n");
|
|
return NULL;
|
|
}
|
|
|
|
dev_priv->drrs.type = dev_priv->vbt.drrs_type;
|
|
|
|
dev_priv->drrs.refresh_rate_type = DRRS_HIGH_RR;
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"seamless DRRS supported for eDP panel.\n");
|
|
return downclock_mode;
|
|
}
|