drm_self_refresh_helper.c (8454B)
1// SPDX-License-Identifier: MIT 2/* 3 * Copyright (C) 2019 Google, Inc. 4 * 5 * Authors: 6 * Sean Paul <seanpaul@chromium.org> 7 */ 8#include <linux/average.h> 9#include <linux/bitops.h> 10#include <linux/slab.h> 11#include <linux/workqueue.h> 12 13#include <drm/drm_atomic.h> 14#include <drm/drm_atomic_helper.h> 15#include <drm/drm_connector.h> 16#include <drm/drm_crtc.h> 17#include <drm/drm_device.h> 18#include <drm/drm_mode_config.h> 19#include <drm/drm_modeset_lock.h> 20#include <drm/drm_print.h> 21#include <drm/drm_self_refresh_helper.h> 22 23/** 24 * DOC: overview 25 * 26 * This helper library provides an easy way for drivers to leverage the atomic 27 * framework to implement panel self refresh (SR) support. Drivers are 28 * responsible for initializing and cleaning up the SR helpers on load/unload 29 * (see &drm_self_refresh_helper_init/&drm_self_refresh_helper_cleanup). 30 * The connector is responsible for setting 31 * &drm_connector_state.self_refresh_aware to true at runtime if it is SR-aware 32 * (meaning it knows how to initiate self refresh on the panel). 33 * 34 * Once a crtc has enabled SR using &drm_self_refresh_helper_init, the 35 * helpers will monitor activity and call back into the driver to enable/disable 36 * SR as appropriate. The best way to think about this is that it's a DPMS 37 * on/off request with &drm_crtc_state.self_refresh_active set in crtc state 38 * that tells you to disable/enable SR on the panel instead of power-cycling it. 39 * 40 * During SR, drivers may choose to fully disable their crtc/encoder/bridge 41 * hardware (in which case no driver changes are necessary), or they can inspect 42 * &drm_crtc_state.self_refresh_active if they want to enter low power mode 43 * without full disable (in case full disable/enable is too slow). 44 * 45 * SR will be deactivated if there are any atomic updates affecting the 46 * pipe that is in SR mode. If a crtc is driving multiple connectors, all 47 * connectors must be SR aware and all will enter/exit SR mode at the same time. 48 * 49 * If the crtc and connector are SR aware, but the panel connected does not 50 * support it (or is otherwise unable to enter SR), the driver should fail 51 * atomic_check when &drm_crtc_state.self_refresh_active is true. 52 */ 53 54#define SELF_REFRESH_AVG_SEED_MS 200 55 56DECLARE_EWMA(psr_time, 4, 4) 57 58struct drm_self_refresh_data { 59 struct drm_crtc *crtc; 60 struct delayed_work entry_work; 61 62 struct mutex avg_mutex; 63 struct ewma_psr_time entry_avg_ms; 64 struct ewma_psr_time exit_avg_ms; 65}; 66 67static void drm_self_refresh_helper_entry_work(struct work_struct *work) 68{ 69 struct drm_self_refresh_data *sr_data = container_of( 70 to_delayed_work(work), 71 struct drm_self_refresh_data, entry_work); 72 struct drm_crtc *crtc = sr_data->crtc; 73 struct drm_device *dev = crtc->dev; 74 struct drm_modeset_acquire_ctx ctx; 75 struct drm_atomic_state *state; 76 struct drm_connector *conn; 77 struct drm_connector_state *conn_state; 78 struct drm_crtc_state *crtc_state; 79 int i, ret = 0; 80 81 drm_modeset_acquire_init(&ctx, 0); 82 83 state = drm_atomic_state_alloc(dev); 84 if (!state) { 85 ret = -ENOMEM; 86 goto out_drop_locks; 87 } 88 89retry: 90 state->acquire_ctx = &ctx; 91 92 crtc_state = drm_atomic_get_crtc_state(state, crtc); 93 if (IS_ERR(crtc_state)) { 94 ret = PTR_ERR(crtc_state); 95 goto out; 96 } 97 98 if (!crtc_state->enable) 99 goto out; 100 101 ret = drm_atomic_add_affected_connectors(state, crtc); 102 if (ret) 103 goto out; 104 105 for_each_new_connector_in_state(state, conn, conn_state, i) { 106 if (!conn_state->self_refresh_aware) 107 goto out; 108 } 109 110 crtc_state->active = false; 111 crtc_state->self_refresh_active = true; 112 113 ret = drm_atomic_commit(state); 114 if (ret) 115 goto out; 116 117out: 118 if (ret == -EDEADLK) { 119 drm_atomic_state_clear(state); 120 ret = drm_modeset_backoff(&ctx); 121 if (!ret) 122 goto retry; 123 } 124 125 drm_atomic_state_put(state); 126 127out_drop_locks: 128 drm_modeset_drop_locks(&ctx); 129 drm_modeset_acquire_fini(&ctx); 130} 131 132/** 133 * drm_self_refresh_helper_update_avg_times - Updates a crtc's SR time averages 134 * @state: the state which has just been applied to hardware 135 * @commit_time_ms: the amount of time in ms that this commit took to complete 136 * @new_self_refresh_mask: bitmask of crtc's that have self_refresh_active in 137 * new state 138 * 139 * Called after &drm_mode_config_funcs.atomic_commit_tail, this function will 140 * update the average entry/exit self refresh times on self refresh transitions. 141 * These averages will be used when calculating how long to delay before 142 * entering self refresh mode after activity. 143 */ 144void 145drm_self_refresh_helper_update_avg_times(struct drm_atomic_state *state, 146 unsigned int commit_time_ms, 147 unsigned int new_self_refresh_mask) 148{ 149 struct drm_crtc *crtc; 150 struct drm_crtc_state *old_crtc_state; 151 int i; 152 153 for_each_old_crtc_in_state(state, crtc, old_crtc_state, i) { 154 bool new_self_refresh_active = new_self_refresh_mask & BIT(i); 155 struct drm_self_refresh_data *sr_data = crtc->self_refresh_data; 156 struct ewma_psr_time *time; 157 158 if (old_crtc_state->self_refresh_active == 159 new_self_refresh_active) 160 continue; 161 162 if (new_self_refresh_active) 163 time = &sr_data->entry_avg_ms; 164 else 165 time = &sr_data->exit_avg_ms; 166 167 mutex_lock(&sr_data->avg_mutex); 168 ewma_psr_time_add(time, commit_time_ms); 169 mutex_unlock(&sr_data->avg_mutex); 170 } 171} 172EXPORT_SYMBOL(drm_self_refresh_helper_update_avg_times); 173 174/** 175 * drm_self_refresh_helper_alter_state - Alters the atomic state for SR exit 176 * @state: the state currently being checked 177 * 178 * Called at the end of atomic check. This function checks the state for flags 179 * incompatible with self refresh exit and changes them. This is a bit 180 * disingenuous since userspace is expecting one thing and we're giving it 181 * another. However in order to keep self refresh entirely hidden from 182 * userspace, this is required. 183 * 184 * At the end, we queue up the self refresh entry work so we can enter PSR after 185 * the desired delay. 186 */ 187void drm_self_refresh_helper_alter_state(struct drm_atomic_state *state) 188{ 189 struct drm_crtc *crtc; 190 struct drm_crtc_state *crtc_state; 191 int i; 192 193 if (state->async_update || !state->allow_modeset) { 194 for_each_old_crtc_in_state(state, crtc, crtc_state, i) { 195 if (crtc_state->self_refresh_active) { 196 state->async_update = false; 197 state->allow_modeset = true; 198 break; 199 } 200 } 201 } 202 203 for_each_new_crtc_in_state(state, crtc, crtc_state, i) { 204 struct drm_self_refresh_data *sr_data; 205 unsigned int delay; 206 207 /* Don't trigger the entry timer when we're already in SR */ 208 if (crtc_state->self_refresh_active) 209 continue; 210 211 sr_data = crtc->self_refresh_data; 212 if (!sr_data) 213 continue; 214 215 mutex_lock(&sr_data->avg_mutex); 216 delay = (ewma_psr_time_read(&sr_data->entry_avg_ms) + 217 ewma_psr_time_read(&sr_data->exit_avg_ms)) * 2; 218 mutex_unlock(&sr_data->avg_mutex); 219 220 mod_delayed_work(system_wq, &sr_data->entry_work, 221 msecs_to_jiffies(delay)); 222 } 223} 224EXPORT_SYMBOL(drm_self_refresh_helper_alter_state); 225 226/** 227 * drm_self_refresh_helper_init - Initializes self refresh helpers for a crtc 228 * @crtc: the crtc which supports self refresh supported displays 229 * 230 * Returns zero if successful or -errno on failure 231 */ 232int drm_self_refresh_helper_init(struct drm_crtc *crtc) 233{ 234 struct drm_self_refresh_data *sr_data = crtc->self_refresh_data; 235 236 /* Helper is already initialized */ 237 if (WARN_ON(sr_data)) 238 return -EINVAL; 239 240 sr_data = kzalloc(sizeof(*sr_data), GFP_KERNEL); 241 if (!sr_data) 242 return -ENOMEM; 243 244 INIT_DELAYED_WORK(&sr_data->entry_work, 245 drm_self_refresh_helper_entry_work); 246 sr_data->crtc = crtc; 247 mutex_init(&sr_data->avg_mutex); 248 ewma_psr_time_init(&sr_data->entry_avg_ms); 249 ewma_psr_time_init(&sr_data->exit_avg_ms); 250 251 /* 252 * Seed the averages so they're non-zero (and sufficiently large 253 * for even poorly performing panels). As time goes on, this will be 254 * averaged out and the values will trend to their true value. 255 */ 256 ewma_psr_time_add(&sr_data->entry_avg_ms, SELF_REFRESH_AVG_SEED_MS); 257 ewma_psr_time_add(&sr_data->exit_avg_ms, SELF_REFRESH_AVG_SEED_MS); 258 259 crtc->self_refresh_data = sr_data; 260 return 0; 261} 262EXPORT_SYMBOL(drm_self_refresh_helper_init); 263 264/** 265 * drm_self_refresh_helper_cleanup - Cleans up self refresh helpers for a crtc 266 * @crtc: the crtc to cleanup 267 */ 268void drm_self_refresh_helper_cleanup(struct drm_crtc *crtc) 269{ 270 struct drm_self_refresh_data *sr_data = crtc->self_refresh_data; 271 272 /* Helper is already uninitialized */ 273 if (!sr_data) 274 return; 275 276 crtc->self_refresh_data = NULL; 277 278 cancel_delayed_work_sync(&sr_data->entry_work); 279 kfree(sr_data); 280} 281EXPORT_SYMBOL(drm_self_refresh_helper_cleanup);