1 /*
2 * Copyright (C) 2014 Red Hat
3 * Copyright (C) 2014 Intel Corp.
4 *
5 * Permission is hereby granted, free of charge, to any person obtaining a
6 * copy of this software and associated documentation files (the "Software"),
7 * to deal in the Software without restriction, including without limitation
8 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
9 * and/or sell copies of the Software, and to permit persons to whom the
10 * Software is furnished to do so, subject to the following conditions:
11 *
12 * The above copyright notice and this permission notice shall be included in
13 * all copies or substantial portions of the Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
19 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
20 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
21 * OTHER DEALINGS IN THE SOFTWARE.
22 *
23 * Authors:
24 * Rob Clark <robdclark@gmail.com>
25 * Daniel Vetter <daniel.vetter@ffwll.ch>
26 */
27
28 #include <linux/dma-fence.h>
29 #include <linux/ktime.h>
30
31 #include <drm/drm_atomic.h>
32 #include <drm/drm_atomic_helper.h>
33 #include <drm/drm_atomic_uapi.h>
34 #include <drm/drm_blend.h>
35 #include <drm/drm_bridge.h>
36 #include <drm/drm_damage_helper.h>
37 #include <drm/drm_device.h>
38 #include <drm/drm_drv.h>
39 #include <drm/drm_framebuffer.h>
40 #include <drm/drm_gem_atomic_helper.h>
41 #include <drm/drm_print.h>
42 #include <drm/drm_self_refresh_helper.h>
43 #include <drm/drm_vblank.h>
44 #include <drm/drm_writeback.h>
45
46 #include "drm_crtc_helper_internal.h"
47 #include "drm_crtc_internal.h"
48
49 /**
50 * DOC: overview
51 *
52 * This helper library provides implementations of check and commit functions on
53 * top of the CRTC modeset helper callbacks and the plane helper callbacks. It
54 * also provides convenience implementations for the atomic state handling
55 * callbacks for drivers which don't need to subclass the drm core structures to
56 * add their own additional internal state.
57 *
58 * This library also provides default implementations for the check callback in
59 * drm_atomic_helper_check() and for the commit callback with
60 * drm_atomic_helper_commit(). But the individual stages and callbacks are
61 * exposed to allow drivers to mix and match and e.g. use the plane helpers only
62 * together with a driver private modeset implementation.
63 *
64 * This library also provides implementations for all the legacy driver
65 * interfaces on top of the atomic interface. See drm_atomic_helper_set_config(),
66 * drm_atomic_helper_disable_plane(), and the various functions to implement
67 * set_property callbacks. New drivers must not implement these functions
68 * themselves but must use the provided helpers.
69 *
70 * The atomic helper uses the same function table structures as all other
71 * modesetting helpers. See the documentation for &struct drm_crtc_helper_funcs,
72 * struct &drm_encoder_helper_funcs and &struct drm_connector_helper_funcs. It
73 * also shares the &struct drm_plane_helper_funcs function table with the plane
74 * helpers.
75 */
76 static void
drm_atomic_helper_plane_changed(struct drm_atomic_state * state,struct drm_plane_state * old_plane_state,struct drm_plane_state * plane_state,struct drm_plane * plane)77 drm_atomic_helper_plane_changed(struct drm_atomic_state *state,
78 struct drm_plane_state *old_plane_state,
79 struct drm_plane_state *plane_state,
80 struct drm_plane *plane)
81 {
82 struct drm_crtc_state *crtc_state;
83
84 if (old_plane_state->crtc) {
85 crtc_state = drm_atomic_get_new_crtc_state(state,
86 old_plane_state->crtc);
87
88 if (WARN_ON(!crtc_state))
89 return;
90
91 crtc_state->planes_changed = true;
92 }
93
94 if (plane_state->crtc) {
95 crtc_state = drm_atomic_get_new_crtc_state(state, plane_state->crtc);
96
97 if (WARN_ON(!crtc_state))
98 return;
99
100 crtc_state->planes_changed = true;
101 }
102 }
103
handle_conflicting_encoders(struct drm_atomic_state * state,bool disable_conflicting_encoders)104 static int handle_conflicting_encoders(struct drm_atomic_state *state,
105 bool disable_conflicting_encoders)
106 {
107 struct drm_connector_state *new_conn_state;
108 struct drm_connector *connector;
109 struct drm_connector_list_iter conn_iter;
110 struct drm_encoder *encoder;
111 unsigned int encoder_mask = 0;
112 int i, ret = 0;
113
114 /*
115 * First loop, find all newly assigned encoders from the connectors
116 * part of the state. If the same encoder is assigned to multiple
117 * connectors bail out.
118 */
119 for_each_new_connector_in_state(state, connector, new_conn_state, i) {
120 const struct drm_connector_helper_funcs *funcs = connector->helper_private;
121 struct drm_encoder *new_encoder;
122
123 if (!new_conn_state->crtc)
124 continue;
125
126 if (funcs->atomic_best_encoder)
127 new_encoder = funcs->atomic_best_encoder(connector,
128 state);
129 else if (funcs->best_encoder)
130 new_encoder = funcs->best_encoder(connector);
131 else
132 new_encoder = drm_connector_get_single_encoder(connector);
133
134 if (new_encoder) {
135 if (encoder_mask & drm_encoder_mask(new_encoder)) {
136 drm_dbg_atomic(connector->dev,
137 "[ENCODER:%d:%s] on [CONNECTOR:%d:%s] already assigned\n",
138 new_encoder->base.id, new_encoder->name,
139 connector->base.id, connector->name);
140
141 return -EINVAL;
142 }
143
144 encoder_mask |= drm_encoder_mask(new_encoder);
145 }
146 }
147
148 if (!encoder_mask)
149 return 0;
150
151 /*
152 * Second loop, iterate over all connectors not part of the state.
153 *
154 * If a conflicting encoder is found and disable_conflicting_encoders
155 * is not set, an error is returned. Userspace can provide a solution
156 * through the atomic ioctl.
157 *
158 * If the flag is set conflicting connectors are removed from the CRTC
159 * and the CRTC is disabled if no encoder is left. This preserves
160 * compatibility with the legacy set_config behavior.
161 */
162 drm_connector_list_iter_begin(state->dev, &conn_iter);
163 drm_for_each_connector_iter(connector, &conn_iter) {
164 struct drm_crtc_state *crtc_state;
165
166 if (drm_atomic_get_new_connector_state(state, connector))
167 continue;
168
169 encoder = connector->state->best_encoder;
170 if (!encoder || !(encoder_mask & drm_encoder_mask(encoder)))
171 continue;
172
173 if (!disable_conflicting_encoders) {
174 drm_dbg_atomic(connector->dev,
175 "[ENCODER:%d:%s] in use on [CRTC:%d:%s] by [CONNECTOR:%d:%s]\n",
176 encoder->base.id, encoder->name,
177 connector->state->crtc->base.id,
178 connector->state->crtc->name,
179 connector->base.id, connector->name);
180 ret = -EINVAL;
181 goto out;
182 }
183
184 new_conn_state = drm_atomic_get_connector_state(state, connector);
185 if (IS_ERR(new_conn_state)) {
186 ret = PTR_ERR(new_conn_state);
187 goto out;
188 }
189
190 drm_dbg_atomic(connector->dev,
191 "[ENCODER:%d:%s] in use on [CRTC:%d:%s], disabling [CONNECTOR:%d:%s]\n",
192 encoder->base.id, encoder->name,
193 new_conn_state->crtc->base.id, new_conn_state->crtc->name,
194 connector->base.id, connector->name);
195
196 crtc_state = drm_atomic_get_new_crtc_state(state, new_conn_state->crtc);
197
198 ret = drm_atomic_set_crtc_for_connector(new_conn_state, NULL);
199 if (ret)
200 goto out;
201
202 if (!crtc_state->connector_mask) {
203 ret = drm_atomic_set_mode_prop_for_crtc(crtc_state,
204 NULL);
205 if (ret < 0)
206 goto out;
207
208 crtc_state->active = false;
209 }
210 }
211 out:
212 drm_connector_list_iter_end(&conn_iter);
213
214 return ret;
215 }
216
217 static void
set_best_encoder(struct drm_atomic_state * state,struct drm_connector_state * conn_state,struct drm_encoder * encoder)218 set_best_encoder(struct drm_atomic_state *state,
219 struct drm_connector_state *conn_state,
220 struct drm_encoder *encoder)
221 {
222 struct drm_crtc_state *crtc_state;
223 struct drm_crtc *crtc;
224
225 if (conn_state->best_encoder) {
226 /* Unset the encoder_mask in the old crtc state. */
227 crtc = conn_state->connector->state->crtc;
228
229 /* A NULL crtc is an error here because we should have
230 * duplicated a NULL best_encoder when crtc was NULL.
231 * As an exception restoring duplicated atomic state
232 * during resume is allowed, so don't warn when
233 * best_encoder is equal to encoder we intend to set.
234 */
235 WARN_ON(!crtc && encoder != conn_state->best_encoder);
236 if (crtc) {
237 crtc_state = drm_atomic_get_new_crtc_state(state, crtc);
238
239 crtc_state->encoder_mask &=
240 ~drm_encoder_mask(conn_state->best_encoder);
241 }
242 }
243
244 if (encoder) {
245 crtc = conn_state->crtc;
246 WARN_ON(!crtc);
247 if (crtc) {
248 crtc_state = drm_atomic_get_new_crtc_state(state, crtc);
249
250 crtc_state->encoder_mask |=
251 drm_encoder_mask(encoder);
252 }
253 }
254
255 conn_state->best_encoder = encoder;
256 }
257
258 static void
steal_encoder(struct drm_atomic_state * state,struct drm_encoder * encoder)259 steal_encoder(struct drm_atomic_state *state,
260 struct drm_encoder *encoder)
261 {
262 struct drm_crtc_state *crtc_state;
263 struct drm_connector *connector;
264 struct drm_connector_state *old_connector_state, *new_connector_state;
265 int i;
266
267 for_each_oldnew_connector_in_state(state, connector, old_connector_state, new_connector_state, i) {
268 struct drm_crtc *encoder_crtc;
269
270 if (new_connector_state->best_encoder != encoder)
271 continue;
272
273 encoder_crtc = old_connector_state->crtc;
274
275 drm_dbg_atomic(encoder->dev,
276 "[ENCODER:%d:%s] in use on [CRTC:%d:%s], stealing it\n",
277 encoder->base.id, encoder->name,
278 encoder_crtc->base.id, encoder_crtc->name);
279
280 set_best_encoder(state, new_connector_state, NULL);
281
282 crtc_state = drm_atomic_get_new_crtc_state(state, encoder_crtc);
283 crtc_state->connectors_changed = true;
284
285 return;
286 }
287 }
288
289 static int
update_connector_routing(struct drm_atomic_state * state,struct drm_connector * connector,struct drm_connector_state * old_connector_state,struct drm_connector_state * new_connector_state,bool added_by_user)290 update_connector_routing(struct drm_atomic_state *state,
291 struct drm_connector *connector,
292 struct drm_connector_state *old_connector_state,
293 struct drm_connector_state *new_connector_state,
294 bool added_by_user)
295 {
296 const struct drm_connector_helper_funcs *funcs;
297 struct drm_encoder *new_encoder;
298 struct drm_crtc_state *crtc_state;
299
300 drm_dbg_atomic(connector->dev, "Updating routing for [CONNECTOR:%d:%s]\n",
301 connector->base.id, connector->name);
302
303 if (old_connector_state->crtc != new_connector_state->crtc) {
304 if (old_connector_state->crtc) {
305 crtc_state = drm_atomic_get_new_crtc_state(state, old_connector_state->crtc);
306 crtc_state->connectors_changed = true;
307 }
308
309 if (new_connector_state->crtc) {
310 crtc_state = drm_atomic_get_new_crtc_state(state, new_connector_state->crtc);
311 crtc_state->connectors_changed = true;
312 }
313 }
314
315 if (!new_connector_state->crtc) {
316 drm_dbg_atomic(connector->dev, "Disabling [CONNECTOR:%d:%s]\n",
317 connector->base.id, connector->name);
318
319 set_best_encoder(state, new_connector_state, NULL);
320
321 return 0;
322 }
323
324 crtc_state = drm_atomic_get_new_crtc_state(state,
325 new_connector_state->crtc);
326 /*
327 * For compatibility with legacy users, we want to make sure that
328 * we allow DPMS On->Off modesets on unregistered connectors. Modesets
329 * which would result in anything else must be considered invalid, to
330 * avoid turning on new displays on dead connectors.
331 *
332 * Since the connector can be unregistered at any point during an
333 * atomic check or commit, this is racy. But that's OK: all we care
334 * about is ensuring that userspace can't do anything but shut off the
335 * display on a connector that was destroyed after it's been notified,
336 * not before.
337 *
338 * Additionally, we also want to ignore connector registration when
339 * we're trying to restore an atomic state during system resume since
340 * there's a chance the connector may have been destroyed during the
341 * process, but it's better to ignore that then cause
342 * drm_atomic_helper_resume() to fail.
343 *
344 * Last, we want to ignore connector registration when the connector
345 * was not pulled in the atomic state by user-space (ie, was pulled
346 * in by the driver, e.g. when updating a DP-MST stream).
347 */
348 if (!state->duplicated && drm_connector_is_unregistered(connector) &&
349 added_by_user && crtc_state->active) {
350 drm_dbg_atomic(connector->dev,
351 "[CONNECTOR:%d:%s] is not registered\n",
352 connector->base.id, connector->name);
353 return -EINVAL;
354 }
355
356 funcs = connector->helper_private;
357
358 if (funcs->atomic_best_encoder)
359 new_encoder = funcs->atomic_best_encoder(connector, state);
360 else if (funcs->best_encoder)
361 new_encoder = funcs->best_encoder(connector);
362 else
363 new_encoder = drm_connector_get_single_encoder(connector);
364
365 if (!new_encoder) {
366 drm_dbg_atomic(connector->dev,
367 "No suitable encoder found for [CONNECTOR:%d:%s]\n",
368 connector->base.id, connector->name);
369 return -EINVAL;
370 }
371
372 if (!drm_encoder_crtc_ok(new_encoder, new_connector_state->crtc)) {
373 drm_dbg_atomic(connector->dev,
374 "[ENCODER:%d:%s] incompatible with [CRTC:%d:%s]\n",
375 new_encoder->base.id,
376 new_encoder->name,
377 new_connector_state->crtc->base.id,
378 new_connector_state->crtc->name);
379 return -EINVAL;
380 }
381
382 if (new_encoder == new_connector_state->best_encoder) {
383 set_best_encoder(state, new_connector_state, new_encoder);
384
385 drm_dbg_atomic(connector->dev,
386 "[CONNECTOR:%d:%s] keeps [ENCODER:%d:%s], now on [CRTC:%d:%s]\n",
387 connector->base.id,
388 connector->name,
389 new_encoder->base.id,
390 new_encoder->name,
391 new_connector_state->crtc->base.id,
392 new_connector_state->crtc->name);
393
394 return 0;
395 }
396
397 steal_encoder(state, new_encoder);
398
399 set_best_encoder(state, new_connector_state, new_encoder);
400
401 crtc_state->connectors_changed = true;
402
403 drm_dbg_atomic(connector->dev,
404 "[CONNECTOR:%d:%s] using [ENCODER:%d:%s] on [CRTC:%d:%s]\n",
405 connector->base.id,
406 connector->name,
407 new_encoder->base.id,
408 new_encoder->name,
409 new_connector_state->crtc->base.id,
410 new_connector_state->crtc->name);
411
412 return 0;
413 }
414
415 static int
mode_fixup(struct drm_atomic_state * state)416 mode_fixup(struct drm_atomic_state *state)
417 {
418 struct drm_crtc *crtc;
419 struct drm_crtc_state *new_crtc_state;
420 struct drm_connector *connector;
421 struct drm_connector_state *new_conn_state;
422 int i;
423 int ret;
424
425 for_each_new_crtc_in_state(state, crtc, new_crtc_state, i) {
426 if (!new_crtc_state->mode_changed &&
427 !new_crtc_state->connectors_changed)
428 continue;
429
430 drm_mode_copy(&new_crtc_state->adjusted_mode, &new_crtc_state->mode);
431 }
432
433 for_each_new_connector_in_state(state, connector, new_conn_state, i) {
434 const struct drm_encoder_helper_funcs *funcs;
435 struct drm_encoder *encoder;
436 struct drm_bridge *bridge;
437
438 WARN_ON(!!new_conn_state->best_encoder != !!new_conn_state->crtc);
439
440 if (!new_conn_state->crtc || !new_conn_state->best_encoder)
441 continue;
442
443 new_crtc_state =
444 drm_atomic_get_new_crtc_state(state, new_conn_state->crtc);
445
446 /*
447 * Each encoder has at most one connector (since we always steal
448 * it away), so we won't call ->mode_fixup twice.
449 */
450 encoder = new_conn_state->best_encoder;
451 funcs = encoder->helper_private;
452
453 bridge = drm_bridge_chain_get_first_bridge(encoder);
454 ret = drm_atomic_bridge_chain_check(bridge,
455 new_crtc_state,
456 new_conn_state);
457 if (ret) {
458 drm_dbg_atomic(encoder->dev, "Bridge atomic check failed\n");
459 return ret;
460 }
461
462 if (funcs && funcs->atomic_check) {
463 ret = funcs->atomic_check(encoder, new_crtc_state,
464 new_conn_state);
465 if (ret) {
466 drm_dbg_atomic(encoder->dev,
467 "[ENCODER:%d:%s] check failed\n",
468 encoder->base.id, encoder->name);
469 return ret;
470 }
471 } else if (funcs && funcs->mode_fixup) {
472 ret = funcs->mode_fixup(encoder, &new_crtc_state->mode,
473 &new_crtc_state->adjusted_mode);
474 if (!ret) {
475 drm_dbg_atomic(encoder->dev,
476 "[ENCODER:%d:%s] fixup failed\n",
477 encoder->base.id, encoder->name);
478 return -EINVAL;
479 }
480 }
481 }
482
483 for_each_new_crtc_in_state(state, crtc, new_crtc_state, i) {
484 const struct drm_crtc_helper_funcs *funcs;
485
486 if (!new_crtc_state->enable)
487 continue;
488
489 if (!new_crtc_state->mode_changed &&
490 !new_crtc_state->connectors_changed)
491 continue;
492
493 funcs = crtc->helper_private;
494 if (!funcs || !funcs->mode_fixup)
495 continue;
496
497 ret = funcs->mode_fixup(crtc, &new_crtc_state->mode,
498 &new_crtc_state->adjusted_mode);
499 if (!ret) {
500 drm_dbg_atomic(crtc->dev, "[CRTC:%d:%s] fixup failed\n",
501 crtc->base.id, crtc->name);
502 return -EINVAL;
503 }
504 }
505
506 return 0;
507 }
508
mode_valid_path(struct drm_connector * connector,struct drm_encoder * encoder,struct drm_crtc * crtc,const struct drm_display_mode * mode)509 static enum drm_mode_status mode_valid_path(struct drm_connector *connector,
510 struct drm_encoder *encoder,
511 struct drm_crtc *crtc,
512 const struct drm_display_mode *mode)
513 {
514 struct drm_bridge *bridge;
515 enum drm_mode_status ret;
516
517 ret = drm_encoder_mode_valid(encoder, mode);
518 if (ret != MODE_OK) {
519 drm_dbg_atomic(encoder->dev,
520 "[ENCODER:%d:%s] mode_valid() failed\n",
521 encoder->base.id, encoder->name);
522 return ret;
523 }
524
525 bridge = drm_bridge_chain_get_first_bridge(encoder);
526 ret = drm_bridge_chain_mode_valid(bridge, &connector->display_info,
527 mode);
528 if (ret != MODE_OK) {
529 drm_dbg_atomic(encoder->dev, "[BRIDGE] mode_valid() failed\n");
530 return ret;
531 }
532
533 ret = drm_crtc_mode_valid(crtc, mode);
534 if (ret != MODE_OK) {
535 drm_dbg_atomic(encoder->dev, "[CRTC:%d:%s] mode_valid() failed\n",
536 crtc->base.id, crtc->name);
537 return ret;
538 }
539
540 return ret;
541 }
542
543 static int
mode_valid(struct drm_atomic_state * state)544 mode_valid(struct drm_atomic_state *state)
545 {
546 struct drm_connector_state *conn_state;
547 struct drm_connector *connector;
548 int i;
549
550 for_each_new_connector_in_state(state, connector, conn_state, i) {
551 struct drm_encoder *encoder = conn_state->best_encoder;
552 struct drm_crtc *crtc = conn_state->crtc;
553 struct drm_crtc_state *crtc_state;
554 enum drm_mode_status mode_status;
555 const struct drm_display_mode *mode;
556
557 if (!crtc || !encoder)
558 continue;
559
560 crtc_state = drm_atomic_get_new_crtc_state(state, crtc);
561 if (!crtc_state)
562 continue;
563 if (!crtc_state->mode_changed && !crtc_state->connectors_changed)
564 continue;
565
566 mode = &crtc_state->mode;
567
568 mode_status = mode_valid_path(connector, encoder, crtc, mode);
569 if (mode_status != MODE_OK)
570 return -EINVAL;
571 }
572
573 return 0;
574 }
575
576 /**
577 * drm_atomic_helper_check_modeset - validate state object for modeset changes
578 * @dev: DRM device
579 * @state: the driver state object
580 *
581 * Check the state object to see if the requested state is physically possible.
582 * This does all the CRTC and connector related computations for an atomic
583 * update and adds any additional connectors needed for full modesets. It calls
584 * the various per-object callbacks in the follow order:
585 *
586 * 1. &drm_connector_helper_funcs.atomic_best_encoder for determining the new encoder.
587 * 2. &drm_connector_helper_funcs.atomic_check to validate the connector state.
588 * 3. If it's determined a modeset is needed then all connectors on the affected
589 * CRTC are added and &drm_connector_helper_funcs.atomic_check is run on them.
590 * 4. &drm_encoder_helper_funcs.mode_valid, &drm_bridge_funcs.mode_valid and
591 * &drm_crtc_helper_funcs.mode_valid are called on the affected components.
592 * 5. &drm_bridge_funcs.mode_fixup is called on all encoder bridges.
593 * 6. &drm_encoder_helper_funcs.atomic_check is called to validate any encoder state.
594 * This function is only called when the encoder will be part of a configured CRTC,
595 * it must not be used for implementing connector property validation.
596 * If this function is NULL, &drm_atomic_encoder_helper_funcs.mode_fixup is called
597 * instead.
598 * 7. &drm_crtc_helper_funcs.mode_fixup is called last, to fix up the mode with CRTC constraints.
599 *
600 * &drm_crtc_state.mode_changed is set when the input mode is changed.
601 * &drm_crtc_state.connectors_changed is set when a connector is added or
602 * removed from the CRTC. &drm_crtc_state.active_changed is set when
603 * &drm_crtc_state.active changes, which is used for DPMS.
604 * &drm_crtc_state.no_vblank is set from the result of drm_dev_has_vblank().
605 * See also: drm_atomic_crtc_needs_modeset()
606 *
607 * IMPORTANT:
608 *
609 * Drivers which set &drm_crtc_state.mode_changed (e.g. in their
610 * &drm_plane_helper_funcs.atomic_check hooks if a plane update can't be done
611 * without a full modeset) _must_ call this function after that change. It is
612 * permitted to call this function multiple times for the same update, e.g.
613 * when the &drm_crtc_helper_funcs.atomic_check functions depend upon the
614 * adjusted dotclock for fifo space allocation and watermark computation.
615 *
616 * RETURNS:
617 * Zero for success or -errno
618 */
619 int
drm_atomic_helper_check_modeset(struct drm_device * dev,struct drm_atomic_state * state)620 drm_atomic_helper_check_modeset(struct drm_device *dev,
621 struct drm_atomic_state *state)
622 {
623 struct drm_crtc *crtc;
624 struct drm_crtc_state *old_crtc_state, *new_crtc_state;
625 struct drm_connector *connector;
626 struct drm_connector_state *old_connector_state, *new_connector_state;
627 int i, ret;
628 unsigned int connectors_mask = 0, user_connectors_mask = 0;
629
630 for_each_oldnew_connector_in_state(state, connector, old_connector_state, new_connector_state, i)
631 user_connectors_mask |= BIT(i);
632
633 for_each_oldnew_crtc_in_state(state, crtc, old_crtc_state, new_crtc_state, i) {
634 bool has_connectors =
635 !!new_crtc_state->connector_mask;
636
637 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
638
639 if (!drm_mode_equal(&old_crtc_state->mode, &new_crtc_state->mode)) {
640 drm_dbg_atomic(dev, "[CRTC:%d:%s] mode changed\n",
641 crtc->base.id, crtc->name);
642 new_crtc_state->mode_changed = true;
643 }
644
645 if (old_crtc_state->enable != new_crtc_state->enable) {
646 drm_dbg_atomic(dev, "[CRTC:%d:%s] enable changed\n",
647 crtc->base.id, crtc->name);
648
649 /*
650 * For clarity this assignment is done here, but
651 * enable == 0 is only true when there are no
652 * connectors and a NULL mode.
653 *
654 * The other way around is true as well. enable != 0
655 * implies that connectors are attached and a mode is set.
656 */
657 new_crtc_state->mode_changed = true;
658 new_crtc_state->connectors_changed = true;
659 }
660
661 if (old_crtc_state->active != new_crtc_state->active) {
662 drm_dbg_atomic(dev, "[CRTC:%d:%s] active changed\n",
663 crtc->base.id, crtc->name);
664 new_crtc_state->active_changed = true;
665 }
666
667 if (new_crtc_state->enable != has_connectors) {
668 drm_dbg_atomic(dev, "[CRTC:%d:%s] enabled/connectors mismatch\n",
669 crtc->base.id, crtc->name);
670
671 return -EINVAL;
672 }
673
674 if (drm_dev_has_vblank(dev))
675 new_crtc_state->no_vblank = false;
676 else
677 new_crtc_state->no_vblank = true;
678 }
679
680 ret = handle_conflicting_encoders(state, false);
681 if (ret)
682 return ret;
683
684 for_each_oldnew_connector_in_state(state, connector, old_connector_state, new_connector_state, i) {
685 const struct drm_connector_helper_funcs *funcs = connector->helper_private;
686
687 WARN_ON(!drm_modeset_is_locked(&dev->mode_config.connection_mutex));
688
689 /*
690 * This only sets crtc->connectors_changed for routing changes,
691 * drivers must set crtc->connectors_changed themselves when
692 * connector properties need to be updated.
693 */
694 ret = update_connector_routing(state, connector,
695 old_connector_state,
696 new_connector_state,
697 BIT(i) & user_connectors_mask);
698 if (ret)
699 return ret;
700 if (old_connector_state->crtc) {
701 new_crtc_state = drm_atomic_get_new_crtc_state(state,
702 old_connector_state->crtc);
703 if (old_connector_state->link_status !=
704 new_connector_state->link_status)
705 new_crtc_state->connectors_changed = true;
706
707 if (old_connector_state->max_requested_bpc !=
708 new_connector_state->max_requested_bpc)
709 new_crtc_state->connectors_changed = true;
710 }
711
712 if (funcs->atomic_check)
713 ret = funcs->atomic_check(connector, state);
714 if (ret) {
715 drm_dbg_atomic(dev,
716 "[CONNECTOR:%d:%s] driver check failed\n",
717 connector->base.id, connector->name);
718 return ret;
719 }
720
721 connectors_mask |= BIT(i);
722 }
723
724 /*
725 * After all the routing has been prepared we need to add in any
726 * connector which is itself unchanged, but whose CRTC changes its
727 * configuration. This must be done before calling mode_fixup in case a
728 * crtc only changed its mode but has the same set of connectors.
729 */
730 for_each_oldnew_crtc_in_state(state, crtc, old_crtc_state, new_crtc_state, i) {
731 if (!drm_atomic_crtc_needs_modeset(new_crtc_state))
732 continue;
733
734 drm_dbg_atomic(dev,
735 "[CRTC:%d:%s] needs all connectors, enable: %c, active: %c\n",
736 crtc->base.id, crtc->name,
737 new_crtc_state->enable ? 'y' : 'n',
738 new_crtc_state->active ? 'y' : 'n');
739
740 ret = drm_atomic_add_affected_connectors(state, crtc);
741 if (ret != 0)
742 return ret;
743
744 ret = drm_atomic_add_affected_planes(state, crtc);
745 if (ret != 0)
746 return ret;
747 }
748
749 /*
750 * Iterate over all connectors again, to make sure atomic_check()
751 * has been called on them when a modeset is forced.
752 */
753 for_each_oldnew_connector_in_state(state, connector, old_connector_state, new_connector_state, i) {
754 const struct drm_connector_helper_funcs *funcs = connector->helper_private;
755
756 if (connectors_mask & BIT(i))
757 continue;
758
759 if (funcs->atomic_check)
760 ret = funcs->atomic_check(connector, state);
761 if (ret) {
762 drm_dbg_atomic(dev,
763 "[CONNECTOR:%d:%s] driver check failed\n",
764 connector->base.id, connector->name);
765 return ret;
766 }
767 }
768
769 /*
770 * Iterate over all connectors again, and add all affected bridges to
771 * the state.
772 */
773 for_each_oldnew_connector_in_state(state, connector,
774 old_connector_state,
775 new_connector_state, i) {
776 struct drm_encoder *encoder;
777
778 encoder = old_connector_state->best_encoder;
779 ret = drm_atomic_add_encoder_bridges(state, encoder);
780 if (ret)
781 return ret;
782
783 encoder = new_connector_state->best_encoder;
784 ret = drm_atomic_add_encoder_bridges(state, encoder);
785 if (ret)
786 return ret;
787 }
788
789 ret = mode_valid(state);
790 if (ret)
791 return ret;
792
793 return mode_fixup(state);
794 }
795 EXPORT_SYMBOL(drm_atomic_helper_check_modeset);
796
797 /**
798 * drm_atomic_helper_check_wb_encoder_state() - Check writeback encoder state
799 * @encoder: encoder state to check
800 * @conn_state: connector state to check
801 *
802 * Checks if the writeback connector state is valid, and returns an error if it
803 * isn't.
804 *
805 * RETURNS:
806 * Zero for success or -errno
807 */
808 int
drm_atomic_helper_check_wb_encoder_state(struct drm_encoder * encoder,struct drm_connector_state * conn_state)809 drm_atomic_helper_check_wb_encoder_state(struct drm_encoder *encoder,
810 struct drm_connector_state *conn_state)
811 {
812 struct drm_writeback_job *wb_job = conn_state->writeback_job;
813 struct drm_property_blob *pixel_format_blob;
814 struct drm_framebuffer *fb;
815 size_t i, nformats;
816 u32 *formats;
817
818 if (!wb_job || !wb_job->fb)
819 return 0;
820
821 pixel_format_blob = wb_job->connector->pixel_formats_blob_ptr;
822 nformats = pixel_format_blob->length / sizeof(u32);
823 formats = pixel_format_blob->data;
824 fb = wb_job->fb;
825
826 for (i = 0; i < nformats; i++)
827 if (fb->format->format == formats[i])
828 return 0;
829
830 drm_dbg_kms(encoder->dev, "Invalid pixel format %p4cc\n", &fb->format->format);
831
832 return -EINVAL;
833 }
834 EXPORT_SYMBOL(drm_atomic_helper_check_wb_encoder_state);
835
836 /**
837 * drm_atomic_helper_check_plane_state() - Check plane state for validity
838 * @plane_state: plane state to check
839 * @crtc_state: CRTC state to check
840 * @min_scale: minimum @src:@dest scaling factor in 16.16 fixed point
841 * @max_scale: maximum @src:@dest scaling factor in 16.16 fixed point
842 * @can_position: is it legal to position the plane such that it
843 * doesn't cover the entire CRTC? This will generally
844 * only be false for primary planes.
845 * @can_update_disabled: can the plane be updated while the CRTC
846 * is disabled?
847 *
848 * Checks that a desired plane update is valid, and updates various
849 * bits of derived state (clipped coordinates etc.). Drivers that provide
850 * their own plane handling rather than helper-provided implementations may
851 * still wish to call this function to avoid duplication of error checking
852 * code.
853 *
854 * RETURNS:
855 * Zero if update appears valid, error code on failure
856 */
drm_atomic_helper_check_plane_state(struct drm_plane_state * plane_state,const struct drm_crtc_state * crtc_state,int min_scale,int max_scale,bool can_position,bool can_update_disabled)857 int drm_atomic_helper_check_plane_state(struct drm_plane_state *plane_state,
858 const struct drm_crtc_state *crtc_state,
859 int min_scale,
860 int max_scale,
861 bool can_position,
862 bool can_update_disabled)
863 {
864 struct drm_framebuffer *fb = plane_state->fb;
865 struct drm_rect *src = &plane_state->src;
866 struct drm_rect *dst = &plane_state->dst;
867 unsigned int rotation = plane_state->rotation;
868 struct drm_rect clip = {};
869 int hscale, vscale;
870
871 WARN_ON(plane_state->crtc && plane_state->crtc != crtc_state->crtc);
872
873 *src = drm_plane_state_src(plane_state);
874 *dst = drm_plane_state_dest(plane_state);
875
876 if (!fb) {
877 plane_state->visible = false;
878 return 0;
879 }
880
881 /* crtc should only be NULL when disabling (i.e., !fb) */
882 if (WARN_ON(!plane_state->crtc)) {
883 plane_state->visible = false;
884 return 0;
885 }
886
887 if (!crtc_state->enable && !can_update_disabled) {
888 drm_dbg_kms(plane_state->plane->dev,
889 "Cannot update plane of a disabled CRTC.\n");
890 return -EINVAL;
891 }
892
893 drm_rect_rotate(src, fb->width << 16, fb->height << 16, rotation);
894
895 /* Check scaling */
896 hscale = drm_rect_calc_hscale(src, dst, min_scale, max_scale);
897 vscale = drm_rect_calc_vscale(src, dst, min_scale, max_scale);
898 if (hscale < 0 || vscale < 0) {
899 drm_dbg_kms(plane_state->plane->dev,
900 "Invalid scaling of plane\n");
901 drm_rect_debug_print("src: ", &plane_state->src, true);
902 drm_rect_debug_print("dst: ", &plane_state->dst, false);
903 return -ERANGE;
904 }
905
906 if (crtc_state->enable)
907 drm_mode_get_hv_timing(&crtc_state->mode, &clip.x2, &clip.y2);
908
909 plane_state->visible = drm_rect_clip_scaled(src, dst, &clip);
910
911 drm_rect_rotate_inv(src, fb->width << 16, fb->height << 16, rotation);
912
913 if (!plane_state->visible)
914 /*
915 * Plane isn't visible; some drivers can handle this
916 * so we just return success here. Drivers that can't
917 * (including those that use the primary plane helper's
918 * update function) will return an error from their
919 * update_plane handler.
920 */
921 return 0;
922
923 if (!can_position && !drm_rect_equals(dst, &clip)) {
924 drm_dbg_kms(plane_state->plane->dev,
925 "Plane must cover entire CRTC\n");
926 drm_rect_debug_print("dst: ", dst, false);
927 drm_rect_debug_print("clip: ", &clip, false);
928 return -EINVAL;
929 }
930
931 return 0;
932 }
933 EXPORT_SYMBOL(drm_atomic_helper_check_plane_state);
934
935 /**
936 * drm_atomic_helper_check_crtc_state() - Check CRTC state for validity
937 * @crtc_state: CRTC state to check
938 * @can_disable_primary_planes: can the CRTC be enabled without a primary plane?
939 *
940 * Checks that a desired CRTC update is valid. Drivers that provide
941 * their own CRTC handling rather than helper-provided implementations may
942 * still wish to call this function to avoid duplication of error checking
943 * code.
944 *
945 * Note that @can_disable_primary_planes only tests if the CRTC can be
946 * enabled without a primary plane. To test if a primary plane can be updated
947 * without a CRTC, use drm_atomic_helper_check_plane_state() in the plane's
948 * atomic check.
949 *
950 * RETURNS:
951 * Zero if update appears valid, error code on failure
952 */
drm_atomic_helper_check_crtc_state(struct drm_crtc_state * crtc_state,bool can_disable_primary_planes)953 int drm_atomic_helper_check_crtc_state(struct drm_crtc_state *crtc_state,
954 bool can_disable_primary_planes)
955 {
956 struct drm_device *dev = crtc_state->crtc->dev;
957
958 if (!crtc_state->enable)
959 return 0;
960
961 /* needs at least one primary plane to be enabled */
962 if (!can_disable_primary_planes) {
963 bool has_primary_plane = false;
964 struct drm_plane *plane;
965
966 drm_for_each_plane_mask(plane, dev, crtc_state->plane_mask) {
967 if (plane->type == DRM_PLANE_TYPE_PRIMARY) {
968 has_primary_plane = true;
969 break;
970 }
971 }
972 if (!has_primary_plane) {
973 drm_dbg_kms(dev, "Cannot enable CRTC without a primary plane.\n");
974 return -EINVAL;
975 }
976 }
977
978 return 0;
979 }
980 EXPORT_SYMBOL(drm_atomic_helper_check_crtc_state);
981
982 /**
983 * drm_atomic_helper_check_planes - validate state object for planes changes
984 * @dev: DRM device
985 * @state: the driver state object
986 *
987 * Check the state object to see if the requested state is physically possible.
988 * This does all the plane update related checks using by calling into the
989 * &drm_crtc_helper_funcs.atomic_check and &drm_plane_helper_funcs.atomic_check
990 * hooks provided by the driver.
991 *
992 * It also sets &drm_crtc_state.planes_changed to indicate that a CRTC has
993 * updated planes.
994 *
995 * RETURNS:
996 * Zero for success or -errno
997 */
998 int
drm_atomic_helper_check_planes(struct drm_device * dev,struct drm_atomic_state * state)999 drm_atomic_helper_check_planes(struct drm_device *dev,
1000 struct drm_atomic_state *state)
1001 {
1002 struct drm_crtc *crtc;
1003 struct drm_crtc_state *new_crtc_state;
1004 struct drm_plane *plane;
1005 struct drm_plane_state *new_plane_state, *old_plane_state;
1006 int i, ret = 0;
1007
1008 for_each_oldnew_plane_in_state(state, plane, old_plane_state, new_plane_state, i) {
1009 const struct drm_plane_helper_funcs *funcs;
1010
1011 WARN_ON(!drm_modeset_is_locked(&plane->mutex));
1012
1013 funcs = plane->helper_private;
1014
1015 drm_atomic_helper_plane_changed(state, old_plane_state, new_plane_state, plane);
1016
1017 drm_atomic_helper_check_plane_damage(state, new_plane_state);
1018
1019 if (!funcs || !funcs->atomic_check)
1020 continue;
1021
1022 ret = funcs->atomic_check(plane, state);
1023 if (ret) {
1024 drm_dbg_atomic(plane->dev,
1025 "[PLANE:%d:%s] atomic driver check failed\n",
1026 plane->base.id, plane->name);
1027 return ret;
1028 }
1029 }
1030
1031 for_each_new_crtc_in_state(state, crtc, new_crtc_state, i) {
1032 const struct drm_crtc_helper_funcs *funcs;
1033
1034 funcs = crtc->helper_private;
1035
1036 if (!funcs || !funcs->atomic_check)
1037 continue;
1038
1039 ret = funcs->atomic_check(crtc, state);
1040 if (ret) {
1041 drm_dbg_atomic(crtc->dev,
1042 "[CRTC:%d:%s] atomic driver check failed\n",
1043 crtc->base.id, crtc->name);
1044 return ret;
1045 }
1046 }
1047
1048 return ret;
1049 }
1050 EXPORT_SYMBOL(drm_atomic_helper_check_planes);
1051
1052 /**
1053 * drm_atomic_helper_check - validate state object
1054 * @dev: DRM device
1055 * @state: the driver state object
1056 *
1057 * Check the state object to see if the requested state is physically possible.
1058 * Only CRTCs and planes have check callbacks, so for any additional (global)
1059 * checking that a driver needs it can simply wrap that around this function.
1060 * Drivers without such needs can directly use this as their
1061 * &drm_mode_config_funcs.atomic_check callback.
1062 *
1063 * This just wraps the two parts of the state checking for planes and modeset
1064 * state in the default order: First it calls drm_atomic_helper_check_modeset()
1065 * and then drm_atomic_helper_check_planes(). The assumption is that the
1066 * @drm_plane_helper_funcs.atomic_check and @drm_crtc_helper_funcs.atomic_check
1067 * functions depend upon an updated adjusted_mode.clock to e.g. properly compute
1068 * watermarks.
1069 *
1070 * Note that zpos normalization will add all enable planes to the state which
1071 * might not desired for some drivers.
1072 * For example enable/disable of a cursor plane which have fixed zpos value
1073 * would trigger all other enabled planes to be forced to the state change.
1074 *
1075 * RETURNS:
1076 * Zero for success or -errno
1077 */
drm_atomic_helper_check(struct drm_device * dev,struct drm_atomic_state * state)1078 int drm_atomic_helper_check(struct drm_device *dev,
1079 struct drm_atomic_state *state)
1080 {
1081 int ret;
1082
1083 ret = drm_atomic_helper_check_modeset(dev, state);
1084 if (ret)
1085 return ret;
1086
1087 if (dev->mode_config.normalize_zpos) {
1088 ret = drm_atomic_normalize_zpos(dev, state);
1089 if (ret)
1090 return ret;
1091 }
1092
1093 ret = drm_atomic_helper_check_planes(dev, state);
1094 if (ret)
1095 return ret;
1096
1097 if (state->legacy_cursor_update)
1098 state->async_update = !drm_atomic_helper_async_check(dev, state);
1099
1100 drm_self_refresh_helper_alter_state(state);
1101
1102 return ret;
1103 }
1104 EXPORT_SYMBOL(drm_atomic_helper_check);
1105
1106 static bool
crtc_needs_disable(struct drm_crtc_state * old_state,struct drm_crtc_state * new_state)1107 crtc_needs_disable(struct drm_crtc_state *old_state,
1108 struct drm_crtc_state *new_state)
1109 {
1110 /*
1111 * No new_state means the CRTC is off, so the only criteria is whether
1112 * it's currently active or in self refresh mode.
1113 */
1114 if (!new_state)
1115 return drm_atomic_crtc_effectively_active(old_state);
1116
1117 /*
1118 * We need to disable bridge(s) and CRTC if we're transitioning out of
1119 * self-refresh and changing CRTCs at the same time, because the
1120 * bridge tracks self-refresh status via CRTC state.
1121 */
1122 if (old_state->self_refresh_active &&
1123 old_state->crtc != new_state->crtc)
1124 return true;
1125
1126 /*
1127 * We also need to run through the crtc_funcs->disable() function if
1128 * the CRTC is currently on, if it's transitioning to self refresh
1129 * mode, or if it's in self refresh mode and needs to be fully
1130 * disabled.
1131 */
1132 return old_state->active ||
1133 (old_state->self_refresh_active && !new_state->active) ||
1134 new_state->self_refresh_active;
1135 }
1136
1137 static void
disable_outputs(struct drm_device * dev,struct drm_atomic_state * old_state)1138 disable_outputs(struct drm_device *dev, struct drm_atomic_state *old_state)
1139 {
1140 struct drm_connector *connector;
1141 struct drm_connector_state *old_conn_state, *new_conn_state;
1142 struct drm_crtc *crtc;
1143 struct drm_crtc_state *old_crtc_state, *new_crtc_state;
1144 int i;
1145
1146 for_each_oldnew_connector_in_state(old_state, connector, old_conn_state, new_conn_state, i) {
1147 const struct drm_encoder_helper_funcs *funcs;
1148 struct drm_encoder *encoder;
1149 struct drm_bridge *bridge;
1150
1151 /*
1152 * Shut down everything that's in the changeset and currently
1153 * still on. So need to check the old, saved state.
1154 */
1155 if (!old_conn_state->crtc)
1156 continue;
1157
1158 old_crtc_state = drm_atomic_get_old_crtc_state(old_state, old_conn_state->crtc);
1159
1160 if (new_conn_state->crtc)
1161 new_crtc_state = drm_atomic_get_new_crtc_state(
1162 old_state,
1163 new_conn_state->crtc);
1164 else
1165 new_crtc_state = NULL;
1166
1167 if (!crtc_needs_disable(old_crtc_state, new_crtc_state) ||
1168 !drm_atomic_crtc_needs_modeset(old_conn_state->crtc->state))
1169 continue;
1170
1171 encoder = old_conn_state->best_encoder;
1172
1173 /* We shouldn't get this far if we didn't previously have
1174 * an encoder.. but WARN_ON() rather than explode.
1175 */
1176 if (WARN_ON(!encoder))
1177 continue;
1178
1179 funcs = encoder->helper_private;
1180
1181 drm_dbg_atomic(dev, "disabling [ENCODER:%d:%s]\n",
1182 encoder->base.id, encoder->name);
1183
1184 /*
1185 * Each encoder has at most one connector (since we always steal
1186 * it away), so we won't call disable hooks twice.
1187 */
1188 bridge = drm_bridge_chain_get_first_bridge(encoder);
1189 drm_atomic_bridge_chain_disable(bridge, old_state);
1190
1191 /* Right function depends upon target state. */
1192 if (funcs) {
1193 if (funcs->atomic_disable)
1194 funcs->atomic_disable(encoder, old_state);
1195 else if (new_conn_state->crtc && funcs->prepare)
1196 funcs->prepare(encoder);
1197 else if (funcs->disable)
1198 funcs->disable(encoder);
1199 else if (funcs->dpms)
1200 funcs->dpms(encoder, DRM_MODE_DPMS_OFF);
1201 }
1202
1203 drm_atomic_bridge_chain_post_disable(bridge, old_state);
1204 }
1205
1206 for_each_oldnew_crtc_in_state(old_state, crtc, old_crtc_state, new_crtc_state, i) {
1207 const struct drm_crtc_helper_funcs *funcs;
1208 int ret;
1209
1210 /* Shut down everything that needs a full modeset. */
1211 if (!drm_atomic_crtc_needs_modeset(new_crtc_state))
1212 continue;
1213
1214 if (!crtc_needs_disable(old_crtc_state, new_crtc_state))
1215 continue;
1216
1217 funcs = crtc->helper_private;
1218
1219 drm_dbg_atomic(dev, "disabling [CRTC:%d:%s]\n",
1220 crtc->base.id, crtc->name);
1221
1222
1223 /* Right function depends upon target state. */
1224 if (new_crtc_state->enable && funcs->prepare)
1225 funcs->prepare(crtc);
1226 else if (funcs->atomic_disable)
1227 funcs->atomic_disable(crtc, old_state);
1228 else if (funcs->disable)
1229 funcs->disable(crtc);
1230 else if (funcs->dpms)
1231 funcs->dpms(crtc, DRM_MODE_DPMS_OFF);
1232
1233 if (!drm_dev_has_vblank(dev))
1234 continue;
1235
1236 ret = drm_crtc_vblank_get(crtc);
1237 /*
1238 * Self-refresh is not a true "disable"; ensure vblank remains
1239 * enabled.
1240 */
1241 if (new_crtc_state->self_refresh_active)
1242 WARN_ONCE(ret != 0,
1243 "driver disabled vblank in self-refresh\n");
1244 else
1245 WARN_ONCE(ret != -EINVAL,
1246 "driver forgot to call drm_crtc_vblank_off()\n");
1247 if (ret == 0)
1248 drm_crtc_vblank_put(crtc);
1249 }
1250 }
1251
1252 /**
1253 * drm_atomic_helper_update_legacy_modeset_state - update legacy modeset state
1254 * @dev: DRM device
1255 * @old_state: atomic state object with old state structures
1256 *
1257 * This function updates all the various legacy modeset state pointers in
1258 * connectors, encoders and CRTCs.
1259 *
1260 * Drivers can use this for building their own atomic commit if they don't have
1261 * a pure helper-based modeset implementation.
1262 *
1263 * Since these updates are not synchronized with lockings, only code paths
1264 * called from &drm_mode_config_helper_funcs.atomic_commit_tail can look at the
1265 * legacy state filled out by this helper. Defacto this means this helper and
1266 * the legacy state pointers are only really useful for transitioning an
1267 * existing driver to the atomic world.
1268 */
1269 void
drm_atomic_helper_update_legacy_modeset_state(struct drm_device * dev,struct drm_atomic_state * old_state)1270 drm_atomic_helper_update_legacy_modeset_state(struct drm_device *dev,
1271 struct drm_atomic_state *old_state)
1272 {
1273 struct drm_connector *connector;
1274 struct drm_connector_state *old_conn_state, *new_conn_state;
1275 struct drm_crtc *crtc;
1276 struct drm_crtc_state *new_crtc_state;
1277 int i;
1278
1279 /* clear out existing links and update dpms */
1280 for_each_oldnew_connector_in_state(old_state, connector, old_conn_state, new_conn_state, i) {
1281 if (connector->encoder) {
1282 WARN_ON(!connector->encoder->crtc);
1283
1284 connector->encoder->crtc = NULL;
1285 connector->encoder = NULL;
1286 }
1287
1288 crtc = new_conn_state->crtc;
1289 if ((!crtc && old_conn_state->crtc) ||
1290 (crtc && drm_atomic_crtc_needs_modeset(crtc->state))) {
1291 int mode = DRM_MODE_DPMS_OFF;
1292
1293 if (crtc && crtc->state->active)
1294 mode = DRM_MODE_DPMS_ON;
1295
1296 connector->dpms = mode;
1297 }
1298 }
1299
1300 /* set new links */
1301 for_each_new_connector_in_state(old_state, connector, new_conn_state, i) {
1302 if (!new_conn_state->crtc)
1303 continue;
1304
1305 if (WARN_ON(!new_conn_state->best_encoder))
1306 continue;
1307
1308 connector->encoder = new_conn_state->best_encoder;
1309 connector->encoder->crtc = new_conn_state->crtc;
1310 }
1311
1312 /* set legacy state in the crtc structure */
1313 for_each_new_crtc_in_state(old_state, crtc, new_crtc_state, i) {
1314 struct drm_plane *primary = crtc->primary;
1315 struct drm_plane_state *new_plane_state;
1316
1317 crtc->mode = new_crtc_state->mode;
1318 crtc->enabled = new_crtc_state->enable;
1319
1320 new_plane_state =
1321 drm_atomic_get_new_plane_state(old_state, primary);
1322
1323 if (new_plane_state && new_plane_state->crtc == crtc) {
1324 crtc->x = new_plane_state->src_x >> 16;
1325 crtc->y = new_plane_state->src_y >> 16;
1326 }
1327 }
1328 }
1329 EXPORT_SYMBOL(drm_atomic_helper_update_legacy_modeset_state);
1330
1331 /**
1332 * drm_atomic_helper_calc_timestamping_constants - update vblank timestamping constants
1333 * @state: atomic state object
1334 *
1335 * Updates the timestamping constants used for precise vblank timestamps
1336 * by calling drm_calc_timestamping_constants() for all enabled crtcs in @state.
1337 */
drm_atomic_helper_calc_timestamping_constants(struct drm_atomic_state * state)1338 void drm_atomic_helper_calc_timestamping_constants(struct drm_atomic_state *state)
1339 {
1340 struct drm_crtc_state *new_crtc_state;
1341 struct drm_crtc *crtc;
1342 int i;
1343
1344 for_each_new_crtc_in_state(state, crtc, new_crtc_state, i) {
1345 if (new_crtc_state->enable)
1346 drm_calc_timestamping_constants(crtc,
1347 &new_crtc_state->adjusted_mode);
1348 }
1349 }
1350 EXPORT_SYMBOL(drm_atomic_helper_calc_timestamping_constants);
1351
1352 static void
crtc_set_mode(struct drm_device * dev,struct drm_atomic_state * old_state)1353 crtc_set_mode(struct drm_device *dev, struct drm_atomic_state *old_state)
1354 {
1355 struct drm_crtc *crtc;
1356 struct drm_crtc_state *new_crtc_state;
1357 struct drm_connector *connector;
1358 struct drm_connector_state *new_conn_state;
1359 int i;
1360
1361 for_each_new_crtc_in_state(old_state, crtc, new_crtc_state, i) {
1362 const struct drm_crtc_helper_funcs *funcs;
1363
1364 if (!new_crtc_state->mode_changed)
1365 continue;
1366
1367 funcs = crtc->helper_private;
1368
1369 if (new_crtc_state->enable && funcs->mode_set_nofb) {
1370 drm_dbg_atomic(dev, "modeset on [CRTC:%d:%s]\n",
1371 crtc->base.id, crtc->name);
1372
1373 funcs->mode_set_nofb(crtc);
1374 }
1375 }
1376
1377 for_each_new_connector_in_state(old_state, connector, new_conn_state, i) {
1378 const struct drm_encoder_helper_funcs *funcs;
1379 struct drm_encoder *encoder;
1380 struct drm_display_mode *mode, *adjusted_mode;
1381 struct drm_bridge *bridge;
1382
1383 if (!new_conn_state->best_encoder)
1384 continue;
1385
1386 encoder = new_conn_state->best_encoder;
1387 funcs = encoder->helper_private;
1388 new_crtc_state = new_conn_state->crtc->state;
1389 mode = &new_crtc_state->mode;
1390 adjusted_mode = &new_crtc_state->adjusted_mode;
1391
1392 if (!new_crtc_state->mode_changed)
1393 continue;
1394
1395 drm_dbg_atomic(dev, "modeset on [ENCODER:%d:%s]\n",
1396 encoder->base.id, encoder->name);
1397
1398 /*
1399 * Each encoder has at most one connector (since we always steal
1400 * it away), so we won't call mode_set hooks twice.
1401 */
1402 if (funcs && funcs->atomic_mode_set) {
1403 funcs->atomic_mode_set(encoder, new_crtc_state,
1404 new_conn_state);
1405 } else if (funcs && funcs->mode_set) {
1406 funcs->mode_set(encoder, mode, adjusted_mode);
1407 }
1408
1409 bridge = drm_bridge_chain_get_first_bridge(encoder);
1410 drm_bridge_chain_mode_set(bridge, mode, adjusted_mode);
1411 }
1412 }
1413
1414 /**
1415 * drm_atomic_helper_commit_modeset_disables - modeset commit to disable outputs
1416 * @dev: DRM device
1417 * @old_state: atomic state object with old state structures
1418 *
1419 * This function shuts down all the outputs that need to be shut down and
1420 * prepares them (if required) with the new mode.
1421 *
1422 * For compatibility with legacy CRTC helpers this should be called before
1423 * drm_atomic_helper_commit_planes(), which is what the default commit function
1424 * does. But drivers with different needs can group the modeset commits together
1425 * and do the plane commits at the end. This is useful for drivers doing runtime
1426 * PM since planes updates then only happen when the CRTC is actually enabled.
1427 */
drm_atomic_helper_commit_modeset_disables(struct drm_device * dev,struct drm_atomic_state * old_state)1428 void drm_atomic_helper_commit_modeset_disables(struct drm_device *dev,
1429 struct drm_atomic_state *old_state)
1430 {
1431 disable_outputs(dev, old_state);
1432
1433 drm_atomic_helper_update_legacy_modeset_state(dev, old_state);
1434 drm_atomic_helper_calc_timestamping_constants(old_state);
1435
1436 crtc_set_mode(dev, old_state);
1437 }
1438 EXPORT_SYMBOL(drm_atomic_helper_commit_modeset_disables);
1439
drm_atomic_helper_commit_writebacks(struct drm_device * dev,struct drm_atomic_state * old_state)1440 static void drm_atomic_helper_commit_writebacks(struct drm_device *dev,
1441 struct drm_atomic_state *old_state)
1442 {
1443 struct drm_connector *connector;
1444 struct drm_connector_state *new_conn_state;
1445 int i;
1446
1447 for_each_new_connector_in_state(old_state, connector, new_conn_state, i) {
1448 const struct drm_connector_helper_funcs *funcs;
1449
1450 funcs = connector->helper_private;
1451 if (!funcs->atomic_commit)
1452 continue;
1453
1454 if (new_conn_state->writeback_job && new_conn_state->writeback_job->fb) {
1455 WARN_ON(connector->connector_type != DRM_MODE_CONNECTOR_WRITEBACK);
1456 funcs->atomic_commit(connector, old_state);
1457 }
1458 }
1459 }
1460
1461 /**
1462 * drm_atomic_helper_commit_modeset_enables - modeset commit to enable outputs
1463 * @dev: DRM device
1464 * @old_state: atomic state object with old state structures
1465 *
1466 * This function enables all the outputs with the new configuration which had to
1467 * be turned off for the update.
1468 *
1469 * For compatibility with legacy CRTC helpers this should be called after
1470 * drm_atomic_helper_commit_planes(), which is what the default commit function
1471 * does. But drivers with different needs can group the modeset commits together
1472 * and do the plane commits at the end. This is useful for drivers doing runtime
1473 * PM since planes updates then only happen when the CRTC is actually enabled.
1474 */
drm_atomic_helper_commit_modeset_enables(struct drm_device * dev,struct drm_atomic_state * old_state)1475 void drm_atomic_helper_commit_modeset_enables(struct drm_device *dev,
1476 struct drm_atomic_state *old_state)
1477 {
1478 struct drm_crtc *crtc;
1479 struct drm_crtc_state *old_crtc_state;
1480 struct drm_crtc_state *new_crtc_state;
1481 struct drm_connector *connector;
1482 struct drm_connector_state *new_conn_state;
1483 int i;
1484
1485 for_each_oldnew_crtc_in_state(old_state, crtc, old_crtc_state, new_crtc_state, i) {
1486 const struct drm_crtc_helper_funcs *funcs;
1487
1488 /* Need to filter out CRTCs where only planes change. */
1489 if (!drm_atomic_crtc_needs_modeset(new_crtc_state))
1490 continue;
1491
1492 if (!new_crtc_state->active)
1493 continue;
1494
1495 funcs = crtc->helper_private;
1496
1497 if (new_crtc_state->enable) {
1498 drm_dbg_atomic(dev, "enabling [CRTC:%d:%s]\n",
1499 crtc->base.id, crtc->name);
1500 if (funcs->atomic_enable)
1501 funcs->atomic_enable(crtc, old_state);
1502 else if (funcs->commit)
1503 funcs->commit(crtc);
1504 }
1505 }
1506
1507 for_each_new_connector_in_state(old_state, connector, new_conn_state, i) {
1508 const struct drm_encoder_helper_funcs *funcs;
1509 struct drm_encoder *encoder;
1510 struct drm_bridge *bridge;
1511
1512 if (!new_conn_state->best_encoder)
1513 continue;
1514
1515 if (!new_conn_state->crtc->state->active ||
1516 !drm_atomic_crtc_needs_modeset(new_conn_state->crtc->state))
1517 continue;
1518
1519 encoder = new_conn_state->best_encoder;
1520 funcs = encoder->helper_private;
1521
1522 drm_dbg_atomic(dev, "enabling [ENCODER:%d:%s]\n",
1523 encoder->base.id, encoder->name);
1524
1525 /*
1526 * Each encoder has at most one connector (since we always steal
1527 * it away), so we won't call enable hooks twice.
1528 */
1529 bridge = drm_bridge_chain_get_first_bridge(encoder);
1530 drm_atomic_bridge_chain_pre_enable(bridge, old_state);
1531
1532 if (funcs) {
1533 if (funcs->atomic_enable)
1534 funcs->atomic_enable(encoder, old_state);
1535 else if (funcs->enable)
1536 funcs->enable(encoder);
1537 else if (funcs->commit)
1538 funcs->commit(encoder);
1539 }
1540
1541 drm_atomic_bridge_chain_enable(bridge, old_state);
1542 }
1543
1544 drm_atomic_helper_commit_writebacks(dev, old_state);
1545 }
1546 EXPORT_SYMBOL(drm_atomic_helper_commit_modeset_enables);
1547
1548 /**
1549 * drm_atomic_helper_wait_for_fences - wait for fences stashed in plane state
1550 * @dev: DRM device
1551 * @state: atomic state object with old state structures
1552 * @pre_swap: If true, do an interruptible wait, and @state is the new state.
1553 * Otherwise @state is the old state.
1554 *
1555 * For implicit sync, driver should fish the exclusive fence out from the
1556 * incoming fb's and stash it in the drm_plane_state. This is called after
1557 * drm_atomic_helper_swap_state() so it uses the current plane state (and
1558 * just uses the atomic state to find the changed planes)
1559 *
1560 * Note that @pre_swap is needed since the point where we block for fences moves
1561 * around depending upon whether an atomic commit is blocking or
1562 * non-blocking. For non-blocking commit all waiting needs to happen after
1563 * drm_atomic_helper_swap_state() is called, but for blocking commits we want
1564 * to wait **before** we do anything that can't be easily rolled back. That is
1565 * before we call drm_atomic_helper_swap_state().
1566 *
1567 * Returns zero if success or < 0 if dma_fence_wait() fails.
1568 */
drm_atomic_helper_wait_for_fences(struct drm_device * dev,struct drm_atomic_state * state,bool pre_swap)1569 int drm_atomic_helper_wait_for_fences(struct drm_device *dev,
1570 struct drm_atomic_state *state,
1571 bool pre_swap)
1572 {
1573 struct drm_plane *plane;
1574 struct drm_plane_state *new_plane_state;
1575 int i, ret;
1576
1577 for_each_new_plane_in_state(state, plane, new_plane_state, i) {
1578 if (!new_plane_state->fence)
1579 continue;
1580
1581 WARN_ON(!new_plane_state->fb);
1582
1583 /*
1584 * If waiting for fences pre-swap (ie: nonblock), userspace can
1585 * still interrupt the operation. Instead of blocking until the
1586 * timer expires, make the wait interruptible.
1587 */
1588 ret = dma_fence_wait(new_plane_state->fence, pre_swap);
1589 if (ret)
1590 return ret;
1591
1592 dma_fence_put(new_plane_state->fence);
1593 new_plane_state->fence = NULL;
1594 }
1595
1596 return 0;
1597 }
1598 EXPORT_SYMBOL(drm_atomic_helper_wait_for_fences);
1599
1600 /**
1601 * drm_atomic_helper_wait_for_vblanks - wait for vblank on CRTCs
1602 * @dev: DRM device
1603 * @old_state: atomic state object with old state structures
1604 *
1605 * Helper to, after atomic commit, wait for vblanks on all affected
1606 * CRTCs (ie. before cleaning up old framebuffers using
1607 * drm_atomic_helper_cleanup_planes()). It will only wait on CRTCs where the
1608 * framebuffers have actually changed to optimize for the legacy cursor and
1609 * plane update use-case.
1610 *
1611 * Drivers using the nonblocking commit tracking support initialized by calling
1612 * drm_atomic_helper_setup_commit() should look at
1613 * drm_atomic_helper_wait_for_flip_done() as an alternative.
1614 */
1615 void
drm_atomic_helper_wait_for_vblanks(struct drm_device * dev,struct drm_atomic_state * old_state)1616 drm_atomic_helper_wait_for_vblanks(struct drm_device *dev,
1617 struct drm_atomic_state *old_state)
1618 {
1619 struct drm_crtc *crtc;
1620 struct drm_crtc_state *old_crtc_state, *new_crtc_state;
1621 int i, ret;
1622 unsigned int crtc_mask = 0;
1623
1624 /*
1625 * Legacy cursor ioctls are completely unsynced, and userspace
1626 * relies on that (by doing tons of cursor updates).
1627 */
1628 if (old_state->legacy_cursor_update)
1629 return;
1630
1631 for_each_oldnew_crtc_in_state(old_state, crtc, old_crtc_state, new_crtc_state, i) {
1632 if (!new_crtc_state->active)
1633 continue;
1634
1635 ret = drm_crtc_vblank_get(crtc);
1636 if (ret != 0)
1637 continue;
1638
1639 crtc_mask |= drm_crtc_mask(crtc);
1640 old_state->crtcs[i].last_vblank_count = drm_crtc_vblank_count(crtc);
1641 }
1642
1643 for_each_old_crtc_in_state(old_state, crtc, old_crtc_state, i) {
1644 if (!(crtc_mask & drm_crtc_mask(crtc)))
1645 continue;
1646
1647 ret = wait_event_timeout(dev->vblank[i].queue,
1648 old_state->crtcs[i].last_vblank_count !=
1649 drm_crtc_vblank_count(crtc),
1650 msecs_to_jiffies(100));
1651
1652 WARN(!ret, "[CRTC:%d:%s] vblank wait timed out\n",
1653 crtc->base.id, crtc->name);
1654
1655 drm_crtc_vblank_put(crtc);
1656 }
1657 }
1658 EXPORT_SYMBOL(drm_atomic_helper_wait_for_vblanks);
1659
1660 /**
1661 * drm_atomic_helper_wait_for_flip_done - wait for all page flips to be done
1662 * @dev: DRM device
1663 * @old_state: atomic state object with old state structures
1664 *
1665 * Helper to, after atomic commit, wait for page flips on all affected
1666 * crtcs (ie. before cleaning up old framebuffers using
1667 * drm_atomic_helper_cleanup_planes()). Compared to
1668 * drm_atomic_helper_wait_for_vblanks() this waits for the completion on all
1669 * CRTCs, assuming that cursors-only updates are signalling their completion
1670 * immediately (or using a different path).
1671 *
1672 * This requires that drivers use the nonblocking commit tracking support
1673 * initialized using drm_atomic_helper_setup_commit().
1674 */
drm_atomic_helper_wait_for_flip_done(struct drm_device * dev,struct drm_atomic_state * old_state)1675 void drm_atomic_helper_wait_for_flip_done(struct drm_device *dev,
1676 struct drm_atomic_state *old_state)
1677 {
1678 struct drm_crtc *crtc;
1679 int i;
1680
1681 for (i = 0; i < dev->mode_config.num_crtc; i++) {
1682 struct drm_crtc_commit *commit = old_state->crtcs[i].commit;
1683 int ret;
1684
1685 crtc = old_state->crtcs[i].ptr;
1686
1687 if (!crtc || !commit)
1688 continue;
1689
1690 ret = wait_for_completion_timeout(&commit->flip_done, 10 * HZ);
1691 if (ret == 0)
1692 drm_err(dev, "[CRTC:%d:%s] flip_done timed out\n",
1693 crtc->base.id, crtc->name);
1694 }
1695
1696 if (old_state->fake_commit)
1697 complete_all(&old_state->fake_commit->flip_done);
1698 }
1699 EXPORT_SYMBOL(drm_atomic_helper_wait_for_flip_done);
1700
1701 /**
1702 * drm_atomic_helper_commit_tail - commit atomic update to hardware
1703 * @old_state: atomic state object with old state structures
1704 *
1705 * This is the default implementation for the
1706 * &drm_mode_config_helper_funcs.atomic_commit_tail hook, for drivers
1707 * that do not support runtime_pm or do not need the CRTC to be
1708 * enabled to perform a commit. Otherwise, see
1709 * drm_atomic_helper_commit_tail_rpm().
1710 *
1711 * Note that the default ordering of how the various stages are called is to
1712 * match the legacy modeset helper library closest.
1713 */
drm_atomic_helper_commit_tail(struct drm_atomic_state * old_state)1714 void drm_atomic_helper_commit_tail(struct drm_atomic_state *old_state)
1715 {
1716 struct drm_device *dev = old_state->dev;
1717
1718 drm_atomic_helper_commit_modeset_disables(dev, old_state);
1719
1720 drm_atomic_helper_commit_planes(dev, old_state, 0);
1721
1722 drm_atomic_helper_commit_modeset_enables(dev, old_state);
1723
1724 drm_atomic_helper_fake_vblank(old_state);
1725
1726 drm_atomic_helper_commit_hw_done(old_state);
1727
1728 drm_atomic_helper_wait_for_vblanks(dev, old_state);
1729
1730 drm_atomic_helper_cleanup_planes(dev, old_state);
1731 }
1732 EXPORT_SYMBOL(drm_atomic_helper_commit_tail);
1733
1734 /**
1735 * drm_atomic_helper_commit_tail_rpm - commit atomic update to hardware
1736 * @old_state: new modeset state to be committed
1737 *
1738 * This is an alternative implementation for the
1739 * &drm_mode_config_helper_funcs.atomic_commit_tail hook, for drivers
1740 * that support runtime_pm or need the CRTC to be enabled to perform a
1741 * commit. Otherwise, one should use the default implementation
1742 * drm_atomic_helper_commit_tail().
1743 */
drm_atomic_helper_commit_tail_rpm(struct drm_atomic_state * old_state)1744 void drm_atomic_helper_commit_tail_rpm(struct drm_atomic_state *old_state)
1745 {
1746 struct drm_device *dev = old_state->dev;
1747
1748 drm_atomic_helper_commit_modeset_disables(dev, old_state);
1749
1750 drm_atomic_helper_commit_modeset_enables(dev, old_state);
1751
1752 drm_atomic_helper_commit_planes(dev, old_state,
1753 DRM_PLANE_COMMIT_ACTIVE_ONLY);
1754
1755 drm_atomic_helper_fake_vblank(old_state);
1756
1757 drm_atomic_helper_commit_hw_done(old_state);
1758
1759 drm_atomic_helper_wait_for_vblanks(dev, old_state);
1760
1761 drm_atomic_helper_cleanup_planes(dev, old_state);
1762 }
1763 EXPORT_SYMBOL(drm_atomic_helper_commit_tail_rpm);
1764
commit_tail(struct drm_atomic_state * old_state)1765 static void commit_tail(struct drm_atomic_state *old_state)
1766 {
1767 struct drm_device *dev = old_state->dev;
1768 const struct drm_mode_config_helper_funcs *funcs;
1769 struct drm_crtc_state *new_crtc_state;
1770 struct drm_crtc *crtc;
1771 ktime_t start;
1772 s64 commit_time_ms;
1773 unsigned int i, new_self_refresh_mask = 0;
1774
1775 funcs = dev->mode_config.helper_private;
1776
1777 /*
1778 * We're measuring the _entire_ commit, so the time will vary depending
1779 * on how many fences and objects are involved. For the purposes of self
1780 * refresh, this is desirable since it'll give us an idea of how
1781 * congested things are. This will inform our decision on how often we
1782 * should enter self refresh after idle.
1783 *
1784 * These times will be averaged out in the self refresh helpers to avoid
1785 * overreacting over one outlier frame
1786 */
1787 start = ktime_get();
1788
1789 drm_atomic_helper_wait_for_fences(dev, old_state, false);
1790
1791 drm_atomic_helper_wait_for_dependencies(old_state);
1792
1793 /*
1794 * We cannot safely access new_crtc_state after
1795 * drm_atomic_helper_commit_hw_done() so figure out which crtc's have
1796 * self-refresh active beforehand:
1797 */
1798 for_each_new_crtc_in_state(old_state, crtc, new_crtc_state, i)
1799 if (new_crtc_state->self_refresh_active)
1800 new_self_refresh_mask |= BIT(i);
1801
1802 if (funcs && funcs->atomic_commit_tail)
1803 funcs->atomic_commit_tail(old_state);
1804 else
1805 drm_atomic_helper_commit_tail(old_state);
1806
1807 commit_time_ms = ktime_ms_delta(ktime_get(), start);
1808 if (commit_time_ms > 0)
1809 drm_self_refresh_helper_update_avg_times(old_state,
1810 (unsigned long)commit_time_ms,
1811 new_self_refresh_mask);
1812
1813 drm_atomic_helper_commit_cleanup_done(old_state);
1814
1815 drm_atomic_state_put(old_state);
1816 }
1817
commit_work(struct work_struct * work)1818 static void commit_work(struct work_struct *work)
1819 {
1820 struct drm_atomic_state *state = container_of(work,
1821 struct drm_atomic_state,
1822 commit_work);
1823 commit_tail(state);
1824 }
1825
1826 /**
1827 * drm_atomic_helper_async_check - check if state can be committed asynchronously
1828 * @dev: DRM device
1829 * @state: the driver state object
1830 *
1831 * This helper will check if it is possible to commit the state asynchronously.
1832 * Async commits are not supposed to swap the states like normal sync commits
1833 * but just do in-place changes on the current state.
1834 *
1835 * It will return 0 if the commit can happen in an asynchronous fashion or error
1836 * if not. Note that error just mean it can't be committed asynchronously, if it
1837 * fails the commit should be treated like a normal synchronous commit.
1838 */
drm_atomic_helper_async_check(struct drm_device * dev,struct drm_atomic_state * state)1839 int drm_atomic_helper_async_check(struct drm_device *dev,
1840 struct drm_atomic_state *state)
1841 {
1842 struct drm_crtc *crtc;
1843 struct drm_crtc_state *crtc_state;
1844 struct drm_plane *plane = NULL;
1845 struct drm_plane_state *old_plane_state = NULL;
1846 struct drm_plane_state *new_plane_state = NULL;
1847 const struct drm_plane_helper_funcs *funcs;
1848 int i, ret, n_planes = 0;
1849
1850 for_each_new_crtc_in_state(state, crtc, crtc_state, i) {
1851 if (drm_atomic_crtc_needs_modeset(crtc_state))
1852 return -EINVAL;
1853 }
1854
1855 for_each_oldnew_plane_in_state(state, plane, old_plane_state, new_plane_state, i)
1856 n_planes++;
1857
1858 /* FIXME: we support only single plane updates for now */
1859 if (n_planes != 1) {
1860 drm_dbg_atomic(dev,
1861 "only single plane async updates are supported\n");
1862 return -EINVAL;
1863 }
1864
1865 if (!new_plane_state->crtc ||
1866 old_plane_state->crtc != new_plane_state->crtc) {
1867 drm_dbg_atomic(dev,
1868 "[PLANE:%d:%s] async update cannot change CRTC\n",
1869 plane->base.id, plane->name);
1870 return -EINVAL;
1871 }
1872
1873 funcs = plane->helper_private;
1874 if (!funcs->atomic_async_update) {
1875 drm_dbg_atomic(dev,
1876 "[PLANE:%d:%s] driver does not support async updates\n",
1877 plane->base.id, plane->name);
1878 return -EINVAL;
1879 }
1880
1881 if (new_plane_state->fence) {
1882 drm_dbg_atomic(dev,
1883 "[PLANE:%d:%s] missing fence for async update\n",
1884 plane->base.id, plane->name);
1885 return -EINVAL;
1886 }
1887
1888 /*
1889 * Don't do an async update if there is an outstanding commit modifying
1890 * the plane. This prevents our async update's changes from getting
1891 * overridden by a previous synchronous update's state.
1892 */
1893 if (old_plane_state->commit &&
1894 !try_wait_for_completion(&old_plane_state->commit->hw_done)) {
1895 drm_dbg_atomic(dev,
1896 "[PLANE:%d:%s] inflight previous commit preventing async commit\n",
1897 plane->base.id, plane->name);
1898 return -EBUSY;
1899 }
1900
1901 ret = funcs->atomic_async_check(plane, state);
1902 if (ret != 0)
1903 drm_dbg_atomic(dev,
1904 "[PLANE:%d:%s] driver async check failed\n",
1905 plane->base.id, plane->name);
1906 return ret;
1907 }
1908 EXPORT_SYMBOL(drm_atomic_helper_async_check);
1909
1910 /**
1911 * drm_atomic_helper_async_commit - commit state asynchronously
1912 * @dev: DRM device
1913 * @state: the driver state object
1914 *
1915 * This function commits a state asynchronously, i.e., not vblank
1916 * synchronized. It should be used on a state only when
1917 * drm_atomic_async_check() succeeds. Async commits are not supposed to swap
1918 * the states like normal sync commits, but just do in-place changes on the
1919 * current state.
1920 *
1921 * TODO: Implement full swap instead of doing in-place changes.
1922 */
drm_atomic_helper_async_commit(struct drm_device * dev,struct drm_atomic_state * state)1923 void drm_atomic_helper_async_commit(struct drm_device *dev,
1924 struct drm_atomic_state *state)
1925 {
1926 struct drm_plane *plane;
1927 struct drm_plane_state *plane_state;
1928 const struct drm_plane_helper_funcs *funcs;
1929 int i;
1930
1931 for_each_new_plane_in_state(state, plane, plane_state, i) {
1932 struct drm_framebuffer *new_fb = plane_state->fb;
1933 struct drm_framebuffer *old_fb = plane->state->fb;
1934
1935 funcs = plane->helper_private;
1936 funcs->atomic_async_update(plane, state);
1937
1938 /*
1939 * ->atomic_async_update() is supposed to update the
1940 * plane->state in-place, make sure at least common
1941 * properties have been properly updated.
1942 */
1943 WARN_ON_ONCE(plane->state->fb != new_fb);
1944 WARN_ON_ONCE(plane->state->crtc_x != plane_state->crtc_x);
1945 WARN_ON_ONCE(plane->state->crtc_y != plane_state->crtc_y);
1946 WARN_ON_ONCE(plane->state->src_x != plane_state->src_x);
1947 WARN_ON_ONCE(plane->state->src_y != plane_state->src_y);
1948
1949 /*
1950 * Make sure the FBs have been swapped so that cleanups in the
1951 * new_state performs a cleanup in the old FB.
1952 */
1953 WARN_ON_ONCE(plane_state->fb != old_fb);
1954 }
1955 }
1956 EXPORT_SYMBOL(drm_atomic_helper_async_commit);
1957
1958 /**
1959 * drm_atomic_helper_commit - commit validated state object
1960 * @dev: DRM device
1961 * @state: the driver state object
1962 * @nonblock: whether nonblocking behavior is requested.
1963 *
1964 * This function commits a with drm_atomic_helper_check() pre-validated state
1965 * object. This can still fail when e.g. the framebuffer reservation fails. This
1966 * function implements nonblocking commits, using
1967 * drm_atomic_helper_setup_commit() and related functions.
1968 *
1969 * Committing the actual hardware state is done through the
1970 * &drm_mode_config_helper_funcs.atomic_commit_tail callback, or its default
1971 * implementation drm_atomic_helper_commit_tail().
1972 *
1973 * RETURNS:
1974 * Zero for success or -errno.
1975 */
drm_atomic_helper_commit(struct drm_device * dev,struct drm_atomic_state * state,bool nonblock)1976 int drm_atomic_helper_commit(struct drm_device *dev,
1977 struct drm_atomic_state *state,
1978 bool nonblock)
1979 {
1980 int ret;
1981
1982 if (state->async_update) {
1983 ret = drm_atomic_helper_prepare_planes(dev, state);
1984 if (ret)
1985 return ret;
1986
1987 drm_atomic_helper_async_commit(dev, state);
1988 drm_atomic_helper_cleanup_planes(dev, state);
1989
1990 return 0;
1991 }
1992
1993 ret = drm_atomic_helper_setup_commit(state, nonblock);
1994 if (ret)
1995 return ret;
1996
1997 INIT_WORK(&state->commit_work, commit_work);
1998
1999 ret = drm_atomic_helper_prepare_planes(dev, state);
2000 if (ret)
2001 return ret;
2002
2003 if (!nonblock) {
2004 ret = drm_atomic_helper_wait_for_fences(dev, state, true);
2005 if (ret)
2006 goto err;
2007 }
2008
2009 /*
2010 * This is the point of no return - everything below never fails except
2011 * when the hw goes bonghits. Which means we can commit the new state on
2012 * the software side now.
2013 */
2014
2015 ret = drm_atomic_helper_swap_state(state, true);
2016 if (ret)
2017 goto err;
2018
2019 /*
2020 * Everything below can be run asynchronously without the need to grab
2021 * any modeset locks at all under one condition: It must be guaranteed
2022 * that the asynchronous work has either been cancelled (if the driver
2023 * supports it, which at least requires that the framebuffers get
2024 * cleaned up with drm_atomic_helper_cleanup_planes()) or completed
2025 * before the new state gets committed on the software side with
2026 * drm_atomic_helper_swap_state().
2027 *
2028 * This scheme allows new atomic state updates to be prepared and
2029 * checked in parallel to the asynchronous completion of the previous
2030 * update. Which is important since compositors need to figure out the
2031 * composition of the next frame right after having submitted the
2032 * current layout.
2033 *
2034 * NOTE: Commit work has multiple phases, first hardware commit, then
2035 * cleanup. We want them to overlap, hence need system_unbound_wq to
2036 * make sure work items don't artificially stall on each another.
2037 */
2038
2039 drm_atomic_state_get(state);
2040 if (nonblock)
2041 queue_work(system_unbound_wq, &state->commit_work);
2042 else
2043 commit_tail(state);
2044
2045 return 0;
2046
2047 err:
2048 drm_atomic_helper_cleanup_planes(dev, state);
2049 return ret;
2050 }
2051 EXPORT_SYMBOL(drm_atomic_helper_commit);
2052
2053 /**
2054 * DOC: implementing nonblocking commit
2055 *
2056 * Nonblocking atomic commits should use struct &drm_crtc_commit to sequence
2057 * different operations against each another. Locks, especially struct
2058 * &drm_modeset_lock, should not be held in worker threads or any other
2059 * asynchronous context used to commit the hardware state.
2060 *
2061 * drm_atomic_helper_commit() implements the recommended sequence for
2062 * nonblocking commits, using drm_atomic_helper_setup_commit() internally:
2063 *
2064 * 1. Run drm_atomic_helper_prepare_planes(). Since this can fail and we
2065 * need to propagate out of memory/VRAM errors to userspace, it must be called
2066 * synchronously.
2067 *
2068 * 2. Synchronize with any outstanding nonblocking commit worker threads which
2069 * might be affected by the new state update. This is handled by
2070 * drm_atomic_helper_setup_commit().
2071 *
2072 * Asynchronous workers need to have sufficient parallelism to be able to run
2073 * different atomic commits on different CRTCs in parallel. The simplest way to
2074 * achieve this is by running them on the &system_unbound_wq work queue. Note
2075 * that drivers are not required to split up atomic commits and run an
2076 * individual commit in parallel - userspace is supposed to do that if it cares.
2077 * But it might be beneficial to do that for modesets, since those necessarily
2078 * must be done as one global operation, and enabling or disabling a CRTC can
2079 * take a long time. But even that is not required.
2080 *
2081 * IMPORTANT: A &drm_atomic_state update for multiple CRTCs is sequenced
2082 * against all CRTCs therein. Therefore for atomic state updates which only flip
2083 * planes the driver must not get the struct &drm_crtc_state of unrelated CRTCs
2084 * in its atomic check code: This would prevent committing of atomic updates to
2085 * multiple CRTCs in parallel. In general, adding additional state structures
2086 * should be avoided as much as possible, because this reduces parallelism in
2087 * (nonblocking) commits, both due to locking and due to commit sequencing
2088 * requirements.
2089 *
2090 * 3. The software state is updated synchronously with
2091 * drm_atomic_helper_swap_state(). Doing this under the protection of all modeset
2092 * locks means concurrent callers never see inconsistent state. Note that commit
2093 * workers do not hold any locks; their access is only coordinated through
2094 * ordering. If workers would access state only through the pointers in the
2095 * free-standing state objects (currently not the case for any driver) then even
2096 * multiple pending commits could be in-flight at the same time.
2097 *
2098 * 4. Schedule a work item to do all subsequent steps, using the split-out
2099 * commit helpers: a) pre-plane commit b) plane commit c) post-plane commit and
2100 * then cleaning up the framebuffers after the old framebuffer is no longer
2101 * being displayed. The scheduled work should synchronize against other workers
2102 * using the &drm_crtc_commit infrastructure as needed. See
2103 * drm_atomic_helper_setup_commit() for more details.
2104 */
2105
stall_checks(struct drm_crtc * crtc,bool nonblock)2106 static int stall_checks(struct drm_crtc *crtc, bool nonblock)
2107 {
2108 struct drm_crtc_commit *commit, *stall_commit = NULL;
2109 bool completed = true;
2110 int i;
2111 long ret = 0;
2112
2113 spin_lock(&crtc->commit_lock);
2114 i = 0;
2115 list_for_each_entry(commit, &crtc->commit_list, commit_entry) {
2116 if (i == 0) {
2117 completed = try_wait_for_completion(&commit->flip_done);
2118 /*
2119 * Userspace is not allowed to get ahead of the previous
2120 * commit with nonblocking ones.
2121 */
2122 if (!completed && nonblock) {
2123 spin_unlock(&crtc->commit_lock);
2124 drm_dbg_atomic(crtc->dev,
2125 "[CRTC:%d:%s] busy with a previous commit\n",
2126 crtc->base.id, crtc->name);
2127
2128 return -EBUSY;
2129 }
2130 } else if (i == 1) {
2131 stall_commit = drm_crtc_commit_get(commit);
2132 break;
2133 }
2134
2135 i++;
2136 }
2137 spin_unlock(&crtc->commit_lock);
2138
2139 if (!stall_commit)
2140 return 0;
2141
2142 /* We don't want to let commits get ahead of cleanup work too much,
2143 * stalling on 2nd previous commit means triple-buffer won't ever stall.
2144 */
2145 ret = wait_for_completion_interruptible_timeout(&stall_commit->cleanup_done,
2146 10*HZ);
2147 if (ret == 0)
2148 drm_err(crtc->dev, "[CRTC:%d:%s] cleanup_done timed out\n",
2149 crtc->base.id, crtc->name);
2150
2151 drm_crtc_commit_put(stall_commit);
2152
2153 return ret < 0 ? ret : 0;
2154 }
2155
release_crtc_commit(struct completion * completion)2156 static void release_crtc_commit(struct completion *completion)
2157 {
2158 struct drm_crtc_commit *commit = container_of(completion,
2159 typeof(*commit),
2160 flip_done);
2161
2162 drm_crtc_commit_put(commit);
2163 }
2164
init_commit(struct drm_crtc_commit * commit,struct drm_crtc * crtc)2165 static void init_commit(struct drm_crtc_commit *commit, struct drm_crtc *crtc)
2166 {
2167 init_completion(&commit->flip_done);
2168 init_completion(&commit->hw_done);
2169 init_completion(&commit->cleanup_done);
2170 INIT_LIST_HEAD(&commit->commit_entry);
2171 kref_init(&commit->ref);
2172 commit->crtc = crtc;
2173 }
2174
2175 static struct drm_crtc_commit *
crtc_or_fake_commit(struct drm_atomic_state * state,struct drm_crtc * crtc)2176 crtc_or_fake_commit(struct drm_atomic_state *state, struct drm_crtc *crtc)
2177 {
2178 if (crtc) {
2179 struct drm_crtc_state *new_crtc_state;
2180
2181 new_crtc_state = drm_atomic_get_new_crtc_state(state, crtc);
2182
2183 return new_crtc_state->commit;
2184 }
2185
2186 if (!state->fake_commit) {
2187 state->fake_commit = kzalloc(sizeof(*state->fake_commit), GFP_KERNEL);
2188 if (!state->fake_commit)
2189 return NULL;
2190
2191 init_commit(state->fake_commit, NULL);
2192 }
2193
2194 return state->fake_commit;
2195 }
2196
2197 /**
2198 * drm_atomic_helper_setup_commit - setup possibly nonblocking commit
2199 * @state: new modeset state to be committed
2200 * @nonblock: whether nonblocking behavior is requested.
2201 *
2202 * This function prepares @state to be used by the atomic helper's support for
2203 * nonblocking commits. Drivers using the nonblocking commit infrastructure
2204 * should always call this function from their
2205 * &drm_mode_config_funcs.atomic_commit hook.
2206 *
2207 * Drivers that need to extend the commit setup to private objects can use the
2208 * &drm_mode_config_helper_funcs.atomic_commit_setup hook.
2209 *
2210 * To be able to use this support drivers need to use a few more helper
2211 * functions. drm_atomic_helper_wait_for_dependencies() must be called before
2212 * actually committing the hardware state, and for nonblocking commits this call
2213 * must be placed in the async worker. See also drm_atomic_helper_swap_state()
2214 * and its stall parameter, for when a driver's commit hooks look at the
2215 * &drm_crtc.state, &drm_plane.state or &drm_connector.state pointer directly.
2216 *
2217 * Completion of the hardware commit step must be signalled using
2218 * drm_atomic_helper_commit_hw_done(). After this step the driver is not allowed
2219 * to read or change any permanent software or hardware modeset state. The only
2220 * exception is state protected by other means than &drm_modeset_lock locks.
2221 * Only the free standing @state with pointers to the old state structures can
2222 * be inspected, e.g. to clean up old buffers using
2223 * drm_atomic_helper_cleanup_planes().
2224 *
2225 * At the very end, before cleaning up @state drivers must call
2226 * drm_atomic_helper_commit_cleanup_done().
2227 *
2228 * This is all implemented by in drm_atomic_helper_commit(), giving drivers a
2229 * complete and easy-to-use default implementation of the atomic_commit() hook.
2230 *
2231 * The tracking of asynchronously executed and still pending commits is done
2232 * using the core structure &drm_crtc_commit.
2233 *
2234 * By default there's no need to clean up resources allocated by this function
2235 * explicitly: drm_atomic_state_default_clear() will take care of that
2236 * automatically.
2237 *
2238 * Returns:
2239 *
2240 * 0 on success. -EBUSY when userspace schedules nonblocking commits too fast,
2241 * -ENOMEM on allocation failures and -EINTR when a signal is pending.
2242 */
drm_atomic_helper_setup_commit(struct drm_atomic_state * state,bool nonblock)2243 int drm_atomic_helper_setup_commit(struct drm_atomic_state *state,
2244 bool nonblock)
2245 {
2246 struct drm_crtc *crtc;
2247 struct drm_crtc_state *old_crtc_state, *new_crtc_state;
2248 struct drm_connector *conn;
2249 struct drm_connector_state *old_conn_state, *new_conn_state;
2250 struct drm_plane *plane;
2251 struct drm_plane_state *old_plane_state, *new_plane_state;
2252 struct drm_crtc_commit *commit;
2253 const struct drm_mode_config_helper_funcs *funcs;
2254 int i, ret;
2255
2256 funcs = state->dev->mode_config.helper_private;
2257
2258 for_each_oldnew_crtc_in_state(state, crtc, old_crtc_state, new_crtc_state, i) {
2259 commit = kzalloc(sizeof(*commit), GFP_KERNEL);
2260 if (!commit)
2261 return -ENOMEM;
2262
2263 init_commit(commit, crtc);
2264
2265 new_crtc_state->commit = commit;
2266
2267 ret = stall_checks(crtc, nonblock);
2268 if (ret)
2269 return ret;
2270
2271 /*
2272 * Drivers only send out events when at least either current or
2273 * new CRTC state is active. Complete right away if everything
2274 * stays off.
2275 */
2276 if (!old_crtc_state->active && !new_crtc_state->active) {
2277 complete_all(&commit->flip_done);
2278 continue;
2279 }
2280
2281 /* Legacy cursor updates are fully unsynced. */
2282 if (state->legacy_cursor_update) {
2283 complete_all(&commit->flip_done);
2284 continue;
2285 }
2286
2287 if (!new_crtc_state->event) {
2288 commit->event = kzalloc(sizeof(*commit->event),
2289 GFP_KERNEL);
2290 if (!commit->event)
2291 return -ENOMEM;
2292
2293 new_crtc_state->event = commit->event;
2294 }
2295
2296 new_crtc_state->event->base.completion = &commit->flip_done;
2297 new_crtc_state->event->base.completion_release = release_crtc_commit;
2298 drm_crtc_commit_get(commit);
2299
2300 commit->abort_completion = true;
2301
2302 state->crtcs[i].commit = commit;
2303 drm_crtc_commit_get(commit);
2304 }
2305
2306 for_each_oldnew_connector_in_state(state, conn, old_conn_state, new_conn_state, i) {
2307 /*
2308 * Userspace is not allowed to get ahead of the previous
2309 * commit with nonblocking ones.
2310 */
2311 if (nonblock && old_conn_state->commit &&
2312 !try_wait_for_completion(&old_conn_state->commit->flip_done)) {
2313 drm_dbg_atomic(conn->dev,
2314 "[CONNECTOR:%d:%s] busy with a previous commit\n",
2315 conn->base.id, conn->name);
2316
2317 return -EBUSY;
2318 }
2319
2320 /* Always track connectors explicitly for e.g. link retraining. */
2321 commit = crtc_or_fake_commit(state, new_conn_state->crtc ?: old_conn_state->crtc);
2322 if (!commit)
2323 return -ENOMEM;
2324
2325 new_conn_state->commit = drm_crtc_commit_get(commit);
2326 }
2327
2328 for_each_oldnew_plane_in_state(state, plane, old_plane_state, new_plane_state, i) {
2329 /*
2330 * Userspace is not allowed to get ahead of the previous
2331 * commit with nonblocking ones.
2332 */
2333 if (nonblock && old_plane_state->commit &&
2334 !try_wait_for_completion(&old_plane_state->commit->flip_done)) {
2335 drm_dbg_atomic(plane->dev,
2336 "[PLANE:%d:%s] busy with a previous commit\n",
2337 plane->base.id, plane->name);
2338
2339 return -EBUSY;
2340 }
2341
2342 /* Always track planes explicitly for async pageflip support. */
2343 commit = crtc_or_fake_commit(state, new_plane_state->crtc ?: old_plane_state->crtc);
2344 if (!commit)
2345 return -ENOMEM;
2346
2347 new_plane_state->commit = drm_crtc_commit_get(commit);
2348 }
2349
2350 if (funcs && funcs->atomic_commit_setup)
2351 return funcs->atomic_commit_setup(state);
2352
2353 return 0;
2354 }
2355 EXPORT_SYMBOL(drm_atomic_helper_setup_commit);
2356
2357 /**
2358 * drm_atomic_helper_wait_for_dependencies - wait for required preceeding commits
2359 * @old_state: atomic state object with old state structures
2360 *
2361 * This function waits for all preceeding commits that touch the same CRTC as
2362 * @old_state to both be committed to the hardware (as signalled by
2363 * drm_atomic_helper_commit_hw_done()) and executed by the hardware (as signalled
2364 * by calling drm_crtc_send_vblank_event() on the &drm_crtc_state.event).
2365 *
2366 * This is part of the atomic helper support for nonblocking commits, see
2367 * drm_atomic_helper_setup_commit() for an overview.
2368 */
drm_atomic_helper_wait_for_dependencies(struct drm_atomic_state * old_state)2369 void drm_atomic_helper_wait_for_dependencies(struct drm_atomic_state *old_state)
2370 {
2371 struct drm_crtc *crtc;
2372 struct drm_crtc_state *old_crtc_state;
2373 struct drm_plane *plane;
2374 struct drm_plane_state *old_plane_state;
2375 struct drm_connector *conn;
2376 struct drm_connector_state *old_conn_state;
2377 int i;
2378 long ret;
2379
2380 for_each_old_crtc_in_state(old_state, crtc, old_crtc_state, i) {
2381 ret = drm_crtc_commit_wait(old_crtc_state->commit);
2382 if (ret)
2383 drm_err(crtc->dev,
2384 "[CRTC:%d:%s] commit wait timed out\n",
2385 crtc->base.id, crtc->name);
2386 }
2387
2388 for_each_old_connector_in_state(old_state, conn, old_conn_state, i) {
2389 ret = drm_crtc_commit_wait(old_conn_state->commit);
2390 if (ret)
2391 drm_err(conn->dev,
2392 "[CONNECTOR:%d:%s] commit wait timed out\n",
2393 conn->base.id, conn->name);
2394 }
2395
2396 for_each_old_plane_in_state(old_state, plane, old_plane_state, i) {
2397 ret = drm_crtc_commit_wait(old_plane_state->commit);
2398 if (ret)
2399 drm_err(plane->dev,
2400 "[PLANE:%d:%s] commit wait timed out\n",
2401 plane->base.id, plane->name);
2402 }
2403 }
2404 EXPORT_SYMBOL(drm_atomic_helper_wait_for_dependencies);
2405
2406 /**
2407 * drm_atomic_helper_fake_vblank - fake VBLANK events if needed
2408 * @old_state: atomic state object with old state structures
2409 *
2410 * This function walks all CRTCs and fakes VBLANK events on those with
2411 * &drm_crtc_state.no_vblank set to true and &drm_crtc_state.event != NULL.
2412 * The primary use of this function is writeback connectors working in oneshot
2413 * mode and faking VBLANK events. In this case they only fake the VBLANK event
2414 * when a job is queued, and any change to the pipeline that does not touch the
2415 * connector is leading to timeouts when calling
2416 * drm_atomic_helper_wait_for_vblanks() or
2417 * drm_atomic_helper_wait_for_flip_done(). In addition to writeback
2418 * connectors, this function can also fake VBLANK events for CRTCs without
2419 * VBLANK interrupt.
2420 *
2421 * This is part of the atomic helper support for nonblocking commits, see
2422 * drm_atomic_helper_setup_commit() for an overview.
2423 */
drm_atomic_helper_fake_vblank(struct drm_atomic_state * old_state)2424 void drm_atomic_helper_fake_vblank(struct drm_atomic_state *old_state)
2425 {
2426 struct drm_crtc_state *new_crtc_state;
2427 struct drm_crtc *crtc;
2428 int i;
2429
2430 for_each_new_crtc_in_state(old_state, crtc, new_crtc_state, i) {
2431 unsigned long flags;
2432
2433 if (!new_crtc_state->no_vblank)
2434 continue;
2435
2436 spin_lock_irqsave(&old_state->dev->event_lock, flags);
2437 if (new_crtc_state->event) {
2438 drm_crtc_send_vblank_event(crtc,
2439 new_crtc_state->event);
2440 new_crtc_state->event = NULL;
2441 }
2442 spin_unlock_irqrestore(&old_state->dev->event_lock, flags);
2443 }
2444 }
2445 EXPORT_SYMBOL(drm_atomic_helper_fake_vblank);
2446
2447 /**
2448 * drm_atomic_helper_commit_hw_done - setup possible nonblocking commit
2449 * @old_state: atomic state object with old state structures
2450 *
2451 * This function is used to signal completion of the hardware commit step. After
2452 * this step the driver is not allowed to read or change any permanent software
2453 * or hardware modeset state. The only exception is state protected by other
2454 * means than &drm_modeset_lock locks.
2455 *
2456 * Drivers should try to postpone any expensive or delayed cleanup work after
2457 * this function is called.
2458 *
2459 * This is part of the atomic helper support for nonblocking commits, see
2460 * drm_atomic_helper_setup_commit() for an overview.
2461 */
drm_atomic_helper_commit_hw_done(struct drm_atomic_state * old_state)2462 void drm_atomic_helper_commit_hw_done(struct drm_atomic_state *old_state)
2463 {
2464 struct drm_crtc *crtc;
2465 struct drm_crtc_state *old_crtc_state, *new_crtc_state;
2466 struct drm_crtc_commit *commit;
2467 int i;
2468
2469 for_each_oldnew_crtc_in_state(old_state, crtc, old_crtc_state, new_crtc_state, i) {
2470 commit = new_crtc_state->commit;
2471 if (!commit)
2472 continue;
2473
2474 /*
2475 * copy new_crtc_state->commit to old_crtc_state->commit,
2476 * it's unsafe to touch new_crtc_state after hw_done,
2477 * but we still need to do so in cleanup_done().
2478 */
2479 if (old_crtc_state->commit)
2480 drm_crtc_commit_put(old_crtc_state->commit);
2481
2482 old_crtc_state->commit = drm_crtc_commit_get(commit);
2483
2484 /* backend must have consumed any event by now */
2485 WARN_ON(new_crtc_state->event);
2486 complete_all(&commit->hw_done);
2487 }
2488
2489 if (old_state->fake_commit) {
2490 complete_all(&old_state->fake_commit->hw_done);
2491 complete_all(&old_state->fake_commit->flip_done);
2492 }
2493 }
2494 EXPORT_SYMBOL(drm_atomic_helper_commit_hw_done);
2495
2496 /**
2497 * drm_atomic_helper_commit_cleanup_done - signal completion of commit
2498 * @old_state: atomic state object with old state structures
2499 *
2500 * This signals completion of the atomic update @old_state, including any
2501 * cleanup work. If used, it must be called right before calling
2502 * drm_atomic_state_put().
2503 *
2504 * This is part of the atomic helper support for nonblocking commits, see
2505 * drm_atomic_helper_setup_commit() for an overview.
2506 */
drm_atomic_helper_commit_cleanup_done(struct drm_atomic_state * old_state)2507 void drm_atomic_helper_commit_cleanup_done(struct drm_atomic_state *old_state)
2508 {
2509 struct drm_crtc *crtc;
2510 struct drm_crtc_state *old_crtc_state;
2511 struct drm_crtc_commit *commit;
2512 int i;
2513
2514 for_each_old_crtc_in_state(old_state, crtc, old_crtc_state, i) {
2515 commit = old_crtc_state->commit;
2516 if (WARN_ON(!commit))
2517 continue;
2518
2519 complete_all(&commit->cleanup_done);
2520 WARN_ON(!try_wait_for_completion(&commit->hw_done));
2521
2522 spin_lock(&crtc->commit_lock);
2523 list_del(&commit->commit_entry);
2524 spin_unlock(&crtc->commit_lock);
2525 }
2526
2527 if (old_state->fake_commit) {
2528 complete_all(&old_state->fake_commit->cleanup_done);
2529 WARN_ON(!try_wait_for_completion(&old_state->fake_commit->hw_done));
2530 }
2531 }
2532 EXPORT_SYMBOL(drm_atomic_helper_commit_cleanup_done);
2533
2534 /**
2535 * drm_atomic_helper_prepare_planes - prepare plane resources before commit
2536 * @dev: DRM device
2537 * @state: atomic state object with new state structures
2538 *
2539 * This function prepares plane state, specifically framebuffers, for the new
2540 * configuration, by calling &drm_plane_helper_funcs.prepare_fb. If any failure
2541 * is encountered this function will call &drm_plane_helper_funcs.cleanup_fb on
2542 * any already successfully prepared framebuffer.
2543 *
2544 * Returns:
2545 * 0 on success, negative error code on failure.
2546 */
drm_atomic_helper_prepare_planes(struct drm_device * dev,struct drm_atomic_state * state)2547 int drm_atomic_helper_prepare_planes(struct drm_device *dev,
2548 struct drm_atomic_state *state)
2549 {
2550 struct drm_connector *connector;
2551 struct drm_connector_state *new_conn_state;
2552 struct drm_plane *plane;
2553 struct drm_plane_state *new_plane_state;
2554 int ret, i, j;
2555
2556 for_each_new_connector_in_state(state, connector, new_conn_state, i) {
2557 if (!new_conn_state->writeback_job)
2558 continue;
2559
2560 ret = drm_writeback_prepare_job(new_conn_state->writeback_job);
2561 if (ret < 0)
2562 return ret;
2563 }
2564
2565 for_each_new_plane_in_state(state, plane, new_plane_state, i) {
2566 const struct drm_plane_helper_funcs *funcs;
2567
2568 funcs = plane->helper_private;
2569
2570 if (funcs->prepare_fb) {
2571 ret = funcs->prepare_fb(plane, new_plane_state);
2572 if (ret)
2573 goto fail;
2574 } else {
2575 WARN_ON_ONCE(funcs->cleanup_fb);
2576
2577 if (!drm_core_check_feature(dev, DRIVER_GEM))
2578 continue;
2579
2580 ret = drm_gem_plane_helper_prepare_fb(plane, new_plane_state);
2581 if (ret)
2582 goto fail;
2583 }
2584 }
2585
2586 return 0;
2587
2588 fail:
2589 for_each_new_plane_in_state(state, plane, new_plane_state, j) {
2590 const struct drm_plane_helper_funcs *funcs;
2591
2592 if (j >= i)
2593 continue;
2594
2595 funcs = plane->helper_private;
2596
2597 if (funcs->cleanup_fb)
2598 funcs->cleanup_fb(plane, new_plane_state);
2599 }
2600
2601 return ret;
2602 }
2603 EXPORT_SYMBOL(drm_atomic_helper_prepare_planes);
2604
plane_crtc_active(const struct drm_plane_state * state)2605 static bool plane_crtc_active(const struct drm_plane_state *state)
2606 {
2607 return state->crtc && state->crtc->state->active;
2608 }
2609
2610 /**
2611 * drm_atomic_helper_commit_planes - commit plane state
2612 * @dev: DRM device
2613 * @old_state: atomic state object with old state structures
2614 * @flags: flags for committing plane state
2615 *
2616 * This function commits the new plane state using the plane and atomic helper
2617 * functions for planes and CRTCs. It assumes that the atomic state has already
2618 * been pushed into the relevant object state pointers, since this step can no
2619 * longer fail.
2620 *
2621 * It still requires the global state object @old_state to know which planes and
2622 * crtcs need to be updated though.
2623 *
2624 * Note that this function does all plane updates across all CRTCs in one step.
2625 * If the hardware can't support this approach look at
2626 * drm_atomic_helper_commit_planes_on_crtc() instead.
2627 *
2628 * Plane parameters can be updated by applications while the associated CRTC is
2629 * disabled. The DRM/KMS core will store the parameters in the plane state,
2630 * which will be available to the driver when the CRTC is turned on. As a result
2631 * most drivers don't need to be immediately notified of plane updates for a
2632 * disabled CRTC.
2633 *
2634 * Unless otherwise needed, drivers are advised to set the ACTIVE_ONLY flag in
2635 * @flags in order not to receive plane update notifications related to a
2636 * disabled CRTC. This avoids the need to manually ignore plane updates in
2637 * driver code when the driver and/or hardware can't or just don't need to deal
2638 * with updates on disabled CRTCs, for example when supporting runtime PM.
2639 *
2640 * Drivers may set the NO_DISABLE_AFTER_MODESET flag in @flags if the relevant
2641 * display controllers require to disable a CRTC's planes when the CRTC is
2642 * disabled. This function would skip the &drm_plane_helper_funcs.atomic_disable
2643 * call for a plane if the CRTC of the old plane state needs a modesetting
2644 * operation. Of course, the drivers need to disable the planes in their CRTC
2645 * disable callbacks since no one else would do that.
2646 *
2647 * The drm_atomic_helper_commit() default implementation doesn't set the
2648 * ACTIVE_ONLY flag to most closely match the behaviour of the legacy helpers.
2649 * This should not be copied blindly by drivers.
2650 */
drm_atomic_helper_commit_planes(struct drm_device * dev,struct drm_atomic_state * old_state,uint32_t flags)2651 void drm_atomic_helper_commit_planes(struct drm_device *dev,
2652 struct drm_atomic_state *old_state,
2653 uint32_t flags)
2654 {
2655 struct drm_crtc *crtc;
2656 struct drm_crtc_state *old_crtc_state, *new_crtc_state;
2657 struct drm_plane *plane;
2658 struct drm_plane_state *old_plane_state, *new_plane_state;
2659 int i;
2660 bool active_only = flags & DRM_PLANE_COMMIT_ACTIVE_ONLY;
2661 bool no_disable = flags & DRM_PLANE_COMMIT_NO_DISABLE_AFTER_MODESET;
2662
2663 for_each_oldnew_crtc_in_state(old_state, crtc, old_crtc_state, new_crtc_state, i) {
2664 const struct drm_crtc_helper_funcs *funcs;
2665
2666 funcs = crtc->helper_private;
2667
2668 if (!funcs || !funcs->atomic_begin)
2669 continue;
2670
2671 if (active_only && !new_crtc_state->active)
2672 continue;
2673
2674 funcs->atomic_begin(crtc, old_state);
2675 }
2676
2677 for_each_oldnew_plane_in_state(old_state, plane, old_plane_state, new_plane_state, i) {
2678 const struct drm_plane_helper_funcs *funcs;
2679 bool disabling;
2680
2681 funcs = plane->helper_private;
2682
2683 if (!funcs)
2684 continue;
2685
2686 disabling = drm_atomic_plane_disabling(old_plane_state,
2687 new_plane_state);
2688
2689 if (active_only) {
2690 /*
2691 * Skip planes related to inactive CRTCs. If the plane
2692 * is enabled use the state of the current CRTC. If the
2693 * plane is being disabled use the state of the old
2694 * CRTC to avoid skipping planes being disabled on an
2695 * active CRTC.
2696 */
2697 if (!disabling && !plane_crtc_active(new_plane_state))
2698 continue;
2699 if (disabling && !plane_crtc_active(old_plane_state))
2700 continue;
2701 }
2702
2703 /*
2704 * Special-case disabling the plane if drivers support it.
2705 */
2706 if (disabling && funcs->atomic_disable) {
2707 struct drm_crtc_state *crtc_state;
2708
2709 crtc_state = old_plane_state->crtc->state;
2710
2711 if (drm_atomic_crtc_needs_modeset(crtc_state) &&
2712 no_disable)
2713 continue;
2714
2715 funcs->atomic_disable(plane, old_state);
2716 } else if (new_plane_state->crtc || disabling) {
2717 funcs->atomic_update(plane, old_state);
2718 }
2719 }
2720
2721 for_each_oldnew_crtc_in_state(old_state, crtc, old_crtc_state, new_crtc_state, i) {
2722 const struct drm_crtc_helper_funcs *funcs;
2723
2724 funcs = crtc->helper_private;
2725
2726 if (!funcs || !funcs->atomic_flush)
2727 continue;
2728
2729 if (active_only && !new_crtc_state->active)
2730 continue;
2731
2732 funcs->atomic_flush(crtc, old_state);
2733 }
2734 }
2735 EXPORT_SYMBOL(drm_atomic_helper_commit_planes);
2736
2737 /**
2738 * drm_atomic_helper_commit_planes_on_crtc - commit plane state for a CRTC
2739 * @old_crtc_state: atomic state object with the old CRTC state
2740 *
2741 * This function commits the new plane state using the plane and atomic helper
2742 * functions for planes on the specific CRTC. It assumes that the atomic state
2743 * has already been pushed into the relevant object state pointers, since this
2744 * step can no longer fail.
2745 *
2746 * This function is useful when plane updates should be done CRTC-by-CRTC
2747 * instead of one global step like drm_atomic_helper_commit_planes() does.
2748 *
2749 * This function can only be savely used when planes are not allowed to move
2750 * between different CRTCs because this function doesn't handle inter-CRTC
2751 * dependencies. Callers need to ensure that either no such dependencies exist,
2752 * resolve them through ordering of commit calls or through some other means.
2753 */
2754 void
drm_atomic_helper_commit_planes_on_crtc(struct drm_crtc_state * old_crtc_state)2755 drm_atomic_helper_commit_planes_on_crtc(struct drm_crtc_state *old_crtc_state)
2756 {
2757 const struct drm_crtc_helper_funcs *crtc_funcs;
2758 struct drm_crtc *crtc = old_crtc_state->crtc;
2759 struct drm_atomic_state *old_state = old_crtc_state->state;
2760 struct drm_crtc_state *new_crtc_state =
2761 drm_atomic_get_new_crtc_state(old_state, crtc);
2762 struct drm_plane *plane;
2763 unsigned int plane_mask;
2764
2765 plane_mask = old_crtc_state->plane_mask;
2766 plane_mask |= new_crtc_state->plane_mask;
2767
2768 crtc_funcs = crtc->helper_private;
2769 if (crtc_funcs && crtc_funcs->atomic_begin)
2770 crtc_funcs->atomic_begin(crtc, old_state);
2771
2772 drm_for_each_plane_mask(plane, crtc->dev, plane_mask) {
2773 struct drm_plane_state *old_plane_state =
2774 drm_atomic_get_old_plane_state(old_state, plane);
2775 struct drm_plane_state *new_plane_state =
2776 drm_atomic_get_new_plane_state(old_state, plane);
2777 const struct drm_plane_helper_funcs *plane_funcs;
2778
2779 plane_funcs = plane->helper_private;
2780
2781 if (!old_plane_state || !plane_funcs)
2782 continue;
2783
2784 WARN_ON(new_plane_state->crtc &&
2785 new_plane_state->crtc != crtc);
2786
2787 if (drm_atomic_plane_disabling(old_plane_state, new_plane_state) &&
2788 plane_funcs->atomic_disable)
2789 plane_funcs->atomic_disable(plane, old_state);
2790 else if (new_plane_state->crtc ||
2791 drm_atomic_plane_disabling(old_plane_state, new_plane_state))
2792 plane_funcs->atomic_update(plane, old_state);
2793 }
2794
2795 if (crtc_funcs && crtc_funcs->atomic_flush)
2796 crtc_funcs->atomic_flush(crtc, old_state);
2797 }
2798 EXPORT_SYMBOL(drm_atomic_helper_commit_planes_on_crtc);
2799
2800 /**
2801 * drm_atomic_helper_disable_planes_on_crtc - helper to disable CRTC's planes
2802 * @old_crtc_state: atomic state object with the old CRTC state
2803 * @atomic: if set, synchronize with CRTC's atomic_begin/flush hooks
2804 *
2805 * Disables all planes associated with the given CRTC. This can be
2806 * used for instance in the CRTC helper atomic_disable callback to disable
2807 * all planes.
2808 *
2809 * If the atomic-parameter is set the function calls the CRTC's
2810 * atomic_begin hook before and atomic_flush hook after disabling the
2811 * planes.
2812 *
2813 * It is a bug to call this function without having implemented the
2814 * &drm_plane_helper_funcs.atomic_disable plane hook.
2815 */
2816 void
drm_atomic_helper_disable_planes_on_crtc(struct drm_crtc_state * old_crtc_state,bool atomic)2817 drm_atomic_helper_disable_planes_on_crtc(struct drm_crtc_state *old_crtc_state,
2818 bool atomic)
2819 {
2820 struct drm_crtc *crtc = old_crtc_state->crtc;
2821 const struct drm_crtc_helper_funcs *crtc_funcs =
2822 crtc->helper_private;
2823 struct drm_plane *plane;
2824
2825 if (atomic && crtc_funcs && crtc_funcs->atomic_begin)
2826 crtc_funcs->atomic_begin(crtc, NULL);
2827
2828 drm_atomic_crtc_state_for_each_plane(plane, old_crtc_state) {
2829 const struct drm_plane_helper_funcs *plane_funcs =
2830 plane->helper_private;
2831
2832 if (!plane_funcs)
2833 continue;
2834
2835 WARN_ON(!plane_funcs->atomic_disable);
2836 if (plane_funcs->atomic_disable)
2837 plane_funcs->atomic_disable(plane, NULL);
2838 }
2839
2840 if (atomic && crtc_funcs && crtc_funcs->atomic_flush)
2841 crtc_funcs->atomic_flush(crtc, NULL);
2842 }
2843 EXPORT_SYMBOL(drm_atomic_helper_disable_planes_on_crtc);
2844
2845 /**
2846 * drm_atomic_helper_cleanup_planes - cleanup plane resources after commit
2847 * @dev: DRM device
2848 * @old_state: atomic state object with old state structures
2849 *
2850 * This function cleans up plane state, specifically framebuffers, from the old
2851 * configuration. Hence the old configuration must be perserved in @old_state to
2852 * be able to call this function.
2853 *
2854 * This function must also be called on the new state when the atomic update
2855 * fails at any point after calling drm_atomic_helper_prepare_planes().
2856 */
drm_atomic_helper_cleanup_planes(struct drm_device * dev,struct drm_atomic_state * old_state)2857 void drm_atomic_helper_cleanup_planes(struct drm_device *dev,
2858 struct drm_atomic_state *old_state)
2859 {
2860 struct drm_plane *plane;
2861 struct drm_plane_state *old_plane_state, *new_plane_state;
2862 int i;
2863
2864 for_each_oldnew_plane_in_state(old_state, plane, old_plane_state, new_plane_state, i) {
2865 const struct drm_plane_helper_funcs *funcs;
2866 struct drm_plane_state *plane_state;
2867
2868 /*
2869 * This might be called before swapping when commit is aborted,
2870 * in which case we have to cleanup the new state.
2871 */
2872 if (old_plane_state == plane->state)
2873 plane_state = new_plane_state;
2874 else
2875 plane_state = old_plane_state;
2876
2877 funcs = plane->helper_private;
2878
2879 if (funcs->cleanup_fb)
2880 funcs->cleanup_fb(plane, plane_state);
2881 }
2882 }
2883 EXPORT_SYMBOL(drm_atomic_helper_cleanup_planes);
2884
2885 /**
2886 * drm_atomic_helper_swap_state - store atomic state into current sw state
2887 * @state: atomic state
2888 * @stall: stall for preceding commits
2889 *
2890 * This function stores the atomic state into the current state pointers in all
2891 * driver objects. It should be called after all failing steps have been done
2892 * and succeeded, but before the actual hardware state is committed.
2893 *
2894 * For cleanup and error recovery the current state for all changed objects will
2895 * be swapped into @state.
2896 *
2897 * With that sequence it fits perfectly into the plane prepare/cleanup sequence:
2898 *
2899 * 1. Call drm_atomic_helper_prepare_planes() with the staged atomic state.
2900 *
2901 * 2. Do any other steps that might fail.
2902 *
2903 * 3. Put the staged state into the current state pointers with this function.
2904 *
2905 * 4. Actually commit the hardware state.
2906 *
2907 * 5. Call drm_atomic_helper_cleanup_planes() with @state, which since step 3
2908 * contains the old state. Also do any other cleanup required with that state.
2909 *
2910 * @stall must be set when nonblocking commits for this driver directly access
2911 * the &drm_plane.state, &drm_crtc.state or &drm_connector.state pointer. With
2912 * the current atomic helpers this is almost always the case, since the helpers
2913 * don't pass the right state structures to the callbacks.
2914 *
2915 * Returns:
2916 *
2917 * Returns 0 on success. Can return -ERESTARTSYS when @stall is true and the
2918 * waiting for the previous commits has been interrupted.
2919 */
drm_atomic_helper_swap_state(struct drm_atomic_state * state,bool stall)2920 int drm_atomic_helper_swap_state(struct drm_atomic_state *state,
2921 bool stall)
2922 {
2923 int i, ret;
2924 struct drm_connector *connector;
2925 struct drm_connector_state *old_conn_state, *new_conn_state;
2926 struct drm_crtc *crtc;
2927 struct drm_crtc_state *old_crtc_state, *new_crtc_state;
2928 struct drm_plane *plane;
2929 struct drm_plane_state *old_plane_state, *new_plane_state;
2930 struct drm_crtc_commit *commit;
2931 struct drm_private_obj *obj;
2932 struct drm_private_state *old_obj_state, *new_obj_state;
2933
2934 if (stall) {
2935 /*
2936 * We have to stall for hw_done here before
2937 * drm_atomic_helper_wait_for_dependencies() because flip
2938 * depth > 1 is not yet supported by all drivers. As long as
2939 * obj->state is directly dereferenced anywhere in the drivers
2940 * atomic_commit_tail function, then it's unsafe to swap state
2941 * before drm_atomic_helper_commit_hw_done() is called.
2942 */
2943
2944 for_each_old_crtc_in_state(state, crtc, old_crtc_state, i) {
2945 commit = old_crtc_state->commit;
2946
2947 if (!commit)
2948 continue;
2949
2950 ret = wait_for_completion_interruptible(&commit->hw_done);
2951 if (ret)
2952 return ret;
2953 }
2954
2955 for_each_old_connector_in_state(state, connector, old_conn_state, i) {
2956 commit = old_conn_state->commit;
2957
2958 if (!commit)
2959 continue;
2960
2961 ret = wait_for_completion_interruptible(&commit->hw_done);
2962 if (ret)
2963 return ret;
2964 }
2965
2966 for_each_old_plane_in_state(state, plane, old_plane_state, i) {
2967 commit = old_plane_state->commit;
2968
2969 if (!commit)
2970 continue;
2971
2972 ret = wait_for_completion_interruptible(&commit->hw_done);
2973 if (ret)
2974 return ret;
2975 }
2976 }
2977
2978 for_each_oldnew_connector_in_state(state, connector, old_conn_state, new_conn_state, i) {
2979 WARN_ON(connector->state != old_conn_state);
2980
2981 old_conn_state->state = state;
2982 new_conn_state->state = NULL;
2983
2984 state->connectors[i].state = old_conn_state;
2985 connector->state = new_conn_state;
2986 }
2987
2988 for_each_oldnew_crtc_in_state(state, crtc, old_crtc_state, new_crtc_state, i) {
2989 WARN_ON(crtc->state != old_crtc_state);
2990
2991 old_crtc_state->state = state;
2992 new_crtc_state->state = NULL;
2993
2994 state->crtcs[i].state = old_crtc_state;
2995 crtc->state = new_crtc_state;
2996
2997 if (new_crtc_state->commit) {
2998 spin_lock(&crtc->commit_lock);
2999 list_add(&new_crtc_state->commit->commit_entry,
3000 &crtc->commit_list);
3001 spin_unlock(&crtc->commit_lock);
3002
3003 new_crtc_state->commit->event = NULL;
3004 }
3005 }
3006
3007 for_each_oldnew_plane_in_state(state, plane, old_plane_state, new_plane_state, i) {
3008 WARN_ON(plane->state != old_plane_state);
3009
3010 old_plane_state->state = state;
3011 new_plane_state->state = NULL;
3012
3013 state->planes[i].state = old_plane_state;
3014 plane->state = new_plane_state;
3015 }
3016
3017 for_each_oldnew_private_obj_in_state(state, obj, old_obj_state, new_obj_state, i) {
3018 WARN_ON(obj->state != old_obj_state);
3019
3020 old_obj_state->state = state;
3021 new_obj_state->state = NULL;
3022
3023 state->private_objs[i].state = old_obj_state;
3024 obj->state = new_obj_state;
3025 }
3026
3027 return 0;
3028 }
3029 EXPORT_SYMBOL(drm_atomic_helper_swap_state);
3030
3031 /**
3032 * drm_atomic_helper_update_plane - Helper for primary plane update using atomic
3033 * @plane: plane object to update
3034 * @crtc: owning CRTC of owning plane
3035 * @fb: framebuffer to flip onto plane
3036 * @crtc_x: x offset of primary plane on @crtc
3037 * @crtc_y: y offset of primary plane on @crtc
3038 * @crtc_w: width of primary plane rectangle on @crtc
3039 * @crtc_h: height of primary plane rectangle on @crtc
3040 * @src_x: x offset of @fb for panning
3041 * @src_y: y offset of @fb for panning
3042 * @src_w: width of source rectangle in @fb
3043 * @src_h: height of source rectangle in @fb
3044 * @ctx: lock acquire context
3045 *
3046 * Provides a default plane update handler using the atomic driver interface.
3047 *
3048 * RETURNS:
3049 * Zero on success, error code on failure
3050 */
drm_atomic_helper_update_plane(struct drm_plane * plane,struct drm_crtc * crtc,struct drm_framebuffer * fb,int crtc_x,int crtc_y,unsigned int crtc_w,unsigned int crtc_h,uint32_t src_x,uint32_t src_y,uint32_t src_w,uint32_t src_h,struct drm_modeset_acquire_ctx * ctx)3051 int drm_atomic_helper_update_plane(struct drm_plane *plane,
3052 struct drm_crtc *crtc,
3053 struct drm_framebuffer *fb,
3054 int crtc_x, int crtc_y,
3055 unsigned int crtc_w, unsigned int crtc_h,
3056 uint32_t src_x, uint32_t src_y,
3057 uint32_t src_w, uint32_t src_h,
3058 struct drm_modeset_acquire_ctx *ctx)
3059 {
3060 struct drm_atomic_state *state;
3061 struct drm_plane_state *plane_state;
3062 int ret = 0;
3063
3064 state = drm_atomic_state_alloc(plane->dev);
3065 if (!state)
3066 return -ENOMEM;
3067
3068 state->acquire_ctx = ctx;
3069 plane_state = drm_atomic_get_plane_state(state, plane);
3070 if (IS_ERR(plane_state)) {
3071 ret = PTR_ERR(plane_state);
3072 goto fail;
3073 }
3074
3075 ret = drm_atomic_set_crtc_for_plane(plane_state, crtc);
3076 if (ret != 0)
3077 goto fail;
3078 drm_atomic_set_fb_for_plane(plane_state, fb);
3079 plane_state->crtc_x = crtc_x;
3080 plane_state->crtc_y = crtc_y;
3081 plane_state->crtc_w = crtc_w;
3082 plane_state->crtc_h = crtc_h;
3083 plane_state->src_x = src_x;
3084 plane_state->src_y = src_y;
3085 plane_state->src_w = src_w;
3086 plane_state->src_h = src_h;
3087
3088 if (plane == crtc->cursor)
3089 state->legacy_cursor_update = true;
3090
3091 ret = drm_atomic_commit(state);
3092 fail:
3093 drm_atomic_state_put(state);
3094 return ret;
3095 }
3096 EXPORT_SYMBOL(drm_atomic_helper_update_plane);
3097
3098 /**
3099 * drm_atomic_helper_disable_plane - Helper for primary plane disable using * atomic
3100 * @plane: plane to disable
3101 * @ctx: lock acquire context
3102 *
3103 * Provides a default plane disable handler using the atomic driver interface.
3104 *
3105 * RETURNS:
3106 * Zero on success, error code on failure
3107 */
drm_atomic_helper_disable_plane(struct drm_plane * plane,struct drm_modeset_acquire_ctx * ctx)3108 int drm_atomic_helper_disable_plane(struct drm_plane *plane,
3109 struct drm_modeset_acquire_ctx *ctx)
3110 {
3111 struct drm_atomic_state *state;
3112 struct drm_plane_state *plane_state;
3113 int ret = 0;
3114
3115 state = drm_atomic_state_alloc(plane->dev);
3116 if (!state)
3117 return -ENOMEM;
3118
3119 state->acquire_ctx = ctx;
3120 plane_state = drm_atomic_get_plane_state(state, plane);
3121 if (IS_ERR(plane_state)) {
3122 ret = PTR_ERR(plane_state);
3123 goto fail;
3124 }
3125
3126 if (plane_state->crtc && plane_state->crtc->cursor == plane)
3127 plane_state->state->legacy_cursor_update = true;
3128
3129 ret = __drm_atomic_helper_disable_plane(plane, plane_state);
3130 if (ret != 0)
3131 goto fail;
3132
3133 ret = drm_atomic_commit(state);
3134 fail:
3135 drm_atomic_state_put(state);
3136 return ret;
3137 }
3138 EXPORT_SYMBOL(drm_atomic_helper_disable_plane);
3139
3140 /**
3141 * drm_atomic_helper_set_config - set a new config from userspace
3142 * @set: mode set configuration
3143 * @ctx: lock acquisition context
3144 *
3145 * Provides a default CRTC set_config handler using the atomic driver interface.
3146 *
3147 * NOTE: For backwards compatibility with old userspace this automatically
3148 * resets the "link-status" property to GOOD, to force any link
3149 * re-training. The SETCRTC ioctl does not define whether an update does
3150 * need a full modeset or just a plane update, hence we're allowed to do
3151 * that. See also drm_connector_set_link_status_property().
3152 *
3153 * Returns:
3154 * Returns 0 on success, negative errno numbers on failure.
3155 */
drm_atomic_helper_set_config(struct drm_mode_set * set,struct drm_modeset_acquire_ctx * ctx)3156 int drm_atomic_helper_set_config(struct drm_mode_set *set,
3157 struct drm_modeset_acquire_ctx *ctx)
3158 {
3159 struct drm_atomic_state *state;
3160 struct drm_crtc *crtc = set->crtc;
3161 int ret = 0;
3162
3163 state = drm_atomic_state_alloc(crtc->dev);
3164 if (!state)
3165 return -ENOMEM;
3166
3167 state->acquire_ctx = ctx;
3168 ret = __drm_atomic_helper_set_config(set, state);
3169 if (ret != 0)
3170 goto fail;
3171
3172 ret = handle_conflicting_encoders(state, true);
3173 if (ret)
3174 goto fail;
3175
3176 ret = drm_atomic_commit(state);
3177
3178 fail:
3179 drm_atomic_state_put(state);
3180 return ret;
3181 }
3182 EXPORT_SYMBOL(drm_atomic_helper_set_config);
3183
3184 /**
3185 * drm_atomic_helper_disable_all - disable all currently active outputs
3186 * @dev: DRM device
3187 * @ctx: lock acquisition context
3188 *
3189 * Loops through all connectors, finding those that aren't turned off and then
3190 * turns them off by setting their DPMS mode to OFF and deactivating the CRTC
3191 * that they are connected to.
3192 *
3193 * This is used for example in suspend/resume to disable all currently active
3194 * functions when suspending. If you just want to shut down everything at e.g.
3195 * driver unload, look at drm_atomic_helper_shutdown().
3196 *
3197 * Note that if callers haven't already acquired all modeset locks this might
3198 * return -EDEADLK, which must be handled by calling drm_modeset_backoff().
3199 *
3200 * Returns:
3201 * 0 on success or a negative error code on failure.
3202 *
3203 * See also:
3204 * drm_atomic_helper_suspend(), drm_atomic_helper_resume() and
3205 * drm_atomic_helper_shutdown().
3206 */
drm_atomic_helper_disable_all(struct drm_device * dev,struct drm_modeset_acquire_ctx * ctx)3207 int drm_atomic_helper_disable_all(struct drm_device *dev,
3208 struct drm_modeset_acquire_ctx *ctx)
3209 {
3210 struct drm_atomic_state *state;
3211 struct drm_connector_state *conn_state;
3212 struct drm_connector *conn;
3213 struct drm_plane_state *plane_state;
3214 struct drm_plane *plane;
3215 struct drm_crtc_state *crtc_state;
3216 struct drm_crtc *crtc;
3217 int ret, i;
3218
3219 state = drm_atomic_state_alloc(dev);
3220 if (!state)
3221 return -ENOMEM;
3222
3223 state->acquire_ctx = ctx;
3224
3225 drm_for_each_crtc(crtc, dev) {
3226 crtc_state = drm_atomic_get_crtc_state(state, crtc);
3227 if (IS_ERR(crtc_state)) {
3228 ret = PTR_ERR(crtc_state);
3229 goto free;
3230 }
3231
3232 crtc_state->active = false;
3233
3234 ret = drm_atomic_set_mode_prop_for_crtc(crtc_state, NULL);
3235 if (ret < 0)
3236 goto free;
3237
3238 ret = drm_atomic_add_affected_planes(state, crtc);
3239 if (ret < 0)
3240 goto free;
3241
3242 ret = drm_atomic_add_affected_connectors(state, crtc);
3243 if (ret < 0)
3244 goto free;
3245 }
3246
3247 for_each_new_connector_in_state(state, conn, conn_state, i) {
3248 ret = drm_atomic_set_crtc_for_connector(conn_state, NULL);
3249 if (ret < 0)
3250 goto free;
3251 }
3252
3253 for_each_new_plane_in_state(state, plane, plane_state, i) {
3254 ret = drm_atomic_set_crtc_for_plane(plane_state, NULL);
3255 if (ret < 0)
3256 goto free;
3257
3258 drm_atomic_set_fb_for_plane(plane_state, NULL);
3259 }
3260
3261 ret = drm_atomic_commit(state);
3262 free:
3263 drm_atomic_state_put(state);
3264 return ret;
3265 }
3266 EXPORT_SYMBOL(drm_atomic_helper_disable_all);
3267
3268 /**
3269 * drm_atomic_helper_shutdown - shutdown all CRTC
3270 * @dev: DRM device
3271 *
3272 * This shuts down all CRTC, which is useful for driver unloading. Shutdown on
3273 * suspend should instead be handled with drm_atomic_helper_suspend(), since
3274 * that also takes a snapshot of the modeset state to be restored on resume.
3275 *
3276 * This is just a convenience wrapper around drm_atomic_helper_disable_all(),
3277 * and it is the atomic version of drm_crtc_force_disable_all().
3278 */
drm_atomic_helper_shutdown(struct drm_device * dev)3279 void drm_atomic_helper_shutdown(struct drm_device *dev)
3280 {
3281 struct drm_modeset_acquire_ctx ctx;
3282 int ret;
3283
3284 DRM_MODESET_LOCK_ALL_BEGIN(dev, ctx, 0, ret);
3285
3286 ret = drm_atomic_helper_disable_all(dev, &ctx);
3287 if (ret)
3288 drm_err(dev,
3289 "Disabling all crtc's during unload failed with %i\n",
3290 ret);
3291
3292 DRM_MODESET_LOCK_ALL_END(dev, ctx, ret);
3293 }
3294 EXPORT_SYMBOL(drm_atomic_helper_shutdown);
3295
3296 /**
3297 * drm_atomic_helper_duplicate_state - duplicate an atomic state object
3298 * @dev: DRM device
3299 * @ctx: lock acquisition context
3300 *
3301 * Makes a copy of the current atomic state by looping over all objects and
3302 * duplicating their respective states. This is used for example by suspend/
3303 * resume support code to save the state prior to suspend such that it can
3304 * be restored upon resume.
3305 *
3306 * Note that this treats atomic state as persistent between save and restore.
3307 * Drivers must make sure that this is possible and won't result in confusion
3308 * or erroneous behaviour.
3309 *
3310 * Note that if callers haven't already acquired all modeset locks this might
3311 * return -EDEADLK, which must be handled by calling drm_modeset_backoff().
3312 *
3313 * Returns:
3314 * A pointer to the copy of the atomic state object on success or an
3315 * ERR_PTR()-encoded error code on failure.
3316 *
3317 * See also:
3318 * drm_atomic_helper_suspend(), drm_atomic_helper_resume()
3319 */
3320 struct drm_atomic_state *
drm_atomic_helper_duplicate_state(struct drm_device * dev,struct drm_modeset_acquire_ctx * ctx)3321 drm_atomic_helper_duplicate_state(struct drm_device *dev,
3322 struct drm_modeset_acquire_ctx *ctx)
3323 {
3324 struct drm_atomic_state *state;
3325 struct drm_connector *conn;
3326 struct drm_connector_list_iter conn_iter;
3327 struct drm_plane *plane;
3328 struct drm_crtc *crtc;
3329 int err = 0;
3330
3331 state = drm_atomic_state_alloc(dev);
3332 if (!state)
3333 return ERR_PTR(-ENOMEM);
3334
3335 state->acquire_ctx = ctx;
3336 state->duplicated = true;
3337
3338 drm_for_each_crtc(crtc, dev) {
3339 struct drm_crtc_state *crtc_state;
3340
3341 crtc_state = drm_atomic_get_crtc_state(state, crtc);
3342 if (IS_ERR(crtc_state)) {
3343 err = PTR_ERR(crtc_state);
3344 goto free;
3345 }
3346 }
3347
3348 drm_for_each_plane(plane, dev) {
3349 struct drm_plane_state *plane_state;
3350
3351 plane_state = drm_atomic_get_plane_state(state, plane);
3352 if (IS_ERR(plane_state)) {
3353 err = PTR_ERR(plane_state);
3354 goto free;
3355 }
3356 }
3357
3358 drm_connector_list_iter_begin(dev, &conn_iter);
3359 drm_for_each_connector_iter(conn, &conn_iter) {
3360 struct drm_connector_state *conn_state;
3361
3362 conn_state = drm_atomic_get_connector_state(state, conn);
3363 if (IS_ERR(conn_state)) {
3364 err = PTR_ERR(conn_state);
3365 drm_connector_list_iter_end(&conn_iter);
3366 goto free;
3367 }
3368 }
3369 drm_connector_list_iter_end(&conn_iter);
3370
3371 /* clear the acquire context so that it isn't accidentally reused */
3372 state->acquire_ctx = NULL;
3373
3374 free:
3375 if (err < 0) {
3376 drm_atomic_state_put(state);
3377 state = ERR_PTR(err);
3378 }
3379
3380 return state;
3381 }
3382 EXPORT_SYMBOL(drm_atomic_helper_duplicate_state);
3383
3384 /**
3385 * drm_atomic_helper_suspend - subsystem-level suspend helper
3386 * @dev: DRM device
3387 *
3388 * Duplicates the current atomic state, disables all active outputs and then
3389 * returns a pointer to the original atomic state to the caller. Drivers can
3390 * pass this pointer to the drm_atomic_helper_resume() helper upon resume to
3391 * restore the output configuration that was active at the time the system
3392 * entered suspend.
3393 *
3394 * Note that it is potentially unsafe to use this. The atomic state object
3395 * returned by this function is assumed to be persistent. Drivers must ensure
3396 * that this holds true. Before calling this function, drivers must make sure
3397 * to suspend fbdev emulation so that nothing can be using the device.
3398 *
3399 * Returns:
3400 * A pointer to a copy of the state before suspend on success or an ERR_PTR()-
3401 * encoded error code on failure. Drivers should store the returned atomic
3402 * state object and pass it to the drm_atomic_helper_resume() helper upon
3403 * resume.
3404 *
3405 * See also:
3406 * drm_atomic_helper_duplicate_state(), drm_atomic_helper_disable_all(),
3407 * drm_atomic_helper_resume(), drm_atomic_helper_commit_duplicated_state()
3408 */
drm_atomic_helper_suspend(struct drm_device * dev)3409 struct drm_atomic_state *drm_atomic_helper_suspend(struct drm_device *dev)
3410 {
3411 struct drm_modeset_acquire_ctx ctx;
3412 struct drm_atomic_state *state;
3413 int err;
3414
3415 /* This can never be returned, but it makes the compiler happy */
3416 state = ERR_PTR(-EINVAL);
3417
3418 DRM_MODESET_LOCK_ALL_BEGIN(dev, ctx, 0, err);
3419
3420 state = drm_atomic_helper_duplicate_state(dev, &ctx);
3421 if (IS_ERR(state))
3422 goto unlock;
3423
3424 err = drm_atomic_helper_disable_all(dev, &ctx);
3425 if (err < 0) {
3426 drm_atomic_state_put(state);
3427 state = ERR_PTR(err);
3428 goto unlock;
3429 }
3430
3431 unlock:
3432 DRM_MODESET_LOCK_ALL_END(dev, ctx, err);
3433 if (err)
3434 return ERR_PTR(err);
3435
3436 return state;
3437 }
3438 EXPORT_SYMBOL(drm_atomic_helper_suspend);
3439
3440 /**
3441 * drm_atomic_helper_commit_duplicated_state - commit duplicated state
3442 * @state: duplicated atomic state to commit
3443 * @ctx: pointer to acquire_ctx to use for commit.
3444 *
3445 * The state returned by drm_atomic_helper_duplicate_state() and
3446 * drm_atomic_helper_suspend() is partially invalid, and needs to
3447 * be fixed up before commit.
3448 *
3449 * Returns:
3450 * 0 on success or a negative error code on failure.
3451 *
3452 * See also:
3453 * drm_atomic_helper_suspend()
3454 */
drm_atomic_helper_commit_duplicated_state(struct drm_atomic_state * state,struct drm_modeset_acquire_ctx * ctx)3455 int drm_atomic_helper_commit_duplicated_state(struct drm_atomic_state *state,
3456 struct drm_modeset_acquire_ctx *ctx)
3457 {
3458 int i, ret;
3459 struct drm_plane *plane;
3460 struct drm_plane_state *new_plane_state;
3461 struct drm_connector *connector;
3462 struct drm_connector_state *new_conn_state;
3463 struct drm_crtc *crtc;
3464 struct drm_crtc_state *new_crtc_state;
3465
3466 state->acquire_ctx = ctx;
3467
3468 for_each_new_plane_in_state(state, plane, new_plane_state, i)
3469 state->planes[i].old_state = plane->state;
3470
3471 for_each_new_crtc_in_state(state, crtc, new_crtc_state, i)
3472 state->crtcs[i].old_state = crtc->state;
3473
3474 for_each_new_connector_in_state(state, connector, new_conn_state, i)
3475 state->connectors[i].old_state = connector->state;
3476
3477 ret = drm_atomic_commit(state);
3478
3479 state->acquire_ctx = NULL;
3480
3481 return ret;
3482 }
3483 EXPORT_SYMBOL(drm_atomic_helper_commit_duplicated_state);
3484
3485 /**
3486 * drm_atomic_helper_resume - subsystem-level resume helper
3487 * @dev: DRM device
3488 * @state: atomic state to resume to
3489 *
3490 * Calls drm_mode_config_reset() to synchronize hardware and software states,
3491 * grabs all modeset locks and commits the atomic state object. This can be
3492 * used in conjunction with the drm_atomic_helper_suspend() helper to
3493 * implement suspend/resume for drivers that support atomic mode-setting.
3494 *
3495 * Returns:
3496 * 0 on success or a negative error code on failure.
3497 *
3498 * See also:
3499 * drm_atomic_helper_suspend()
3500 */
drm_atomic_helper_resume(struct drm_device * dev,struct drm_atomic_state * state)3501 int drm_atomic_helper_resume(struct drm_device *dev,
3502 struct drm_atomic_state *state)
3503 {
3504 struct drm_modeset_acquire_ctx ctx;
3505 int err;
3506
3507 drm_mode_config_reset(dev);
3508
3509 DRM_MODESET_LOCK_ALL_BEGIN(dev, ctx, 0, err);
3510
3511 err = drm_atomic_helper_commit_duplicated_state(state, &ctx);
3512
3513 DRM_MODESET_LOCK_ALL_END(dev, ctx, err);
3514 drm_atomic_state_put(state);
3515
3516 return err;
3517 }
3518 EXPORT_SYMBOL(drm_atomic_helper_resume);
3519
page_flip_common(struct drm_atomic_state * state,struct drm_crtc * crtc,struct drm_framebuffer * fb,struct drm_pending_vblank_event * event,uint32_t flags)3520 static int page_flip_common(struct drm_atomic_state *state,
3521 struct drm_crtc *crtc,
3522 struct drm_framebuffer *fb,
3523 struct drm_pending_vblank_event *event,
3524 uint32_t flags)
3525 {
3526 struct drm_plane *plane = crtc->primary;
3527 struct drm_plane_state *plane_state;
3528 struct drm_crtc_state *crtc_state;
3529 int ret = 0;
3530
3531 crtc_state = drm_atomic_get_crtc_state(state, crtc);
3532 if (IS_ERR(crtc_state))
3533 return PTR_ERR(crtc_state);
3534
3535 crtc_state->event = event;
3536 crtc_state->async_flip = flags & DRM_MODE_PAGE_FLIP_ASYNC;
3537
3538 plane_state = drm_atomic_get_plane_state(state, plane);
3539 if (IS_ERR(plane_state))
3540 return PTR_ERR(plane_state);
3541
3542 ret = drm_atomic_set_crtc_for_plane(plane_state, crtc);
3543 if (ret != 0)
3544 return ret;
3545 drm_atomic_set_fb_for_plane(plane_state, fb);
3546
3547 /* Make sure we don't accidentally do a full modeset. */
3548 state->allow_modeset = false;
3549 if (!crtc_state->active) {
3550 drm_dbg_atomic(crtc->dev,
3551 "[CRTC:%d:%s] disabled, rejecting legacy flip\n",
3552 crtc->base.id, crtc->name);
3553 return -EINVAL;
3554 }
3555
3556 return ret;
3557 }
3558
3559 /**
3560 * drm_atomic_helper_page_flip - execute a legacy page flip
3561 * @crtc: DRM CRTC
3562 * @fb: DRM framebuffer
3563 * @event: optional DRM event to signal upon completion
3564 * @flags: flip flags for non-vblank sync'ed updates
3565 * @ctx: lock acquisition context
3566 *
3567 * Provides a default &drm_crtc_funcs.page_flip implementation
3568 * using the atomic driver interface.
3569 *
3570 * Returns:
3571 * Returns 0 on success, negative errno numbers on failure.
3572 *
3573 * See also:
3574 * drm_atomic_helper_page_flip_target()
3575 */
drm_atomic_helper_page_flip(struct drm_crtc * crtc,struct drm_framebuffer * fb,struct drm_pending_vblank_event * event,uint32_t flags,struct drm_modeset_acquire_ctx * ctx)3576 int drm_atomic_helper_page_flip(struct drm_crtc *crtc,
3577 struct drm_framebuffer *fb,
3578 struct drm_pending_vblank_event *event,
3579 uint32_t flags,
3580 struct drm_modeset_acquire_ctx *ctx)
3581 {
3582 struct drm_plane *plane = crtc->primary;
3583 struct drm_atomic_state *state;
3584 int ret = 0;
3585
3586 state = drm_atomic_state_alloc(plane->dev);
3587 if (!state)
3588 return -ENOMEM;
3589
3590 state->acquire_ctx = ctx;
3591
3592 ret = page_flip_common(state, crtc, fb, event, flags);
3593 if (ret != 0)
3594 goto fail;
3595
3596 ret = drm_atomic_nonblocking_commit(state);
3597 fail:
3598 drm_atomic_state_put(state);
3599 return ret;
3600 }
3601 EXPORT_SYMBOL(drm_atomic_helper_page_flip);
3602
3603 /**
3604 * drm_atomic_helper_page_flip_target - do page flip on target vblank period.
3605 * @crtc: DRM CRTC
3606 * @fb: DRM framebuffer
3607 * @event: optional DRM event to signal upon completion
3608 * @flags: flip flags for non-vblank sync'ed updates
3609 * @target: specifying the target vblank period when the flip to take effect
3610 * @ctx: lock acquisition context
3611 *
3612 * Provides a default &drm_crtc_funcs.page_flip_target implementation.
3613 * Similar to drm_atomic_helper_page_flip() with extra parameter to specify
3614 * target vblank period to flip.
3615 *
3616 * Returns:
3617 * Returns 0 on success, negative errno numbers on failure.
3618 */
drm_atomic_helper_page_flip_target(struct drm_crtc * crtc,struct drm_framebuffer * fb,struct drm_pending_vblank_event * event,uint32_t flags,uint32_t target,struct drm_modeset_acquire_ctx * ctx)3619 int drm_atomic_helper_page_flip_target(struct drm_crtc *crtc,
3620 struct drm_framebuffer *fb,
3621 struct drm_pending_vblank_event *event,
3622 uint32_t flags,
3623 uint32_t target,
3624 struct drm_modeset_acquire_ctx *ctx)
3625 {
3626 struct drm_plane *plane = crtc->primary;
3627 struct drm_atomic_state *state;
3628 struct drm_crtc_state *crtc_state;
3629 int ret = 0;
3630
3631 state = drm_atomic_state_alloc(plane->dev);
3632 if (!state)
3633 return -ENOMEM;
3634
3635 state->acquire_ctx = ctx;
3636
3637 ret = page_flip_common(state, crtc, fb, event, flags);
3638 if (ret != 0)
3639 goto fail;
3640
3641 crtc_state = drm_atomic_get_new_crtc_state(state, crtc);
3642 if (WARN_ON(!crtc_state)) {
3643 ret = -EINVAL;
3644 goto fail;
3645 }
3646 crtc_state->target_vblank = target;
3647
3648 ret = drm_atomic_nonblocking_commit(state);
3649 fail:
3650 drm_atomic_state_put(state);
3651 return ret;
3652 }
3653 EXPORT_SYMBOL(drm_atomic_helper_page_flip_target);
3654
3655 /**
3656 * drm_atomic_helper_bridge_propagate_bus_fmt() - Propagate output format to
3657 * the input end of a bridge
3658 * @bridge: bridge control structure
3659 * @bridge_state: new bridge state
3660 * @crtc_state: new CRTC state
3661 * @conn_state: new connector state
3662 * @output_fmt: tested output bus format
3663 * @num_input_fmts: will contain the size of the returned array
3664 *
3665 * This helper is a pluggable implementation of the
3666 * &drm_bridge_funcs.atomic_get_input_bus_fmts operation for bridges that don't
3667 * modify the bus configuration between their input and their output. It
3668 * returns an array of input formats with a single element set to @output_fmt.
3669 *
3670 * RETURNS:
3671 * a valid format array of size @num_input_fmts, or NULL if the allocation
3672 * failed
3673 */
3674 u32 *
drm_atomic_helper_bridge_propagate_bus_fmt(struct drm_bridge * bridge,struct drm_bridge_state * bridge_state,struct drm_crtc_state * crtc_state,struct drm_connector_state * conn_state,u32 output_fmt,unsigned int * num_input_fmts)3675 drm_atomic_helper_bridge_propagate_bus_fmt(struct drm_bridge *bridge,
3676 struct drm_bridge_state *bridge_state,
3677 struct drm_crtc_state *crtc_state,
3678 struct drm_connector_state *conn_state,
3679 u32 output_fmt,
3680 unsigned int *num_input_fmts)
3681 {
3682 u32 *input_fmts;
3683
3684 input_fmts = kzalloc(sizeof(*input_fmts), GFP_KERNEL);
3685 if (!input_fmts) {
3686 *num_input_fmts = 0;
3687 return NULL;
3688 }
3689
3690 *num_input_fmts = 1;
3691 input_fmts[0] = output_fmt;
3692 return input_fmts;
3693 }
3694 EXPORT_SYMBOL(drm_atomic_helper_bridge_propagate_bus_fmt);
3695