1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright (C) 2012 Texas Instruments
7 #include <linux/delay.h>
8 #include <linux/dma-mapping.h>
9 #include <linux/of_graph.h>
10 #include <linux/pm_runtime.h>
12 #include <drm/drm_atomic.h>
13 #include <drm/drm_atomic_helper.h>
14 #include <drm/drm_crtc.h>
15 #include <drm/drm_fb_dma_helper.h>
16 #include <drm/drm_fourcc.h>
17 #include <drm/drm_framebuffer.h>
18 #include <drm/drm_gem_dma_helper.h>
19 #include <drm/drm_modeset_helper_vtables.h>
20 #include <drm/drm_print.h>
21 #include <drm/drm_vblank.h>
23 #include "tilcdc_drv.h"
24 #include "tilcdc_regs.h"
26 #define TILCDC_VBLANK_SAFETY_THRESHOLD_US 1000
27 #define TILCDC_PALETTE_SIZE 32
28 #define TILCDC_PALETTE_FIRST_ENTRY 0x4000
33 struct drm_plane primary;
34 const struct tilcdc_panel_info *info;
35 struct drm_pending_vblank_event *event;
36 struct mutex enable_lock;
39 wait_queue_head_t frame_done_wq;
43 unsigned int lcd_fck_rate;
46 unsigned int hvtotal_us;
48 struct drm_framebuffer *next_fb;
50 /* Only set if an external encoder is connected */
51 bool simulate_vesa_sync;
55 struct work_struct recover_work;
57 dma_addr_t palette_dma_handle;
59 struct completion palette_loaded;
61 #define to_tilcdc_crtc(x) container_of(x, struct tilcdc_crtc, base)
63 static void set_scanout(struct drm_crtc *crtc, struct drm_framebuffer *fb)
65 struct drm_device *dev = crtc->dev;
66 struct tilcdc_drm_private *priv = dev->dev_private;
67 struct drm_gem_dma_object *gem;
68 dma_addr_t start, end;
69 u64 dma_base_and_ceiling;
71 gem = drm_fb_dma_get_gem_obj(fb, 0);
73 start = gem->dma_addr + fb->offsets[0] +
74 crtc->y * fb->pitches[0] +
75 crtc->x * fb->format->cpp[0];
77 end = start + (crtc->mode.vdisplay * fb->pitches[0]);
79 /* Write LCDC_DMA_FB_BASE_ADDR_0_REG and LCDC_DMA_FB_CEILING_ADDR_0_REG
80 * with a single insruction, if available. This should make it more
81 * unlikely that LCDC would fetch the DMA addresses in the middle of
87 dma_base_and_ceiling = (u64)end << 32 | start;
88 tilcdc_write64(dev, LCDC_DMA_FB_BASE_ADDR_0_REG, dma_base_and_ceiling);
92 * The driver currently only supports only true color formats. For
93 * true color the palette block is bypassed, but a 32 byte palette
94 * should still be loaded. The first 16-bit entry must be 0x4000 while
95 * all other entries must be zeroed.
97 static void tilcdc_crtc_load_palette(struct drm_crtc *crtc)
99 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
100 struct drm_device *dev = crtc->dev;
101 struct tilcdc_drm_private *priv = dev->dev_private;
104 reinit_completion(&tilcdc_crtc->palette_loaded);
106 /* Tell the LCDC where the palette is located. */
107 tilcdc_write(dev, LCDC_DMA_FB_BASE_ADDR_0_REG,
108 tilcdc_crtc->palette_dma_handle);
109 tilcdc_write(dev, LCDC_DMA_FB_CEILING_ADDR_0_REG,
110 (u32) tilcdc_crtc->palette_dma_handle +
111 TILCDC_PALETTE_SIZE - 1);
113 /* Set dma load mode for palette loading only. */
114 tilcdc_write_mask(dev, LCDC_RASTER_CTRL_REG,
115 LCDC_PALETTE_LOAD_MODE(PALETTE_ONLY),
116 LCDC_PALETTE_LOAD_MODE_MASK);
118 /* Enable DMA Palette Loaded Interrupt */
120 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_V1_PL_INT_ENA);
122 tilcdc_write(dev, LCDC_INT_ENABLE_SET_REG, LCDC_V2_PL_INT_ENA);
124 /* Enable LCDC DMA and wait for palette to be loaded. */
125 tilcdc_clear_irqstatus(dev, 0xffffffff);
126 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
128 ret = wait_for_completion_timeout(&tilcdc_crtc->palette_loaded,
129 msecs_to_jiffies(50));
131 dev_err(dev->dev, "%s: Palette loading timeout", __func__);
133 /* Disable LCDC DMA and DMA Palette Loaded Interrupt. */
134 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
136 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_V1_PL_INT_ENA);
138 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG, LCDC_V2_PL_INT_ENA);
141 static void tilcdc_crtc_enable_irqs(struct drm_device *dev)
143 struct tilcdc_drm_private *priv = dev->dev_private;
145 tilcdc_clear_irqstatus(dev, 0xffffffff);
147 if (priv->rev == 1) {
148 tilcdc_set(dev, LCDC_RASTER_CTRL_REG,
149 LCDC_V1_SYNC_LOST_INT_ENA | LCDC_V1_FRAME_DONE_INT_ENA |
150 LCDC_V1_UNDERFLOW_INT_ENA);
152 tilcdc_write(dev, LCDC_INT_ENABLE_SET_REG,
153 LCDC_V2_UNDERFLOW_INT_ENA |
154 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
158 static void tilcdc_crtc_disable_irqs(struct drm_device *dev)
160 struct tilcdc_drm_private *priv = dev->dev_private;
162 /* disable irqs that we might have enabled: */
163 if (priv->rev == 1) {
164 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
165 LCDC_V1_SYNC_LOST_INT_ENA | LCDC_V1_FRAME_DONE_INT_ENA |
166 LCDC_V1_UNDERFLOW_INT_ENA | LCDC_V1_PL_INT_ENA);
167 tilcdc_clear(dev, LCDC_DMA_CTRL_REG,
168 LCDC_V1_END_OF_FRAME_INT_ENA);
170 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
171 LCDC_V2_UNDERFLOW_INT_ENA | LCDC_V2_PL_INT_ENA |
172 LCDC_V2_END_OF_FRAME0_INT_ENA |
173 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
177 static void reset(struct drm_crtc *crtc)
179 struct drm_device *dev = crtc->dev;
180 struct tilcdc_drm_private *priv = dev->dev_private;
185 tilcdc_set(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
186 usleep_range(250, 1000);
187 tilcdc_clear(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
191 * Calculate the percentage difference between the requested pixel clock rate
192 * and the effective rate resulting from calculating the clock divider value.
194 static unsigned int tilcdc_pclk_diff(unsigned long rate,
195 unsigned long real_rate)
197 int r = rate / 100, rr = real_rate / 100;
199 return (unsigned int)(abs(((rr - r) * 100) / r));
202 static void tilcdc_crtc_set_clk(struct drm_crtc *crtc)
204 struct drm_device *dev = crtc->dev;
205 struct tilcdc_drm_private *priv = dev->dev_private;
206 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
207 unsigned long clk_rate, real_pclk_rate, pclk_rate;
211 clkdiv = 2; /* first try using a standard divider of 2 */
213 /* mode.clock is in KHz, set_rate wants parameter in Hz */
214 pclk_rate = crtc->mode.clock * 1000;
216 ret = clk_set_rate(priv->clk, pclk_rate * clkdiv);
217 clk_rate = clk_get_rate(priv->clk);
218 real_pclk_rate = clk_rate / clkdiv;
219 if (ret < 0 || tilcdc_pclk_diff(pclk_rate, real_pclk_rate) > 5) {
221 * If we fail to set the clock rate (some architectures don't
222 * use the common clock framework yet and may not implement
223 * all the clk API calls for every clock), try the next best
224 * thing: adjusting the clock divider, unless clk_get_rate()
228 /* Nothing more we can do. Just bail out. */
230 "failed to set the pixel clock - unable to read current lcdc clock rate\n");
234 clkdiv = DIV_ROUND_CLOSEST(clk_rate, pclk_rate);
237 * Emit a warning if the real clock rate resulting from the
238 * calculated divider differs much from the requested rate.
240 * 5% is an arbitrary value - LCDs are usually quite tolerant
241 * about pixel clock rates.
243 real_pclk_rate = clk_rate / clkdiv;
245 if (tilcdc_pclk_diff(pclk_rate, real_pclk_rate) > 5) {
247 "effective pixel clock rate (%luHz) differs from the requested rate (%luHz)\n",
248 real_pclk_rate, pclk_rate);
252 tilcdc_crtc->lcd_fck_rate = clk_rate;
254 DBG("lcd_clk=%u, mode clock=%d, div=%u",
255 tilcdc_crtc->lcd_fck_rate, crtc->mode.clock, clkdiv);
257 /* Configure the LCD clock divisor. */
258 tilcdc_write(dev, LCDC_CTRL_REG, LCDC_CLK_DIVISOR(clkdiv) |
262 tilcdc_set(dev, LCDC_CLK_ENABLE_REG,
263 LCDC_V2_DMA_CLK_EN | LCDC_V2_LIDD_CLK_EN |
264 LCDC_V2_CORE_CLK_EN);
267 static uint tilcdc_mode_hvtotal(const struct drm_display_mode *mode)
269 return (uint) div_u64(1000llu * mode->htotal * mode->vtotal,
273 static void tilcdc_crtc_set_mode(struct drm_crtc *crtc)
275 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
276 struct drm_device *dev = crtc->dev;
277 struct tilcdc_drm_private *priv = dev->dev_private;
278 const struct tilcdc_panel_info *info = tilcdc_crtc->info;
279 uint32_t reg, hbp, hfp, hsw, vbp, vfp, vsw;
280 struct drm_display_mode *mode = &crtc->state->adjusted_mode;
281 struct drm_framebuffer *fb = crtc->primary->state->fb;
289 /* Configure the Burst Size and fifo threshold of DMA: */
290 reg = tilcdc_read(dev, LCDC_DMA_CTRL_REG) & ~0x00000770;
291 switch (info->dma_burst_sz) {
293 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_1);
296 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_2);
299 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_4);
302 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_8);
305 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_16);
308 dev_err(dev->dev, "invalid burst size\n");
311 reg |= (info->fifo_th << 8);
312 tilcdc_write(dev, LCDC_DMA_CTRL_REG, reg);
314 /* Configure timings: */
315 hbp = mode->htotal - mode->hsync_end;
316 hfp = mode->hsync_start - mode->hdisplay;
317 hsw = mode->hsync_end - mode->hsync_start;
318 vbp = mode->vtotal - mode->vsync_end;
319 vfp = mode->vsync_start - mode->vdisplay;
320 vsw = mode->vsync_end - mode->vsync_start;
322 DBG("%dx%d, hbp=%u, hfp=%u, hsw=%u, vbp=%u, vfp=%u, vsw=%u",
323 mode->hdisplay, mode->vdisplay, hbp, hfp, hsw, vbp, vfp, vsw);
325 /* Set AC Bias Period and Number of Transitions per Interrupt: */
326 reg = tilcdc_read(dev, LCDC_RASTER_TIMING_2_REG) & ~0x000fff00;
327 reg |= LCDC_AC_BIAS_FREQUENCY(info->ac_bias) |
328 LCDC_AC_BIAS_TRANSITIONS_PER_INT(info->ac_bias_intrpt);
331 * subtract one from hfp, hbp, hsw because the hardware uses
334 if (priv->rev == 2) {
335 /* clear bits we're going to set */
337 reg |= ((hfp-1) & 0x300) >> 8;
338 reg |= ((hbp-1) & 0x300) >> 4;
339 reg |= ((hsw-1) & 0x3c0) << 21;
341 tilcdc_write(dev, LCDC_RASTER_TIMING_2_REG, reg);
343 reg = (((mode->hdisplay >> 4) - 1) << 4) |
344 (((hbp-1) & 0xff) << 24) |
345 (((hfp-1) & 0xff) << 16) |
346 (((hsw-1) & 0x3f) << 10);
348 reg |= (((mode->hdisplay >> 4) - 1) & 0x40) >> 3;
349 tilcdc_write(dev, LCDC_RASTER_TIMING_0_REG, reg);
351 reg = ((mode->vdisplay - 1) & 0x3ff) |
352 ((vbp & 0xff) << 24) |
353 ((vfp & 0xff) << 16) |
354 (((vsw-1) & 0x3f) << 10);
355 tilcdc_write(dev, LCDC_RASTER_TIMING_1_REG, reg);
358 * be sure to set Bit 10 for the V2 LCDC controller,
359 * otherwise limited to 1024 pixels width, stopping
360 * 1920x1080 being supported.
362 if (priv->rev == 2) {
363 if ((mode->vdisplay - 1) & 0x400) {
364 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG,
367 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG,
372 /* Configure display type: */
373 reg = tilcdc_read(dev, LCDC_RASTER_CTRL_REG) &
374 ~(LCDC_TFT_MODE | LCDC_MONO_8BIT_MODE | LCDC_MONOCHROME_MODE |
375 LCDC_V2_TFT_24BPP_MODE | LCDC_V2_TFT_24BPP_UNPACK |
376 0x000ff000 /* Palette Loading Delay bits */);
377 reg |= LCDC_TFT_MODE; /* no monochrome/passive support */
378 if (info->tft_alt_mode)
379 reg |= LCDC_TFT_ALT_ENABLE;
380 if (priv->rev == 2) {
381 switch (fb->format->format) {
382 case DRM_FORMAT_BGR565:
383 case DRM_FORMAT_RGB565:
385 case DRM_FORMAT_XBGR8888:
386 case DRM_FORMAT_XRGB8888:
387 reg |= LCDC_V2_TFT_24BPP_UNPACK;
389 case DRM_FORMAT_BGR888:
390 case DRM_FORMAT_RGB888:
391 reg |= LCDC_V2_TFT_24BPP_MODE;
394 dev_err(dev->dev, "invalid pixel format\n");
398 reg |= info->fdd << 12;
399 tilcdc_write(dev, LCDC_RASTER_CTRL_REG, reg);
401 if (info->invert_pxl_clk)
402 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
404 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
407 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
409 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
412 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
414 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
416 if (mode->flags & DRM_MODE_FLAG_NHSYNC)
417 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
419 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
421 if (mode->flags & DRM_MODE_FLAG_NVSYNC)
422 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
424 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
426 if (info->raster_order)
427 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
429 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
431 tilcdc_crtc_set_clk(crtc);
433 tilcdc_crtc_load_palette(crtc);
435 set_scanout(crtc, fb);
437 drm_mode_copy(&crtc->hwmode, &crtc->state->adjusted_mode);
439 tilcdc_crtc->hvtotal_us =
440 tilcdc_mode_hvtotal(&crtc->hwmode);
443 static void tilcdc_crtc_enable(struct drm_crtc *crtc)
445 struct drm_device *dev = crtc->dev;
446 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
449 mutex_lock(&tilcdc_crtc->enable_lock);
450 if (tilcdc_crtc->enabled || tilcdc_crtc->shutdown) {
451 mutex_unlock(&tilcdc_crtc->enable_lock);
455 pm_runtime_get_sync(dev->dev);
459 tilcdc_crtc_set_mode(crtc);
461 tilcdc_crtc_enable_irqs(dev);
463 tilcdc_clear(dev, LCDC_DMA_CTRL_REG, LCDC_DUAL_FRAME_BUFFER_ENABLE);
464 tilcdc_write_mask(dev, LCDC_RASTER_CTRL_REG,
465 LCDC_PALETTE_LOAD_MODE(DATA_ONLY),
466 LCDC_PALETTE_LOAD_MODE_MASK);
468 /* There is no real chance for a race here as the time stamp
469 * is taken before the raster DMA is started. The spin-lock is
470 * taken to have a memory barrier after taking the time-stamp
471 * and to avoid a context switch between taking the stamp and
472 * enabling the raster.
474 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
475 tilcdc_crtc->last_vblank = ktime_get();
476 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
477 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
479 drm_crtc_vblank_on(crtc);
481 tilcdc_crtc->enabled = true;
482 mutex_unlock(&tilcdc_crtc->enable_lock);
485 static void tilcdc_crtc_atomic_enable(struct drm_crtc *crtc,
486 struct drm_atomic_state *state)
488 tilcdc_crtc_enable(crtc);
491 static void tilcdc_crtc_off(struct drm_crtc *crtc, bool shutdown)
493 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
494 struct drm_device *dev = crtc->dev;
497 mutex_lock(&tilcdc_crtc->enable_lock);
499 tilcdc_crtc->shutdown = true;
500 if (!tilcdc_crtc->enabled) {
501 mutex_unlock(&tilcdc_crtc->enable_lock);
504 tilcdc_crtc->frame_done = false;
505 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
508 * Wait for framedone irq which will still come before putting
511 ret = wait_event_timeout(tilcdc_crtc->frame_done_wq,
512 tilcdc_crtc->frame_done,
513 msecs_to_jiffies(500));
515 dev_err(dev->dev, "%s: timeout waiting for framedone\n",
518 drm_crtc_vblank_off(crtc);
520 spin_lock_irq(&crtc->dev->event_lock);
522 if (crtc->state->event) {
523 drm_crtc_send_vblank_event(crtc, crtc->state->event);
524 crtc->state->event = NULL;
527 spin_unlock_irq(&crtc->dev->event_lock);
529 tilcdc_crtc_disable_irqs(dev);
531 pm_runtime_put_sync(dev->dev);
533 tilcdc_crtc->enabled = false;
534 mutex_unlock(&tilcdc_crtc->enable_lock);
537 static void tilcdc_crtc_disable(struct drm_crtc *crtc)
539 tilcdc_crtc_off(crtc, false);
542 static void tilcdc_crtc_atomic_disable(struct drm_crtc *crtc,
543 struct drm_atomic_state *state)
545 tilcdc_crtc_disable(crtc);
548 static void tilcdc_crtc_atomic_flush(struct drm_crtc *crtc,
549 struct drm_atomic_state *state)
551 if (!crtc->state->event)
554 spin_lock_irq(&crtc->dev->event_lock);
555 drm_crtc_send_vblank_event(crtc, crtc->state->event);
556 crtc->state->event = NULL;
557 spin_unlock_irq(&crtc->dev->event_lock);
560 void tilcdc_crtc_shutdown(struct drm_crtc *crtc)
562 tilcdc_crtc_off(crtc, true);
565 static bool tilcdc_crtc_is_on(struct drm_crtc *crtc)
567 return crtc->state && crtc->state->enable && crtc->state->active;
570 static void tilcdc_crtc_recover_work(struct work_struct *work)
572 struct tilcdc_crtc *tilcdc_crtc =
573 container_of(work, struct tilcdc_crtc, recover_work);
574 struct drm_crtc *crtc = &tilcdc_crtc->base;
576 dev_info(crtc->dev->dev, "%s: Reset CRTC", __func__);
578 drm_modeset_lock(&crtc->mutex, NULL);
580 if (!tilcdc_crtc_is_on(crtc))
583 tilcdc_crtc_disable(crtc);
584 tilcdc_crtc_enable(crtc);
586 drm_modeset_unlock(&crtc->mutex);
589 static void tilcdc_crtc_destroy(struct drm_crtc *crtc)
591 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
593 tilcdc_crtc_shutdown(crtc);
595 flush_workqueue(priv->wq);
597 of_node_put(crtc->port);
598 drm_crtc_cleanup(crtc);
601 int tilcdc_crtc_update_fb(struct drm_crtc *crtc,
602 struct drm_framebuffer *fb,
603 struct drm_pending_vblank_event *event)
605 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
606 struct drm_device *dev = crtc->dev;
608 if (tilcdc_crtc->event) {
609 dev_err(dev->dev, "already pending page flip!\n");
613 tilcdc_crtc->event = event;
615 mutex_lock(&tilcdc_crtc->enable_lock);
617 if (tilcdc_crtc->enabled) {
622 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
624 next_vblank = ktime_add_us(tilcdc_crtc->last_vblank,
625 tilcdc_crtc->hvtotal_us);
626 tdiff = ktime_to_us(ktime_sub(next_vblank, ktime_get()));
628 if (tdiff < TILCDC_VBLANK_SAFETY_THRESHOLD_US)
629 tilcdc_crtc->next_fb = fb;
631 set_scanout(crtc, fb);
633 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
636 mutex_unlock(&tilcdc_crtc->enable_lock);
641 static bool tilcdc_crtc_mode_fixup(struct drm_crtc *crtc,
642 const struct drm_display_mode *mode,
643 struct drm_display_mode *adjusted_mode)
645 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
647 if (!tilcdc_crtc->simulate_vesa_sync)
651 * tilcdc does not generate VESA-compliant sync but aligns
652 * VS on the second edge of HS instead of first edge.
653 * We use adjusted_mode, to fixup sync by aligning both rising
654 * edges and add HSKEW offset to fix the sync.
656 adjusted_mode->hskew = mode->hsync_end - mode->hsync_start;
657 adjusted_mode->flags |= DRM_MODE_FLAG_HSKEW;
659 if (mode->flags & DRM_MODE_FLAG_NHSYNC) {
660 adjusted_mode->flags |= DRM_MODE_FLAG_PHSYNC;
661 adjusted_mode->flags &= ~DRM_MODE_FLAG_NHSYNC;
663 adjusted_mode->flags |= DRM_MODE_FLAG_NHSYNC;
664 adjusted_mode->flags &= ~DRM_MODE_FLAG_PHSYNC;
670 static int tilcdc_crtc_atomic_check(struct drm_crtc *crtc,
671 struct drm_atomic_state *state)
673 struct drm_crtc_state *crtc_state = drm_atomic_get_new_crtc_state(state,
675 /* If we are not active we don't care */
676 if (!crtc_state->active)
679 if (state->planes[0].ptr != crtc->primary ||
680 state->planes[0].state == NULL ||
681 state->planes[0].state->crtc != crtc) {
682 dev_dbg(crtc->dev->dev, "CRTC primary plane must be present");
689 static int tilcdc_crtc_enable_vblank(struct drm_crtc *crtc)
691 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
692 struct drm_device *dev = crtc->dev;
693 struct tilcdc_drm_private *priv = dev->dev_private;
696 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
698 tilcdc_clear_irqstatus(dev, LCDC_END_OF_FRAME0);
701 tilcdc_set(dev, LCDC_DMA_CTRL_REG,
702 LCDC_V1_END_OF_FRAME_INT_ENA);
704 tilcdc_set(dev, LCDC_INT_ENABLE_SET_REG,
705 LCDC_V2_END_OF_FRAME0_INT_ENA);
707 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
712 static void tilcdc_crtc_disable_vblank(struct drm_crtc *crtc)
714 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
715 struct drm_device *dev = crtc->dev;
716 struct tilcdc_drm_private *priv = dev->dev_private;
719 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
722 tilcdc_clear(dev, LCDC_DMA_CTRL_REG,
723 LCDC_V1_END_OF_FRAME_INT_ENA);
725 tilcdc_clear(dev, LCDC_INT_ENABLE_SET_REG,
726 LCDC_V2_END_OF_FRAME0_INT_ENA);
728 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
731 static void tilcdc_crtc_reset(struct drm_crtc *crtc)
733 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
734 struct drm_device *dev = crtc->dev;
737 drm_atomic_helper_crtc_reset(crtc);
739 /* Turn the raster off if it for some reason is on. */
740 pm_runtime_get_sync(dev->dev);
741 if (tilcdc_read(dev, LCDC_RASTER_CTRL_REG) & LCDC_RASTER_ENABLE) {
742 /* Enable DMA Frame Done Interrupt */
743 tilcdc_write(dev, LCDC_INT_ENABLE_SET_REG, LCDC_FRAME_DONE);
744 tilcdc_clear_irqstatus(dev, 0xffffffff);
746 tilcdc_crtc->frame_done = false;
747 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
749 ret = wait_event_timeout(tilcdc_crtc->frame_done_wq,
750 tilcdc_crtc->frame_done,
751 msecs_to_jiffies(500));
753 dev_err(dev->dev, "%s: timeout waiting for framedone\n",
756 pm_runtime_put_sync(dev->dev);
759 static const struct drm_crtc_funcs tilcdc_crtc_funcs = {
760 .destroy = tilcdc_crtc_destroy,
761 .set_config = drm_atomic_helper_set_config,
762 .page_flip = drm_atomic_helper_page_flip,
763 .reset = tilcdc_crtc_reset,
764 .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state,
765 .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state,
766 .enable_vblank = tilcdc_crtc_enable_vblank,
767 .disable_vblank = tilcdc_crtc_disable_vblank,
770 static enum drm_mode_status
771 tilcdc_crtc_mode_valid(struct drm_crtc *crtc,
772 const struct drm_display_mode *mode)
774 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
775 unsigned int bandwidth;
776 uint32_t hbp, hfp, hsw, vbp, vfp, vsw;
779 * check to see if the width is within the range that
780 * the LCD Controller physically supports
782 if (mode->hdisplay > priv->max_width)
783 return MODE_VIRTUAL_X;
785 /* width must be multiple of 16 */
786 if (mode->hdisplay & 0xf)
787 return MODE_VIRTUAL_X;
789 if (mode->vdisplay > 2048)
790 return MODE_VIRTUAL_Y;
792 DBG("Processing mode %dx%d@%d with pixel clock %d",
793 mode->hdisplay, mode->vdisplay,
794 drm_mode_vrefresh(mode), mode->clock);
796 hbp = mode->htotal - mode->hsync_end;
797 hfp = mode->hsync_start - mode->hdisplay;
798 hsw = mode->hsync_end - mode->hsync_start;
799 vbp = mode->vtotal - mode->vsync_end;
800 vfp = mode->vsync_start - mode->vdisplay;
801 vsw = mode->vsync_end - mode->vsync_start;
803 if ((hbp-1) & ~0x3ff) {
804 DBG("Pruning mode: Horizontal Back Porch out of range");
805 return MODE_HBLANK_WIDE;
808 if ((hfp-1) & ~0x3ff) {
809 DBG("Pruning mode: Horizontal Front Porch out of range");
810 return MODE_HBLANK_WIDE;
813 if ((hsw-1) & ~0x3ff) {
814 DBG("Pruning mode: Horizontal Sync Width out of range");
815 return MODE_HSYNC_WIDE;
819 DBG("Pruning mode: Vertical Back Porch out of range");
820 return MODE_VBLANK_WIDE;
824 DBG("Pruning mode: Vertical Front Porch out of range");
825 return MODE_VBLANK_WIDE;
828 if ((vsw-1) & ~0x3f) {
829 DBG("Pruning mode: Vertical Sync Width out of range");
830 return MODE_VSYNC_WIDE;
834 * some devices have a maximum allowed pixel clock
835 * configured from the DT
837 if (mode->clock > priv->max_pixelclock) {
838 DBG("Pruning mode: pixel clock too high");
839 return MODE_CLOCK_HIGH;
843 * some devices further limit the max horizontal resolution
844 * configured from the DT
846 if (mode->hdisplay > priv->max_width)
847 return MODE_BAD_WIDTH;
849 /* filter out modes that would require too much memory bandwidth: */
850 bandwidth = mode->hdisplay * mode->vdisplay *
851 drm_mode_vrefresh(mode);
852 if (bandwidth > priv->max_bandwidth) {
853 DBG("Pruning mode: exceeds defined bandwidth limit");
860 static const struct drm_crtc_helper_funcs tilcdc_crtc_helper_funcs = {
861 .mode_valid = tilcdc_crtc_mode_valid,
862 .mode_fixup = tilcdc_crtc_mode_fixup,
863 .atomic_check = tilcdc_crtc_atomic_check,
864 .atomic_enable = tilcdc_crtc_atomic_enable,
865 .atomic_disable = tilcdc_crtc_atomic_disable,
866 .atomic_flush = tilcdc_crtc_atomic_flush,
869 void tilcdc_crtc_set_panel_info(struct drm_crtc *crtc,
870 const struct tilcdc_panel_info *info)
872 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
873 tilcdc_crtc->info = info;
876 void tilcdc_crtc_set_simulate_vesa_sync(struct drm_crtc *crtc,
877 bool simulate_vesa_sync)
879 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
881 tilcdc_crtc->simulate_vesa_sync = simulate_vesa_sync;
884 void tilcdc_crtc_update_clk(struct drm_crtc *crtc)
886 struct drm_device *dev = crtc->dev;
887 struct tilcdc_drm_private *priv = dev->dev_private;
888 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
890 drm_modeset_lock(&crtc->mutex, NULL);
891 if (tilcdc_crtc->lcd_fck_rate != clk_get_rate(priv->clk)) {
892 if (tilcdc_crtc_is_on(crtc)) {
893 pm_runtime_get_sync(dev->dev);
894 tilcdc_crtc_disable(crtc);
896 tilcdc_crtc_set_clk(crtc);
898 tilcdc_crtc_enable(crtc);
899 pm_runtime_put_sync(dev->dev);
902 drm_modeset_unlock(&crtc->mutex);
905 #define SYNC_LOST_COUNT_LIMIT 50
907 irqreturn_t tilcdc_crtc_irq(struct drm_crtc *crtc)
909 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
910 struct drm_device *dev = crtc->dev;
911 struct tilcdc_drm_private *priv = dev->dev_private;
914 stat = tilcdc_read_irqstatus(dev);
915 tilcdc_clear_irqstatus(dev, stat);
917 if (stat & LCDC_END_OF_FRAME0) {
918 bool skip_event = false;
923 spin_lock(&tilcdc_crtc->irq_lock);
925 tilcdc_crtc->last_vblank = now;
927 if (tilcdc_crtc->next_fb) {
928 set_scanout(crtc, tilcdc_crtc->next_fb);
929 tilcdc_crtc->next_fb = NULL;
933 spin_unlock(&tilcdc_crtc->irq_lock);
935 drm_crtc_handle_vblank(crtc);
938 struct drm_pending_vblank_event *event;
940 spin_lock(&dev->event_lock);
942 event = tilcdc_crtc->event;
943 tilcdc_crtc->event = NULL;
945 drm_crtc_send_vblank_event(crtc, event);
947 spin_unlock(&dev->event_lock);
950 if (tilcdc_crtc->frame_intact)
951 tilcdc_crtc->sync_lost_count = 0;
953 tilcdc_crtc->frame_intact = true;
956 if (stat & LCDC_FIFO_UNDERFLOW)
957 dev_err_ratelimited(dev->dev, "%s(0x%08x): FIFO underflow",
960 if (stat & LCDC_PL_LOAD_DONE) {
961 complete(&tilcdc_crtc->palette_loaded);
963 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
966 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
970 if (stat & LCDC_SYNC_LOST) {
971 dev_err_ratelimited(dev->dev, "%s(0x%08x): Sync lost",
973 tilcdc_crtc->frame_intact = false;
974 if (priv->rev == 1) {
975 reg = tilcdc_read(dev, LCDC_RASTER_CTRL_REG);
976 if (reg & LCDC_RASTER_ENABLE) {
977 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
979 tilcdc_set(dev, LCDC_RASTER_CTRL_REG,
983 if (tilcdc_crtc->sync_lost_count++ >
984 SYNC_LOST_COUNT_LIMIT) {
986 "%s(0x%08x): Sync lost flood detected, recovering",
988 queue_work(system_wq,
989 &tilcdc_crtc->recover_work);
990 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
992 tilcdc_crtc->sync_lost_count = 0;
997 if (stat & LCDC_FRAME_DONE) {
998 tilcdc_crtc->frame_done = true;
999 wake_up(&tilcdc_crtc->frame_done_wq);
1000 /* rev 1 lcdc appears to hang if irq is not disabled here */
1002 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
1003 LCDC_V1_FRAME_DONE_INT_ENA);
1006 /* For revision 2 only */
1007 if (priv->rev == 2) {
1008 /* Indicate to LCDC that the interrupt service routine has
1009 * completed, see 13.3.6.1.6 in AM335x TRM.
1011 tilcdc_write(dev, LCDC_END_OF_INT_IND_REG, 0);
1017 int tilcdc_crtc_create(struct drm_device *dev)
1019 struct tilcdc_drm_private *priv = dev->dev_private;
1020 struct tilcdc_crtc *tilcdc_crtc;
1021 struct drm_crtc *crtc;
1024 tilcdc_crtc = devm_kzalloc(dev->dev, sizeof(*tilcdc_crtc), GFP_KERNEL);
1028 init_completion(&tilcdc_crtc->palette_loaded);
1029 tilcdc_crtc->palette_base = dmam_alloc_coherent(dev->dev,
1030 TILCDC_PALETTE_SIZE,
1031 &tilcdc_crtc->palette_dma_handle,
1032 GFP_KERNEL | __GFP_ZERO);
1033 if (!tilcdc_crtc->palette_base)
1035 *tilcdc_crtc->palette_base = TILCDC_PALETTE_FIRST_ENTRY;
1037 crtc = &tilcdc_crtc->base;
1039 ret = tilcdc_plane_init(dev, &tilcdc_crtc->primary);
1043 mutex_init(&tilcdc_crtc->enable_lock);
1045 init_waitqueue_head(&tilcdc_crtc->frame_done_wq);
1047 spin_lock_init(&tilcdc_crtc->irq_lock);
1048 INIT_WORK(&tilcdc_crtc->recover_work, tilcdc_crtc_recover_work);
1050 ret = drm_crtc_init_with_planes(dev, crtc,
1051 &tilcdc_crtc->primary,
1058 drm_crtc_helper_add(crtc, &tilcdc_crtc_helper_funcs);
1060 if (priv->is_componentized) {
1061 crtc->port = of_graph_get_port_by_id(dev->dev->of_node, 0);
1062 if (!crtc->port) { /* This should never happen */
1063 dev_err(dev->dev, "Port node not found in %pOF\n",
1074 tilcdc_crtc_destroy(crtc);