2 * Copyright (C) 2008 Maarten Maathuis.
5 * Permission is hereby granted, free of charge, to any person obtaining
6 * a copy of this software and associated documentation files (the
7 * "Software"), to deal in the Software without restriction, including
8 * without limitation the rights to use, copy, modify, merge, publish,
9 * distribute, sublicense, and/or sell copies of the Software, and to
10 * permit persons to whom the Software is furnished to do so, subject to
11 * the following conditions:
13 * The above copyright notice and this permission notice (including the
14 * next paragraph) shall be included in all copies or substantial
15 * portions of the Software.
17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
18 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
19 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
20 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
21 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
22 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
23 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
28 #include "drm_crtc_helper.h"
30 #include "nouveau_fbcon.h"
31 #include "nouveau_hw.h"
32 #include "nouveau_crtc.h"
33 #include "nouveau_dma.h"
34 #include "nouveau_gem.h"
35 #include "nouveau_connector.h"
36 #include "nv50_display.h"
38 #include "nouveau_fence.h"
40 #include <subdev/bios/gpio.h>
41 #include <subdev/gpio.h>
42 #include <engine/disp.h>
45 nouveau_user_framebuffer_destroy(struct drm_framebuffer
*drm_fb
)
47 struct nouveau_framebuffer
*fb
= nouveau_framebuffer(drm_fb
);
50 drm_gem_object_unreference_unlocked(fb
->nvbo
->gem
);
52 drm_framebuffer_cleanup(drm_fb
);
57 nouveau_user_framebuffer_create_handle(struct drm_framebuffer
*drm_fb
,
58 struct drm_file
*file_priv
,
61 struct nouveau_framebuffer
*fb
= nouveau_framebuffer(drm_fb
);
63 return drm_gem_handle_create(file_priv
, fb
->nvbo
->gem
, handle
);
66 static const struct drm_framebuffer_funcs nouveau_framebuffer_funcs
= {
67 .destroy
= nouveau_user_framebuffer_destroy
,
68 .create_handle
= nouveau_user_framebuffer_create_handle
,
72 nouveau_framebuffer_init(struct drm_device
*dev
,
73 struct nouveau_framebuffer
*nv_fb
,
74 struct drm_mode_fb_cmd2
*mode_cmd
,
75 struct nouveau_bo
*nvbo
)
77 struct nouveau_drm
*drm
= nouveau_drm(dev
);
78 struct drm_framebuffer
*fb
= &nv_fb
->base
;
81 ret
= drm_framebuffer_init(dev
, fb
, &nouveau_framebuffer_funcs
);
86 drm_helper_mode_fill_fb_struct(fb
, mode_cmd
);
89 if (nv_device(drm
->device
)->card_type
>= NV_50
) {
90 u32 tile_flags
= nouveau_bo_tile_layout(nvbo
);
91 if (tile_flags
== 0x7a00 ||
93 nv_fb
->r_dma
= NvEvoFB32
;
95 if (tile_flags
== 0x7000)
96 nv_fb
->r_dma
= NvEvoFB16
;
98 nv_fb
->r_dma
= NvEvoVRAM_LP
;
101 case 8: nv_fb
->r_format
= NV50_EVO_CRTC_FB_DEPTH_8
; break;
102 case 15: nv_fb
->r_format
= NV50_EVO_CRTC_FB_DEPTH_15
; break;
103 case 16: nv_fb
->r_format
= NV50_EVO_CRTC_FB_DEPTH_16
; break;
105 case 32: nv_fb
->r_format
= NV50_EVO_CRTC_FB_DEPTH_24
; break;
106 case 30: nv_fb
->r_format
= NV50_EVO_CRTC_FB_DEPTH_30
; break;
108 NV_ERROR(drm
, "unknown depth %d\n", fb
->depth
);
112 if (nv_device(drm
->device
)->chipset
== 0x50)
113 nv_fb
->r_format
|= (tile_flags
<< 8);
116 if (nv_device(drm
->device
)->card_type
< NV_D0
)
117 nv_fb
->r_pitch
= 0x00100000 | fb
->pitches
[0];
119 nv_fb
->r_pitch
= 0x01000000 | fb
->pitches
[0];
121 u32 mode
= nvbo
->tile_mode
;
122 if (nv_device(drm
->device
)->card_type
>= NV_C0
)
124 nv_fb
->r_pitch
= ((fb
->pitches
[0] / 4) << 4) | mode
;
131 static struct drm_framebuffer
*
132 nouveau_user_framebuffer_create(struct drm_device
*dev
,
133 struct drm_file
*file_priv
,
134 struct drm_mode_fb_cmd2
*mode_cmd
)
136 struct nouveau_framebuffer
*nouveau_fb
;
137 struct drm_gem_object
*gem
;
140 gem
= drm_gem_object_lookup(dev
, file_priv
, mode_cmd
->handles
[0]);
142 return ERR_PTR(-ENOENT
);
144 nouveau_fb
= kzalloc(sizeof(struct nouveau_framebuffer
), GFP_KERNEL
);
146 return ERR_PTR(-ENOMEM
);
148 ret
= nouveau_framebuffer_init(dev
, nouveau_fb
, mode_cmd
, nouveau_gem_object(gem
));
150 drm_gem_object_unreference(gem
);
154 return &nouveau_fb
->base
;
157 static const struct drm_mode_config_funcs nouveau_mode_config_funcs
= {
158 .fb_create
= nouveau_user_framebuffer_create
,
159 .output_poll_changed
= nouveau_fbcon_output_poll_changed
,
163 struct nouveau_drm_prop_enum_list
{
169 static struct nouveau_drm_prop_enum_list underscan
[] = {
170 { 6, UNDERSCAN_AUTO
, "auto" },
171 { 6, UNDERSCAN_OFF
, "off" },
172 { 6, UNDERSCAN_ON
, "on" },
176 static struct nouveau_drm_prop_enum_list dither_mode
[] = {
177 { 7, DITHERING_MODE_AUTO
, "auto" },
178 { 7, DITHERING_MODE_OFF
, "off" },
179 { 1, DITHERING_MODE_ON
, "on" },
180 { 6, DITHERING_MODE_STATIC2X2
, "static 2x2" },
181 { 6, DITHERING_MODE_DYNAMIC2X2
, "dynamic 2x2" },
182 { 4, DITHERING_MODE_TEMPORAL
, "temporal" },
186 static struct nouveau_drm_prop_enum_list dither_depth
[] = {
187 { 6, DITHERING_DEPTH_AUTO
, "auto" },
188 { 6, DITHERING_DEPTH_6BPC
, "6 bpc" },
189 { 6, DITHERING_DEPTH_8BPC
, "8 bpc" },
193 #define PROP_ENUM(p,gen,n,list) do { \
194 struct nouveau_drm_prop_enum_list *l = (list); \
196 while (l->gen_mask) { \
197 if (l->gen_mask & (1 << (gen))) \
202 p = drm_property_create(dev, DRM_MODE_PROP_ENUM, n, c); \
205 while (p && l->gen_mask) { \
206 if (l->gen_mask & (1 << (gen))) { \
207 drm_property_add_enum(p, c, l->type, l->name); \
216 nouveau_display_init(struct drm_device
*dev
)
218 struct nouveau_drm
*drm
= nouveau_drm(dev
);
219 struct nouveau_display
*disp
= nouveau_display(dev
);
220 struct nouveau_gpio
*gpio
= nouveau_gpio(drm
->device
);
221 struct drm_connector
*connector
;
224 ret
= disp
->init(dev
);
228 /* power on internal panel if it's not already. the init tables of
229 * some vbios default this to off for some reason, causing the
230 * panel to not work after resume
232 if (gpio
&& gpio
->get(gpio
, 0, DCB_GPIO_PANEL_POWER
, 0xff) == 0) {
233 gpio
->set(gpio
, 0, DCB_GPIO_PANEL_POWER
, 0xff, 1);
237 /* enable polling for external displays */
238 drm_kms_helper_poll_enable(dev
);
240 /* enable hotplug interrupts */
241 list_for_each_entry(connector
, &dev
->mode_config
.connector_list
, head
) {
242 struct nouveau_connector
*conn
= nouveau_connector(connector
);
244 gpio
->irq(gpio
, 0, conn
->hpd
, 0xff, true);
251 nouveau_display_fini(struct drm_device
*dev
)
253 struct nouveau_drm
*drm
= nouveau_drm(dev
);
254 struct nouveau_display
*disp
= nouveau_display(dev
);
255 struct nouveau_gpio
*gpio
= nouveau_gpio(drm
->device
);
256 struct drm_connector
*connector
;
258 /* disable hotplug interrupts */
259 list_for_each_entry(connector
, &dev
->mode_config
.connector_list
, head
) {
260 struct nouveau_connector
*conn
= nouveau_connector(connector
);
262 gpio
->irq(gpio
, 0, conn
->hpd
, 0xff, false);
265 drm_kms_helper_poll_disable(dev
);
270 nouveau_display_vblank_notify(void *data
, int crtc
)
272 drm_handle_vblank(data
, crtc
);
276 nouveau_display_vblank_get(void *data
, int crtc
)
278 drm_vblank_get(data
, crtc
);
282 nouveau_display_vblank_put(void *data
, int crtc
)
284 drm_vblank_put(data
, crtc
);
288 nouveau_display_create(struct drm_device
*dev
)
290 struct nouveau_drm
*drm
= nouveau_drm(dev
);
291 struct nouveau_disp
*pdisp
= nouveau_disp(drm
->device
);
292 struct nouveau_display
*disp
;
295 disp
= drm
->display
= kzalloc(sizeof(*disp
), GFP_KERNEL
);
299 pdisp
->vblank
.data
= dev
;
300 pdisp
->vblank
.notify
= nouveau_display_vblank_notify
;
301 pdisp
->vblank
.get
= nouveau_display_vblank_get
;
302 pdisp
->vblank
.put
= nouveau_display_vblank_put
;
304 drm_mode_config_init(dev
);
305 drm_mode_create_scaling_mode_property(dev
);
306 drm_mode_create_dvi_i_properties(dev
);
308 if (nv_device(drm
->device
)->card_type
< NV_50
)
311 if (nv_device(drm
->device
)->card_type
< NV_D0
)
316 PROP_ENUM(disp
->dithering_mode
, gen
, "dithering mode", dither_mode
);
317 PROP_ENUM(disp
->dithering_depth
, gen
, "dithering depth", dither_depth
);
318 PROP_ENUM(disp
->underscan_property
, gen
, "underscan", underscan
);
320 disp
->underscan_hborder_property
=
321 drm_property_create_range(dev
, 0, "underscan hborder", 0, 128);
323 disp
->underscan_vborder_property
=
324 drm_property_create_range(dev
, 0, "underscan vborder", 0, 128);
327 disp
->vibrant_hue_property
=
328 drm_property_create(dev
, DRM_MODE_PROP_RANGE
,
330 disp
->vibrant_hue_property
->values
[0] = 0;
331 disp
->vibrant_hue_property
->values
[1] = 180; /* -90..+90 */
333 disp
->color_vibrance_property
=
334 drm_property_create(dev
, DRM_MODE_PROP_RANGE
,
335 "color vibrance", 2);
336 disp
->color_vibrance_property
->values
[0] = 0;
337 disp
->color_vibrance_property
->values
[1] = 200; /* -100..+100 */
340 dev
->mode_config
.funcs
= &nouveau_mode_config_funcs
;
341 dev
->mode_config
.fb_base
= pci_resource_start(dev
->pdev
, 1);
343 dev
->mode_config
.min_width
= 0;
344 dev
->mode_config
.min_height
= 0;
345 if (nv_device(drm
->device
)->card_type
< NV_10
) {
346 dev
->mode_config
.max_width
= 2048;
347 dev
->mode_config
.max_height
= 2048;
349 if (nv_device(drm
->device
)->card_type
< NV_50
) {
350 dev
->mode_config
.max_width
= 4096;
351 dev
->mode_config
.max_height
= 4096;
353 dev
->mode_config
.max_width
= 8192;
354 dev
->mode_config
.max_height
= 8192;
357 dev
->mode_config
.preferred_depth
= 24;
358 dev
->mode_config
.prefer_shadow
= 1;
360 drm_kms_helper_poll_init(dev
);
361 drm_kms_helper_poll_disable(dev
);
363 if (nv_device(drm
->device
)->card_type
< NV_50
)
364 ret
= nv04_display_create(dev
);
366 if (nv_device(drm
->device
)->card_type
< NV_D0
)
367 ret
= nv50_display_create(dev
);
369 ret
= nvd0_display_create(dev
);
371 goto disp_create_err
;
373 if (dev
->mode_config
.num_crtc
) {
374 ret
= drm_vblank_init(dev
, dev
->mode_config
.num_crtc
);
379 nouveau_backlight_init(dev
);
385 drm_kms_helper_poll_fini(dev
);
386 drm_mode_config_cleanup(dev
);
391 nouveau_display_destroy(struct drm_device
*dev
)
393 struct nouveau_display
*disp
= nouveau_display(dev
);
395 nouveau_backlight_exit(dev
);
396 drm_vblank_cleanup(dev
);
400 drm_kms_helper_poll_fini(dev
);
401 drm_mode_config_cleanup(dev
);
402 nouveau_drm(dev
)->display
= NULL
;
407 nouveau_display_suspend(struct drm_device
*dev
)
409 struct nouveau_drm
*drm
= nouveau_drm(dev
);
410 struct drm_crtc
*crtc
;
412 nouveau_display_fini(dev
);
414 NV_INFO(drm
, "unpinning framebuffer(s)...\n");
415 list_for_each_entry(crtc
, &dev
->mode_config
.crtc_list
, head
) {
416 struct nouveau_framebuffer
*nouveau_fb
;
418 nouveau_fb
= nouveau_framebuffer(crtc
->fb
);
419 if (!nouveau_fb
|| !nouveau_fb
->nvbo
)
422 nouveau_bo_unpin(nouveau_fb
->nvbo
);
425 list_for_each_entry(crtc
, &dev
->mode_config
.crtc_list
, head
) {
426 struct nouveau_crtc
*nv_crtc
= nouveau_crtc(crtc
);
428 nouveau_bo_unmap(nv_crtc
->cursor
.nvbo
);
429 nouveau_bo_unpin(nv_crtc
->cursor
.nvbo
);
436 nouveau_display_resume(struct drm_device
*dev
)
438 struct nouveau_drm
*drm
= nouveau_drm(dev
);
439 struct drm_crtc
*crtc
;
442 list_for_each_entry(crtc
, &dev
->mode_config
.crtc_list
, head
) {
443 struct nouveau_framebuffer
*nouveau_fb
;
445 nouveau_fb
= nouveau_framebuffer(crtc
->fb
);
446 if (!nouveau_fb
|| !nouveau_fb
->nvbo
)
449 nouveau_bo_pin(nouveau_fb
->nvbo
, TTM_PL_FLAG_VRAM
);
452 list_for_each_entry(crtc
, &dev
->mode_config
.crtc_list
, head
) {
453 struct nouveau_crtc
*nv_crtc
= nouveau_crtc(crtc
);
455 ret
= nouveau_bo_pin(nv_crtc
->cursor
.nvbo
, TTM_PL_FLAG_VRAM
);
457 ret
= nouveau_bo_map(nv_crtc
->cursor
.nvbo
);
459 NV_ERROR(drm
, "Could not pin/map cursor.\n");
462 nouveau_fbcon_set_suspend(dev
, 0);
463 nouveau_fbcon_zfill_all(dev
);
465 nouveau_display_init(dev
);
467 /* Force CLUT to get re-loaded during modeset */
468 list_for_each_entry(crtc
, &dev
->mode_config
.crtc_list
, head
) {
469 struct nouveau_crtc
*nv_crtc
= nouveau_crtc(crtc
);
471 nv_crtc
->lut
.depth
= 0;
474 drm_helper_resume_force_mode(dev
);
476 list_for_each_entry(crtc
, &dev
->mode_config
.crtc_list
, head
) {
477 struct nouveau_crtc
*nv_crtc
= nouveau_crtc(crtc
);
478 u32 offset
= nv_crtc
->cursor
.nvbo
->bo
.offset
;
480 nv_crtc
->cursor
.set_offset(nv_crtc
, offset
);
481 nv_crtc
->cursor
.set_pos(nv_crtc
, nv_crtc
->cursor_saved_x
,
482 nv_crtc
->cursor_saved_y
);
487 nouveau_vblank_enable(struct drm_device
*dev
, int crtc
)
489 struct nouveau_device
*device
= nouveau_dev(dev
);
491 if (device
->card_type
>= NV_D0
)
492 nv_mask(device
, 0x6100c0 + (crtc
* 0x800), 1, 1);
494 if (device
->card_type
>= NV_50
)
495 nv_mask(device
, NV50_PDISPLAY_INTR_EN_1
, 0,
496 NV50_PDISPLAY_INTR_EN_1_VBLANK_CRTC_(crtc
));
498 NVWriteCRTC(dev
, crtc
, NV_PCRTC_INTR_EN_0
,
499 NV_PCRTC_INTR_0_VBLANK
);
505 nouveau_vblank_disable(struct drm_device
*dev
, int crtc
)
507 struct nouveau_device
*device
= nouveau_dev(dev
);
509 if (device
->card_type
>= NV_D0
)
510 nv_mask(device
, 0x6100c0 + (crtc
* 0x800), 1, 0);
512 if (device
->card_type
>= NV_50
)
513 nv_mask(device
, NV50_PDISPLAY_INTR_EN_1
,
514 NV50_PDISPLAY_INTR_EN_1_VBLANK_CRTC_(crtc
), 0);
516 NVWriteCRTC(dev
, crtc
, NV_PCRTC_INTR_EN_0
, 0);
520 nouveau_page_flip_reserve(struct nouveau_bo
*old_bo
,
521 struct nouveau_bo
*new_bo
)
525 ret
= nouveau_bo_pin(new_bo
, TTM_PL_FLAG_VRAM
);
529 ret
= ttm_bo_reserve(&new_bo
->bo
, false, false, false, 0);
533 if (likely(old_bo
!= new_bo
)) {
534 ret
= ttm_bo_reserve(&old_bo
->bo
, false, false, false, 0);
542 ttm_bo_unreserve(&new_bo
->bo
);
544 nouveau_bo_unpin(new_bo
);
549 nouveau_page_flip_unreserve(struct nouveau_bo
*old_bo
,
550 struct nouveau_bo
*new_bo
,
551 struct nouveau_fence
*fence
)
553 nouveau_bo_fence(new_bo
, fence
);
554 ttm_bo_unreserve(&new_bo
->bo
);
556 if (likely(old_bo
!= new_bo
)) {
557 nouveau_bo_fence(old_bo
, fence
);
558 ttm_bo_unreserve(&old_bo
->bo
);
561 nouveau_bo_unpin(old_bo
);
565 nouveau_page_flip_emit(struct nouveau_channel
*chan
,
566 struct nouveau_bo
*old_bo
,
567 struct nouveau_bo
*new_bo
,
568 struct nouveau_page_flip_state
*s
,
569 struct nouveau_fence
**pfence
)
571 struct nouveau_fence_chan
*fctx
= chan
->fence
;
572 struct nouveau_drm
*drm
= chan
->drm
;
573 struct drm_device
*dev
= drm
->dev
;
577 /* Queue it to the pending list */
578 spin_lock_irqsave(&dev
->event_lock
, flags
);
579 list_add_tail(&s
->head
, &fctx
->flip
);
580 spin_unlock_irqrestore(&dev
->event_lock
, flags
);
582 /* Synchronize with the old framebuffer */
583 ret
= nouveau_fence_sync(old_bo
->bo
.sync_obj
, chan
);
587 /* Emit the pageflip */
588 ret
= RING_SPACE(chan
, 3);
592 if (nv_device(drm
->device
)->card_type
< NV_C0
) {
593 BEGIN_NV04(chan
, NvSubSw
, NV_SW_PAGE_FLIP
, 1);
594 OUT_RING (chan
, 0x00000000);
595 OUT_RING (chan
, 0x00000000);
597 BEGIN_NVC0(chan
, 0, NV10_SUBCHAN_REF_CNT
, 1);
599 BEGIN_IMC0(chan
, 0, NVSW_SUBCHAN_PAGE_FLIP
, 0x0000);
603 ret
= nouveau_fence_new(chan
, pfence
);
609 spin_lock_irqsave(&dev
->event_lock
, flags
);
611 spin_unlock_irqrestore(&dev
->event_lock
, flags
);
616 nouveau_crtc_page_flip(struct drm_crtc
*crtc
, struct drm_framebuffer
*fb
,
617 struct drm_pending_vblank_event
*event
)
619 struct drm_device
*dev
= crtc
->dev
;
620 struct nouveau_drm
*drm
= nouveau_drm(dev
);
621 struct nouveau_bo
*old_bo
= nouveau_framebuffer(crtc
->fb
)->nvbo
;
622 struct nouveau_bo
*new_bo
= nouveau_framebuffer(fb
)->nvbo
;
623 struct nouveau_page_flip_state
*s
;
624 struct nouveau_channel
*chan
= NULL
;
625 struct nouveau_fence
*fence
;
631 s
= kzalloc(sizeof(*s
), GFP_KERNEL
);
635 /* Don't let the buffers go away while we flip */
636 ret
= nouveau_page_flip_reserve(old_bo
, new_bo
);
640 /* Initialize a page flip struct */
641 *s
= (struct nouveau_page_flip_state
)
642 { { }, event
, nouveau_crtc(crtc
)->index
,
643 fb
->bits_per_pixel
, fb
->pitches
[0], crtc
->x
, crtc
->y
,
646 /* Choose the channel the flip will be handled in */
647 fence
= new_bo
->bo
.sync_obj
;
649 chan
= fence
->channel
;
652 mutex_lock(&chan
->cli
->mutex
);
654 /* Emit a page flip */
655 if (nv_device(drm
->device
)->card_type
>= NV_50
) {
656 if (nv_device(drm
->device
)->card_type
>= NV_D0
)
657 ret
= nvd0_display_flip_next(crtc
, fb
, chan
, 0);
659 ret
= nv50_display_flip_next(crtc
, fb
, chan
);
661 mutex_unlock(&chan
->cli
->mutex
);
666 ret
= nouveau_page_flip_emit(chan
, old_bo
, new_bo
, s
, &fence
);
667 mutex_unlock(&chan
->cli
->mutex
);
671 /* Update the crtc struct and cleanup */
674 nouveau_page_flip_unreserve(old_bo
, new_bo
, fence
);
675 nouveau_fence_unref(&fence
);
679 nouveau_page_flip_unreserve(old_bo
, new_bo
, NULL
);
686 nouveau_finish_page_flip(struct nouveau_channel
*chan
,
687 struct nouveau_page_flip_state
*ps
)
689 struct nouveau_fence_chan
*fctx
= chan
->fence
;
690 struct nouveau_drm
*drm
= chan
->drm
;
691 struct drm_device
*dev
= drm
->dev
;
692 struct nouveau_page_flip_state
*s
;
695 spin_lock_irqsave(&dev
->event_lock
, flags
);
697 if (list_empty(&fctx
->flip
)) {
698 NV_ERROR(drm
, "unexpected pageflip\n");
699 spin_unlock_irqrestore(&dev
->event_lock
, flags
);
703 s
= list_first_entry(&fctx
->flip
, struct nouveau_page_flip_state
, head
);
705 struct drm_pending_vblank_event
*e
= s
->event
;
708 do_gettimeofday(&now
);
709 e
->event
.sequence
= 0;
710 e
->event
.tv_sec
= now
.tv_sec
;
711 e
->event
.tv_usec
= now
.tv_usec
;
712 list_add_tail(&e
->base
.link
, &e
->base
.file_priv
->event_list
);
713 wake_up_interruptible(&e
->base
.file_priv
->event_wait
);
721 spin_unlock_irqrestore(&dev
->event_lock
, flags
);
726 nouveau_flip_complete(void *data
)
728 struct nouveau_channel
*chan
= data
;
729 struct nouveau_drm
*drm
= chan
->drm
;
730 struct nouveau_page_flip_state state
;
732 if (!nouveau_finish_page_flip(chan
, &state
)) {
733 if (nv_device(drm
->device
)->card_type
< NV_50
) {
734 nv_set_crtc_base(drm
->dev
, state
.crtc
, state
.offset
+
735 state
.y
* state
.pitch
+
736 state
.x
* state
.bpp
/ 8);
744 nouveau_display_dumb_create(struct drm_file
*file_priv
, struct drm_device
*dev
,
745 struct drm_mode_create_dumb
*args
)
747 struct nouveau_bo
*bo
;
750 args
->pitch
= roundup(args
->width
* (args
->bpp
/ 8), 256);
751 args
->size
= args
->pitch
* args
->height
;
752 args
->size
= roundup(args
->size
, PAGE_SIZE
);
754 ret
= nouveau_gem_new(dev
, args
->size
, 0, NOUVEAU_GEM_DOMAIN_VRAM
, 0, 0, &bo
);
758 ret
= drm_gem_handle_create(file_priv
, bo
->gem
, &args
->handle
);
759 drm_gem_object_unreference_unlocked(bo
->gem
);
764 nouveau_display_dumb_destroy(struct drm_file
*file_priv
, struct drm_device
*dev
,
767 return drm_gem_handle_delete(file_priv
, handle
);
771 nouveau_display_dumb_map_offset(struct drm_file
*file_priv
,
772 struct drm_device
*dev
,
773 uint32_t handle
, uint64_t *poffset
)
775 struct drm_gem_object
*gem
;
777 gem
= drm_gem_object_lookup(dev
, file_priv
, handle
);
779 struct nouveau_bo
*bo
= gem
->driver_private
;
780 *poffset
= bo
->bo
.addr_space_offset
;
781 drm_gem_object_unreference_unlocked(gem
);