split dev_queue
[cor.git] / include / drm / drm_device.h
blob1acfc3bbd3fb2e29c3fe3f2139076b09152708ff
1 #ifndef _DRM_DEVICE_H_
2 #define _DRM_DEVICE_H_
4 #include <linux/list.h>
5 #include <linux/kref.h>
6 #include <linux/mutex.h>
7 #include <linux/idr.h>
9 #include <drm/drm_hashtab.h>
10 #include <drm/drm_mode_config.h>
12 struct drm_driver;
13 struct drm_minor;
14 struct drm_master;
15 struct drm_device_dma;
16 struct drm_vblank_crtc;
17 struct drm_sg_mem;
18 struct drm_local_map;
19 struct drm_vma_offset_manager;
20 struct drm_vram_mm;
21 struct drm_fb_helper;
23 struct inode;
25 struct pci_dev;
26 struct pci_controller;
29 /**
30 * enum drm_switch_power - power state of drm device
33 enum switch_power_state {
34 /** @DRM_SWITCH_POWER_ON: Power state is ON */
35 DRM_SWITCH_POWER_ON = 0,
37 /** @DRM_SWITCH_POWER_OFF: Power state is OFF */
38 DRM_SWITCH_POWER_OFF = 1,
40 /** @DRM_SWITCH_POWER_CHANGING: Power state is changing */
41 DRM_SWITCH_POWER_CHANGING = 2,
43 /** @DRM_SWITCH_POWER_DYNAMIC_OFF: Suspended */
44 DRM_SWITCH_POWER_DYNAMIC_OFF = 3,
47 /**
48 * struct drm_device - DRM device structure
50 * This structure represent a complete card that
51 * may contain multiple heads.
53 struct drm_device {
54 /**
55 * @legacy_dev_list:
57 * List of devices per driver for stealth attach cleanup
59 struct list_head legacy_dev_list;
61 /** @if_version: Highest interface version set */
62 int if_version;
64 /** @ref: Object ref-count */
65 struct kref ref;
67 /** @dev: Device structure of bus-device */
68 struct device *dev;
70 /** @driver: DRM driver managing the device */
71 struct drm_driver *driver;
73 /**
74 * @dev_private:
76 * DRM driver private data. Instead of using this pointer it is
77 * recommended that drivers use drm_dev_init() and embed struct
78 * &drm_device in their larger per-device structure.
80 void *dev_private;
82 /** @primary: Primary node */
83 struct drm_minor *primary;
85 /** @render: Render node */
86 struct drm_minor *render;
88 /**
89 * @registered:
91 * Internally used by drm_dev_register() and drm_connector_register().
93 bool registered;
95 /**
96 * @master:
98 * Currently active master for this device.
99 * Protected by &master_mutex
101 struct drm_master *master;
104 * @driver_features: per-device driver features
106 * Drivers can clear specific flags here to disallow
107 * certain features on a per-device basis while still
108 * sharing a single &struct drm_driver instance across
109 * all devices.
111 u32 driver_features;
114 * @unplugged:
116 * Flag to tell if the device has been unplugged.
117 * See drm_dev_enter() and drm_dev_is_unplugged().
119 bool unplugged;
121 /** @anon_inode: inode for private address-space */
122 struct inode *anon_inode;
124 /** @unique: Unique name of the device */
125 char *unique;
128 * @struct_mutex:
130 * Lock for others (not &drm_minor.master and &drm_file.is_master)
132 struct mutex struct_mutex;
135 * @master_mutex:
137 * Lock for &drm_minor.master and &drm_file.is_master
139 struct mutex master_mutex;
142 * @open_count:
144 * Usage counter for outstanding files open,
145 * protected by drm_global_mutex
147 int open_count;
149 /** @filelist_mutex: Protects @filelist. */
150 struct mutex filelist_mutex;
152 * @filelist:
154 * List of userspace clients, linked through &drm_file.lhead.
156 struct list_head filelist;
159 * @filelist_internal:
161 * List of open DRM files for in-kernel clients.
162 * Protected by &filelist_mutex.
164 struct list_head filelist_internal;
167 * @clientlist_mutex:
169 * Protects &clientlist access.
171 struct mutex clientlist_mutex;
174 * @clientlist:
176 * List of in-kernel clients. Protected by &clientlist_mutex.
178 struct list_head clientlist;
181 * @irq_enabled:
183 * Indicates that interrupt handling is enabled, specifically vblank
184 * handling. Drivers which don't use drm_irq_install() need to set this
185 * to true manually.
187 bool irq_enabled;
190 * @irq: Used by the drm_irq_install() and drm_irq_unistall() helpers.
192 int irq;
195 * @vblank_disable_immediate:
197 * If true, vblank interrupt will be disabled immediately when the
198 * refcount drops to zero, as opposed to via the vblank disable
199 * timer.
201 * This can be set to true it the hardware has a working vblank counter
202 * with high-precision timestamping (otherwise there are races) and the
203 * driver uses drm_crtc_vblank_on() and drm_crtc_vblank_off()
204 * appropriately. See also @max_vblank_count and
205 * &drm_crtc_funcs.get_vblank_counter.
207 bool vblank_disable_immediate;
210 * @vblank:
212 * Array of vblank tracking structures, one per &struct drm_crtc. For
213 * historical reasons (vblank support predates kernel modesetting) this
214 * is free-standing and not part of &struct drm_crtc itself. It must be
215 * initialized explicitly by calling drm_vblank_init().
217 struct drm_vblank_crtc *vblank;
220 * @vblank_time_lock:
222 * Protects vblank count and time updates during vblank enable/disable
224 spinlock_t vblank_time_lock;
226 * @vbl_lock: Top-level vblank references lock, wraps the low-level
227 * @vblank_time_lock.
229 spinlock_t vbl_lock;
232 * @max_vblank_count:
234 * Maximum value of the vblank registers. This value +1 will result in a
235 * wrap-around of the vblank register. It is used by the vblank core to
236 * handle wrap-arounds.
238 * If set to zero the vblank core will try to guess the elapsed vblanks
239 * between times when the vblank interrupt is disabled through
240 * high-precision timestamps. That approach is suffering from small
241 * races and imprecision over longer time periods, hence exposing a
242 * hardware vblank counter is always recommended.
244 * This is the statically configured device wide maximum. The driver
245 * can instead choose to use a runtime configurable per-crtc value
246 * &drm_vblank_crtc.max_vblank_count, in which case @max_vblank_count
247 * must be left at zero. See drm_crtc_set_max_vblank_count() on how
248 * to use the per-crtc value.
250 * If non-zero, &drm_crtc_funcs.get_vblank_counter must be set.
252 u32 max_vblank_count;
254 /** @vblank_event_list: List of vblank events */
255 struct list_head vblank_event_list;
258 * @event_lock:
260 * Protects @vblank_event_list and event delivery in
261 * general. See drm_send_event() and drm_send_event_locked().
263 spinlock_t event_lock;
265 /** @agp: AGP data */
266 struct drm_agp_head *agp;
268 /** @pdev: PCI device structure */
269 struct pci_dev *pdev;
271 #ifdef __alpha__
272 /** @hose: PCI hose, only used on ALPHA platforms. */
273 struct pci_controller *hose;
274 #endif
275 /** @num_crtcs: Number of CRTCs on this device */
276 unsigned int num_crtcs;
278 /** @mode_config: Current mode config */
279 struct drm_mode_config mode_config;
281 /** @object_name_lock: GEM information */
282 struct mutex object_name_lock;
284 /** @object_name_idr: GEM information */
285 struct idr object_name_idr;
287 /** @vma_offset_manager: GEM information */
288 struct drm_vma_offset_manager *vma_offset_manager;
290 /** @vram_mm: VRAM MM memory manager */
291 struct drm_vram_mm *vram_mm;
294 * @switch_power_state:
296 * Power state of the client.
297 * Used by drivers supporting the switcheroo driver.
298 * The state is maintained in the
299 * &vga_switcheroo_client_ops.set_gpu_state callback
301 enum switch_power_state switch_power_state;
304 * @fb_helper:
306 * Pointer to the fbdev emulation structure.
307 * Set by drm_fb_helper_init() and cleared by drm_fb_helper_fini().
309 struct drm_fb_helper *fb_helper;
311 /* Everything below here is for legacy driver, never use! */
312 /* private: */
313 #if IS_ENABLED(CONFIG_DRM_LEGACY)
314 /* Context handle management - linked list of context handles */
315 struct list_head ctxlist;
317 /* Context handle management - mutex for &ctxlist */
318 struct mutex ctxlist_mutex;
320 /* Context handle management */
321 struct idr ctx_idr;
323 /* Memory management - linked list of regions */
324 struct list_head maplist;
326 /* Memory management - user token hash table for maps */
327 struct drm_open_hash map_hash;
329 /* Context handle management - list of vmas (for debugging) */
330 struct list_head vmalist;
332 /* Optional pointer for DMA support */
333 struct drm_device_dma *dma;
335 /* Context swapping flag */
336 __volatile__ long context_flag;
338 /* Last current context */
339 int last_context;
341 /* Lock for &buf_use and a few other things. */
342 spinlock_t buf_lock;
344 /* Usage counter for buffers in use -- cannot alloc */
345 int buf_use;
347 /* Buffer allocation in progress */
348 atomic_t buf_alloc;
350 struct {
351 int context;
352 struct drm_hw_lock *lock;
353 } sigdata;
355 struct drm_local_map *agp_buffer_map;
356 unsigned int agp_buffer_token;
358 /* Scatter gather memory */
359 struct drm_sg_mem *sg;
360 #endif
363 #endif