/* SPDX-License-Identifier: MIT */ /* * Copyright (C) 2013-2017 Oracle Corporation * This file is based on ast_drv.h * Copyright 2012 Red Hat Inc. * Authors: Dave Airlie * Michael Thayer */ #ifndef __VBOX_DRV_H__ #define __VBOX_DRV_H__ #include #include #include #include #include #include #include #include #include #include #include #include #include #include "vboxvideo_guest.h" #include "vboxvideo_vbe.h" #include "hgsmi_ch_setup.h" #define DRIVER_NAME "vboxvideo" #define DRIVER_DESC "Oracle VM VirtualBox Graphics Card" #define DRIVER_DATE "20130823" #define DRIVER_MAJOR 1 #define DRIVER_MINOR 0 #define DRIVER_PATCHLEVEL 0 #define VBOX_MAX_CURSOR_WIDTH 64 #define VBOX_MAX_CURSOR_HEIGHT 64 #define CURSOR_PIXEL_COUNT (VBOX_MAX_CURSOR_WIDTH * VBOX_MAX_CURSOR_HEIGHT) #define CURSOR_DATA_SIZE (CURSOR_PIXEL_COUNT * 4 + CURSOR_PIXEL_COUNT / 8) #define VBOX_MAX_SCREENS 32 #define GUEST_HEAP_OFFSET(vbox) ((vbox)->full_vram_size - \ VBVA_ADAPTER_INFORMATION_SIZE) #define GUEST_HEAP_SIZE VBVA_ADAPTER_INFORMATION_SIZE #define GUEST_HEAP_USABLE_SIZE (VBVA_ADAPTER_INFORMATION_SIZE - \ sizeof(struct hgsmi_host_flags)) #define HOST_FLAGS_OFFSET GUEST_HEAP_USABLE_SIZE struct vbox_framebuffer { struct drm_framebuffer base; struct drm_gem_object *obj; }; struct vbox_private { /* Must be first; or we must define our own release callback */ struct drm_device ddev; struct drm_fb_helper fb_helper; struct vbox_framebuffer afb; u8 __iomem *guest_heap; u8 __iomem *vbva_buffers; struct gen_pool *guest_pool; struct vbva_buf_ctx *vbva_info; bool any_pitch; u32 num_crtcs; /* Amount of available VRAM, including space used for buffers. */ u32 full_vram_size; /* Amount of available VRAM, not including space used for buffers. */ u32 available_vram_size; /* Array of structures for receiving mode hints. */ struct vbva_modehint *last_mode_hints; int fb_mtrr; struct { struct ttm_bo_device bdev; } ttm; struct mutex hw_mutex; /* protects modeset and accel/vbva accesses */ /* * We decide whether or not user-space supports display hot-plug * depending on whether they react to a hot-plug event after the initial * mode query. */ bool initial_mode_queried; struct work_struct hotplug_work; u32 input_mapping_width; u32 input_mapping_height; /* * Is user-space using an X.Org-style layout of one large frame-buffer * encompassing all screen ones or is the fbdev console active? */ bool single_framebuffer; u8 cursor_data[CURSOR_DATA_SIZE]; }; #undef CURSOR_PIXEL_COUNT #undef CURSOR_DATA_SIZE struct vbox_gem_object; struct vbox_connector { struct drm_connector base; char name[32]; struct vbox_crtc *vbox_crtc; struct { u32 width; u32 height; bool disconnected; } mode_hint; }; struct vbox_crtc { struct drm_crtc base; bool disconnected; unsigned int crtc_id; u32 fb_offset; bool cursor_enabled; u32 x_hint; u32 y_hint; /* * When setting a mode we not only pass the mode to the hypervisor, * but also information on how to map / translate input coordinates * for the emulated USB tablet. This input-mapping may change when * the mode on *another* crtc changes. * * This means that sometimes we must do a modeset on other crtc-s then * the one being changed to update the input-mapping. Including crtc-s * which may be disabled inside the guest (shown as a black window * on the host unless closed by the user). * * With atomic modesetting the mode-info of disabled crtcs gets zeroed * yet we need it when updating the input-map to avoid resizing the * window as a side effect of a mode_set on another crtc. Therefor we * cache the info of the last mode below. */ u32 width; u32 height; u32 x; u32 y; }; struct vbox_encoder { struct drm_encoder base; }; #define to_vbox_crtc(x) container_of(x, struct vbox_crtc, base) #define to_vbox_connector(x) container_of(x, struct vbox_connector, base) #define to_vbox_encoder(x) container_of(x, struct vbox_encoder, base) #define to_vbox_framebuffer(x) container_of(x, struct vbox_framebuffer, base) bool vbox_check_supported(u16 id); int vbox_hw_init(struct vbox_private *vbox); void vbox_hw_fini(struct vbox_private *vbox); int vbox_mode_init(struct vbox_private *vbox); void vbox_mode_fini(struct vbox_private *vbox); void vbox_report_caps(struct vbox_private *vbox); void vbox_framebuffer_dirty_rectangles(struct drm_framebuffer *fb, struct drm_clip_rect *rects, unsigned int num_rects); int vbox_framebuffer_init(struct vbox_private *vbox, struct vbox_framebuffer *vbox_fb, const struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_object *obj); int vboxfb_create(struct drm_fb_helper *helper, struct drm_fb_helper_surface_size *sizes); void vbox_fbdev_fini(struct vbox_private *vbox); struct vbox_bo { struct ttm_buffer_object bo; struct ttm_placement placement; struct ttm_bo_kmap_obj kmap; struct drm_gem_object gem; struct ttm_place placements[3]; int pin_count; }; #define gem_to_vbox_bo(gobj) container_of((gobj), struct vbox_bo, gem) static inline struct vbox_bo *vbox_bo(struct ttm_buffer_object *bo) { return container_of(bo, struct vbox_bo, bo); } #define to_vbox_obj(x) container_of(x, struct vbox_gem_object, base) static inline u64 vbox_bo_gpu_offset(struct vbox_bo *bo) { return bo->bo.offset; } int vbox_dumb_create(struct drm_file *file, struct drm_device *dev, struct drm_mode_create_dumb *args); void vbox_gem_free_object(struct drm_gem_object *obj); int vbox_dumb_mmap_offset(struct drm_file *file, struct drm_device *dev, u32 handle, u64 *offset); #define DRM_FILE_PAGE_OFFSET (0x10000000ULL >> PAGE_SHIFT) int vbox_mm_init(struct vbox_private *vbox); void vbox_mm_fini(struct vbox_private *vbox); int vbox_bo_create(struct vbox_private *vbox, int size, int align, u32 flags, struct vbox_bo **pvboxbo); int vbox_gem_create(struct vbox_private *vbox, u32 size, bool iskernel, struct drm_gem_object **obj); int vbox_bo_pin(struct vbox_bo *bo, u32 pl_flag); int vbox_bo_unpin(struct vbox_bo *bo); static inline int vbox_bo_reserve(struct vbox_bo *bo, bool no_wait) { int ret; ret = ttm_bo_reserve(&bo->bo, true, no_wait, NULL); if (ret) { if (ret != -ERESTARTSYS && ret != -EBUSY) DRM_ERROR("reserve failed %p\n", bo); return ret; } return 0; } static inline void vbox_bo_unreserve(struct vbox_bo *bo) { ttm_bo_unreserve(&bo->bo); } void vbox_ttm_placement(struct vbox_bo *bo, int domain); int vbox_bo_push_sysram(struct vbox_bo *bo); int vbox_mmap(struct file *filp, struct vm_area_struct *vma); void *vbox_bo_kmap(struct vbox_bo *bo); void vbox_bo_kunmap(struct vbox_bo *bo); /* vbox_prime.c */ int vbox_gem_prime_pin(struct drm_gem_object *obj); void vbox_gem_prime_unpin(struct drm_gem_object *obj); struct sg_table *vbox_gem_prime_get_sg_table(struct drm_gem_object *obj); struct drm_gem_object *vbox_gem_prime_import_sg_table( struct drm_device *dev, struct dma_buf_attachment *attach, struct sg_table *table); void *vbox_gem_prime_vmap(struct drm_gem_object *obj); void vbox_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr); int vbox_gem_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *area); /* vbox_irq.c */ int vbox_irq_init(struct vbox_private *vbox); void vbox_irq_fini(struct vbox_private *vbox); void vbox_report_hotplug(struct vbox_private *vbox); irqreturn_t vbox_irq_handler(int irq, void *arg); /* vbox_hgsmi.c */ void *hgsmi_buffer_alloc(struct gen_pool *guest_pool, size_t size, u8 channel, u16 channel_info); void hgsmi_buffer_free(struct gen_pool *guest_pool, void *buf); int hgsmi_buffer_submit(struct gen_pool *guest_pool, void *buf); static inline void vbox_write_ioport(u16 index, u16 data) { outw(index, VBE_DISPI_IOPORT_INDEX); outw(data, VBE_DISPI_IOPORT_DATA); } #endif