summaryrefslogtreecommitdiff
path: root/drivers/gpu/drm/qxl/qxl_fb.c
diff options
context:
space:
mode:
Diffstat (limited to 'drivers/gpu/drm/qxl/qxl_fb.c')
-rw-r--r--drivers/gpu/drm/qxl/qxl_fb.c784
1 files changed, 784 insertions, 0 deletions
diff --git a/drivers/gpu/drm/qxl/qxl_fb.c b/drivers/gpu/drm/qxl/qxl_fb.c
new file mode 100644
index 00000000000..c2a9fb51e7b
--- /dev/null
+++ b/drivers/gpu/drm/qxl/qxl_fb.c
@@ -0,0 +1,784 @@
+/*
+ * Copyright © 2007 David Airlie
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
+ * DEALINGS IN THE SOFTWARE.
+ *
+ * Authors:
+ * David Airlie
+ */
+ /*
+ * Modularization
+ */
+
+#include <linux/module.h>
+#include <linux/fb.h>
+
+#include "drmP.h"
+#include "drm.h"
+#include "drm_crtc.h"
+#include "drm_crtc_helper.h"
+#include "qxl_drm.h"
+#include "qxl_drv.h"
+
+#include "qxl_object.h"
+#include "drm_fb_helper.h"
+
+#define QXL_DIRTY_DELAY (HZ / 30)
+
+struct qxl_fbdev {
+ struct drm_fb_helper helper;
+ struct qxl_framebuffer qfb;
+ struct list_head fbdev_list;
+ struct qxl_device *qdev;
+
+ /* dirty memory logging */
+ struct {
+ spinlock_t lock;
+ bool active;
+ unsigned x1;
+ unsigned y1;
+ unsigned x2;
+ unsigned y2;
+ } dirty;
+};
+
+static void qxl_fb_image_init(struct qxl_fb_image *qxl_fb_image,
+ struct qxl_device *qdev, struct fb_info *info,
+ const struct fb_image *image)
+{
+ qxl_fb_image->qdev = qdev;
+ if (info) {
+ qxl_fb_image->visual = info->fix.visual;
+ if (qxl_fb_image->visual == FB_VISUAL_TRUECOLOR ||
+ qxl_fb_image->visual == FB_VISUAL_DIRECTCOLOR) {
+ memcpy(&qxl_fb_image->pseudo_palette, info->pseudo_palette,
+ sizeof(qxl_fb_image->pseudo_palette));
+ }
+ } else {
+ /* fallback */
+ if (image->depth == 1) {
+ qxl_fb_image->visual = FB_VISUAL_MONO10;
+ } else {
+ qxl_fb_image->visual = FB_VISUAL_DIRECTCOLOR;
+ }
+ }
+ if (image) {
+ memcpy(&qxl_fb_image->fb_image, image,
+ sizeof(qxl_fb_image->fb_image));
+ }
+}
+
+static void qxl_fb_dirty_flush(struct fb_info *info)
+{
+ struct qxl_fbdev *qfbdev = info->par;
+ struct qxl_device *qdev = qfbdev->qdev;
+ struct qxl_fb_image qxl_fb_image;
+ struct fb_image *image = &qxl_fb_image.fb_image;
+ u32 x1, x2, y1, y2;
+ int stride = qfbdev->qfb.base.pitches[0] * 4; /* TODO: hard coding 32 bpp */
+
+ x1 = qfbdev->dirty.x1;
+ x2 = qfbdev->dirty.x2;
+ y1 = qfbdev->dirty.y1;
+ y2 = qfbdev->dirty.y2;
+ /*
+ * we are using a shadow draw buffer, at qdev->surface0_shadow
+ */
+ qxl_io_log(qdev, "dirty x[%d, %d], y[%d, %d]", x1, x2, y1, y2);
+ image->dx = x1;
+ image->dy = y1;
+ image->width = x2 - x1;
+ image->height = y2 - y1;
+ image->fg_color = 0xffffffff; /* unused, just to avoid uninitialized warnings */
+ image->bg_color = 0;
+ image->depth = 32; /* TODO: take from somewhere? */
+ image->cmap.start = 0;
+ image->cmap.len = 0;
+ image->cmap.red = NULL;
+ image->cmap.green = NULL;
+ image->cmap.blue = NULL;
+ image->cmap.transp = NULL;
+ image->data = qdev->surface0_shadow + (x1 * 4) + (stride * y1);
+
+ qxl_fb_image_init(&qxl_fb_image, qdev, info, NULL);
+ qxl_draw_opaque_fb(&qxl_fb_image, stride);
+ qfbdev->dirty.x1 = 0;
+ qfbdev->dirty.x2 = 0;
+ qfbdev->dirty.y1 = 0;
+ qfbdev->dirty.y2 = 0;
+}
+
+static void qxl_deferred_io(struct fb_info *info,
+ struct list_head *pagelist)
+{
+ struct qxl_fbdev *qfbdev = info->par;
+ unsigned long start, end, min, max;
+ struct page *page;
+ int y1, y2;
+
+ min = ULONG_MAX;
+ max = 0;
+ list_for_each_entry(page, pagelist, lru) {
+ start = page->index << PAGE_SHIFT;
+ end = start + PAGE_SIZE - 1;
+ min = min(min, start);
+ max = max(max, end);
+ }
+
+ if (min < max) {
+ y1 = min / info->fix.line_length;
+ y2 = (max / info->fix.line_length) + 1;
+
+ // TODO: add spin lock?
+ //spin_lock_irqsave(&qfbdev->dirty.lock, flags);
+ qfbdev->dirty.x1 = 0;
+ qfbdev->dirty.y1 = y1;
+ qfbdev->dirty.x2 = info->var.xres;
+ qfbdev->dirty.y2 = y2;
+ //spin_unlock_irqrestore(&qfbdev->dirty.lock, flags);
+ }
+
+ qxl_fb_dirty_flush(info);
+};
+
+
+struct fb_deferred_io qxl_defio = {
+ .delay = QXL_DIRTY_DELAY,
+ .deferred_io = qxl_deferred_io,
+};
+
+static void qxl_fb_do_work_items(struct qxl_device *qdev);
+
+static void qxl_fb_fillrect(struct fb_info *info,
+ const struct fb_fillrect *fb_rect)
+{
+ struct qxl_fbdev *qfbdev = info->par;
+ struct qxl_device *qdev = qfbdev->qdev;
+ struct qxl_rect rect;
+ uint32_t color;
+ int x = fb_rect->dx;
+ int y = fb_rect->dy;
+ int width = fb_rect->width;
+ int height = fb_rect->height;
+ uint16_t rop;
+ struct qxl_draw_fill qxl_draw_fill_rec;
+
+ if (qxl_debug_disable_fb) {
+ QXL_INFO_ONCE(qdev, "(skipped) %s:%d\n", __func__, __LINE__);
+ return;
+ }
+
+ if (info->fix.visual == FB_VISUAL_TRUECOLOR ||
+ info->fix.visual == FB_VISUAL_DIRECTCOLOR )
+ color = ((u32 *) (info->pseudo_palette))[fb_rect->color];
+ else
+ color = fb_rect->color;
+ rect.left = x;
+ rect.right = x + width;
+ rect.top = y;
+ rect.bottom = y + height;
+ switch (fb_rect->rop) {
+ case ROP_XOR:
+ rop = SPICE_ROPD_OP_XOR;
+ break;
+ case ROP_COPY:
+ rop = SPICE_ROPD_OP_PUT;
+ break;
+ default:
+ printk( KERN_ERR "qxl_fb_fillrect(): unknown rop, defaulting to SPICE_ROPD_OP_PUT\n");
+ rop = SPICE_ROPD_OP_PUT;
+ }
+ qxl_draw_fill_rec.qdev = qdev;
+ qxl_draw_fill_rec.rect = rect;
+ qxl_draw_fill_rec.color = color;
+ qxl_draw_fill_rec.rop = rop;
+ if (in_interrupt()) {
+ unsigned long flags;
+ qxl_io_log(qdev, "%s: TODO use RCU, mysterious locks with spin_lock\n",
+ __func__);
+ return;
+ spin_lock_irqsave(&qdev->fb_workqueue_spinlock, flags);
+ if (qxl_fb_queue_draw_fill(&qxl_draw_fill_rec)) {
+ qxl_io_log(qdev, "%s: failed to queue work item in interrupt context.\n"
+ "fb_fillrect = %x\n", __FUNCTION__, fb_rect);
+ }
+ spin_unlock_irqrestore(&qdev->fb_workqueue_spinlock, flags);
+ return;
+ }
+ /* ensure proper order of rendering operations - TODO: must do this
+ * for everything. */
+ qxl_fb_do_work_items(qdev);
+ qxl_draw_fill(&qxl_draw_fill_rec);
+ //cfb_fillrect(info, rect);
+ //qxl_update_screen(qfbdev->qdev);
+}
+
+static void qxl_fb_copyarea(struct fb_info *info,
+ const struct fb_copyarea *region)
+{
+ struct qxl_fbdev *qfbdev = info->par;
+
+ if (qxl_debug_disable_fb) {
+ QXL_INFO_ONCE(qfbdev->qdev, "skipped %s:%d\n", __func__, __LINE__);
+ return;
+ }
+ qxl_draw_copyarea(qfbdev->qdev,
+ region->width, region->height,
+ region->sx, region->sy,
+ region->dx, region->dy);
+}
+
+#if 0
+static u32 checksum(struct fb_info *info)
+{
+ u8 *addr;
+ u32 sum = 0;
+ // for now checksum the pixmap - later check other stuff
+ for (addr = info->pixmap.addr; addr < info->pixmap.addr + info->pixmap.size; ++addr)
+ sum += *addr;
+ return sum;
+}
+#endif
+
+static struct qxl_fb_work_item *get_work_item(struct qxl_device *qdev, int type)
+{
+ struct qxl_fb_work_item *item;
+
+ if (list_empty(&qdev->fb_work_item_free)) {
+ // not much we can do, really - log this to qemu at least
+ return NULL;
+ }
+ item = list_first_entry(&qdev->fb_work_item_free, struct qxl_fb_work_item, head);
+ list_del(&item->head);
+ item->type = type;
+ return item;
+}
+
+static void push_fb_work_item(struct qxl_device *qdev, struct qxl_fb_work_item *item)
+{
+ list_add_tail(&item->head, &qdev->fb_work_item_pending);
+ queue_work(qdev->fb_workqueue, &qdev->fb_work);
+}
+
+/* not static for debugfs */
+int qxl_fb_queue_draw_fill(struct qxl_draw_fill *qxl_draw_fill_rec)
+{
+ struct qxl_device *qdev = qxl_draw_fill_rec->qdev;
+ struct qxl_fb_work_item *item = get_work_item(qdev, QXL_FB_WORK_ITEM_DRAW_FILL);
+
+ if (!item) {
+ return -ENOMEM;
+ }
+ item->qxl_draw_fill = *qxl_draw_fill_rec;
+ push_fb_work_item(qdev, item);
+ return 0;
+}
+
+/* not static for debugfs */
+int qxl_fb_queue_imageblit(struct qxl_device *qdev,
+ struct qxl_fb_image *qxl_fb_image,
+ struct fb_info *info,
+ const struct fb_image *image)
+{
+ struct qxl_fb_work_item *item = get_work_item(qdev, QXL_FB_WORK_ITEM_IMAGE);
+
+ if (!item) {
+ return -ENOMEM;
+ }
+ if (!qxl_fb_image) {
+ qxl_fb_image_init(&item->qxl_fb_image, qdev, info, image);
+ } else {
+ /* debugfs path */
+ memcpy(&item->qxl_fb_image, qxl_fb_image, sizeof(*qxl_fb_image));
+ }
+ push_fb_work_item(qdev, item);
+ return 0;
+}
+
+static void qxl_fb_imageblit_safe(struct qxl_fb_image *qxl_fb_image)
+{
+ //DRM_INFO("%s:%d %d,%d (%d,%d)\n", __func__, __LINE__,
+ // image->width, image->height, image->dx, image->dy);
+ qxl_draw_opaque_fb(qxl_fb_image, 0);
+}
+
+static void qxl_fb_imageblit(struct fb_info *info,
+ const struct fb_image *image)
+{
+ struct qxl_fbdev *qfbdev = info->par;
+ struct qxl_device *qdev = qfbdev->qdev;
+ struct qxl_fb_image qxl_fb_image;
+
+ if (qxl_debug_disable_fb) {
+ QXL_INFO_ONCE(qfbdev->qdev, "%s: skipping\n", __FUNCTION__);
+ return;
+ }
+ if (in_interrupt()) {
+ /* we cannot do any ttm_bo allocation since that will fail on
+ * ioremap_wc..__get_vm_area_node, so queue the work item instead
+ * This can happen from printk inside an interrupt context, i.e.:
+ * smp_apic_timer_interrupt..check_cpu_stall */
+ unsigned long flags;
+ qxl_io_log(qdev, "%s: TODO use RCU, mysterious locks with spin_lock\n",
+ __func__);
+ return;
+ spin_lock_irqsave(&qdev->fb_workqueue_spinlock, flags);
+ if (qxl_fb_queue_imageblit(qdev, NULL, info, image)) {
+ qxl_io_log(qdev, "%s: failed to queue work item in interrupt context.\n"
+ "info = %x, image = %x\n", __FUNCTION__, info, image);
+ }
+ spin_unlock_irqrestore(&qdev->fb_workqueue_spinlock, flags);
+ return;
+ }
+ /* ensure proper order of rendering operations - TODO: must do this
+ * for everything. */
+ qxl_fb_do_work_items(qfbdev->qdev);
+ qxl_fb_image_init(&qxl_fb_image, qfbdev->qdev, info, image);
+ qxl_fb_imageblit_safe(&qxl_fb_image);
+}
+
+static void qxl_fb_do_work_items(struct qxl_device *qdev)
+{
+ struct qxl_fb_work_item *item;
+ struct qxl_fb_work_item *n;
+ unsigned long flags;
+
+ if (in_interrupt()) {
+ qxl_io_log(qdev, "%s: wtf - should not happen\n", __FUNCTION__);
+ return;
+ }
+
+ spin_lock_irqsave(&qdev->fb_workqueue_spinlock, flags);
+ list_for_each_entry_safe(item, n, &qdev->fb_work_item_pending, head) {
+ spin_unlock_irqrestore(&qdev->fb_workqueue_spinlock, flags);
+ switch (item->type) {
+ case QXL_FB_WORK_ITEM_IMAGE:
+ qxl_fb_imageblit_safe(&item->qxl_fb_image);
+ break;
+ case QXL_FB_WORK_ITEM_DRAW_FILL:
+ qxl_draw_fill(&item->qxl_draw_fill);
+ break;
+ default:
+ qxl_io_log(qdev, "%s: invalid fb work item type %d\n", item->type);
+ }
+ spin_lock_irqsave(&qdev->fb_workqueue_spinlock, flags);
+ list_del(&item->head);
+ list_add_tail(&item->head, &qdev->fb_work_item_free);
+ spin_unlock_irqrestore(&qdev->fb_workqueue_spinlock, flags);
+ }
+ spin_unlock_irqrestore(&qdev->fb_workqueue_spinlock, flags);
+}
+
+static void qxl_fb_work(struct work_struct *work)
+{
+ struct qxl_device *qdev = container_of(work, struct qxl_device, fb_work);
+
+ qxl_fb_do_work_items(qdev);
+}
+
+int qxl_fb_init(struct qxl_device *qdev)
+{
+ int i;
+
+ qdev->fb_workqueue = alloc_workqueue("qxl_fb", WQ_NON_REENTRANT, 0);
+ if (!qdev->fb_workqueue)
+ return -ENOMEM;
+ INIT_LIST_HEAD(&qdev->fb_work_item_pending);
+ INIT_LIST_HEAD(&qdev->fb_work_item_free);
+ INIT_WORK(&qdev->fb_work, qxl_fb_work);
+ spin_lock_init(&qdev->fb_workqueue_spinlock);
+ /* no kmalloc in interrupt context - not sure if I'm fixing a problem
+ * or introducing a limitation */
+ for (i = 0 ; i < sizeof(qdev->fb_work_items) / sizeof(qdev->fb_work_items[0]) ; ++i) {
+ list_add_tail(&qdev->fb_work_items[i].head, &qdev->fb_work_item_free);
+ }
+ return 0;
+}
+
+
+struct qxl_fb_par {
+ u32 pseudo_palette[17]; /* why not 257? Xorg wants 256 elements. */
+};
+
+static int qxl_fb_pan_display(struct fb_var_screeninfo *var,
+ struct fb_info *info)
+{
+ return 0;
+}
+
+static int qxl_fb_blank(int blank, struct fb_info *info)
+{
+ return 0;
+}
+
+
+static struct fb_ops qxlfb_ops = {
+ .owner = THIS_MODULE,
+ .fb_check_var = drm_fb_helper_check_var,
+ .fb_set_par = drm_fb_helper_set_par, // TODO: like vmwgfx
+ .fb_fillrect = qxl_fb_fillrect,
+ .fb_copyarea = qxl_fb_copyarea,
+ .fb_imageblit = qxl_fb_imageblit,
+ .fb_pan_display = qxl_fb_pan_display,
+ .fb_blank = qxl_fb_blank,
+ .fb_setcmap = drm_fb_helper_setcmap,
+ .fb_debug_enter = drm_fb_helper_debug_enter,
+ .fb_debug_leave = drm_fb_helper_debug_leave,
+};
+
+static void qxlfb_destroy_pinned_object(struct drm_gem_object *gobj)
+{
+ struct qxl_bo *qbo = gem_to_qxl_bo(gobj);
+ int ret;
+
+ ret = qxl_bo_reserve(qbo, false);
+ if (likely(ret == 0)) {
+ qxl_bo_kunmap(qbo);
+ qxl_bo_unpin(qbo);
+ qxl_bo_unreserve(qbo);
+ }
+ drm_gem_object_unreference_unlocked(gobj);
+}
+
+int qxl_get_handle_for_primary_fb(struct qxl_device *qdev,
+ struct drm_file *file_priv,
+ uint32_t *handle)
+{
+ int r;
+ struct drm_gem_object *gobj = qdev->fbdev_qfb->obj;
+
+ BUG_ON(!gobj);
+ /* drm_get_handle_create adds a reference - good */
+ r = drm_gem_handle_create(file_priv, gobj, handle);
+ if (r) {
+ return r;
+ }
+ return 0;
+}
+
+static int qxlfb_create_pinned_object(struct qxl_fbdev *qfbdev,
+ struct drm_mode_fb_cmd2 *mode_cmd,
+ struct drm_gem_object **gobj_p)
+{
+ struct qxl_device *qdev = qfbdev->qdev;
+ struct drm_gem_object *gobj = NULL;
+ struct qxl_bo *qbo = NULL;
+ int ret;
+ int aligned_size, size;
+ int height = mode_cmd->height;
+ int bpp;
+ int depth;
+
+ drm_fb_get_bpp_depth(mode_cmd->pixel_format, &bpp, &depth);
+
+ size = mode_cmd->pitches[0] * height;
+ aligned_size = ALIGN(size, PAGE_SIZE);
+ // TODO: unallocate and reallocate surface0 for real. Hack to just
+ // have a large enough surface0 for 1024x768 Xorg 32bpp mode
+ ret = qxl_gem_object_create(qdev, 1024*768*4 /* aligned_size */, 0,
+ QXL_GEM_DOMAIN_VRAM,
+ false, /* is discardable */
+ false, /* is kernel (false means device) */
+ &gobj);
+ if (ret) {
+ printk(KERN_ERR "failed to allocate framebuffer (%d)\n",
+ aligned_size);
+ return -ENOMEM;
+ }
+ qbo = gem_to_qxl_bo(gobj);
+
+ ret = qxl_bo_reserve(qbo, false);
+ if (unlikely(ret != 0))
+ goto out_unref;
+ ret = qxl_bo_pin(qbo, QXL_GEM_DOMAIN_VRAM, NULL);
+ if (ret) {
+ qxl_bo_unreserve(qbo);
+ goto out_unref;
+ }
+ ret = qxl_bo_kmap(qbo, NULL);
+ qxl_bo_unreserve(qbo); // unreserve, will be mmaped
+ if (ret) {
+ goto out_unref;
+ }
+
+ *gobj_p = gobj;
+ return 0;
+out_unref:
+ qxlfb_destroy_pinned_object(gobj);
+ *gobj_p = NULL;
+ return ret;
+}
+
+static int qxlfb_create(struct qxl_fbdev *qfbdev,
+ struct drm_fb_helper_surface_size *sizes)
+{
+ struct qxl_device *qdev = qfbdev->qdev;
+ struct fb_info *info;
+ struct drm_framebuffer *fb = NULL;
+ struct drm_mode_fb_cmd2 mode_cmd;
+ struct drm_gem_object *gobj = NULL;
+ struct qxl_bo *qbo = NULL;
+ struct device *device = &qdev->pdev->dev;
+ int ret;
+ int size;
+ int bpp = sizes->surface_bpp;
+ int depth = sizes->surface_depth;
+
+ mode_cmd.width = sizes->surface_width;
+ mode_cmd.height = sizes->surface_height;
+
+ mode_cmd.pitches[0] = ALIGN(mode_cmd.width * ((bpp + 1) / 8), 64);
+ mode_cmd.pixel_format = drm_mode_legacy_fb_format(bpp, depth);
+
+ ret = qxlfb_create_pinned_object(qfbdev, &mode_cmd, &gobj);
+ qdev->surface0_bo = qbo = gem_to_qxl_bo(gobj);
+ QXL_INFO(qdev, "%s: %dx%d %d\n", __func__, mode_cmd.width,
+ mode_cmd.height, mode_cmd.pitches[0]);
+ qdev->surface0_shadow = kmalloc(mode_cmd.pitches[0] * mode_cmd.height, GFP_KERNEL);
+ BUG_ON(!qdev->surface0_shadow); /* TODO: what's the usual response to memory allocation errors? that one. */
+ QXL_INFO(qdev, "surface0 at gpu offset %lld, mmap_offset %lld (virt %p, shadow %p)\n",
+ qxl_bo_gpu_offset(qdev->surface0_bo),
+ qxl_bo_mmap_offset(qdev->surface0_bo),
+ qdev->surface0_bo->kptr,
+ qdev->surface0_shadow);
+ size = mode_cmd.pitches[0] * mode_cmd.height;
+
+ info = framebuffer_alloc(0, device);
+ if (info == NULL) {
+ ret = -ENOMEM;
+ goto out_unref;
+ }
+
+ info->par = qfbdev;
+
+ qxl_framebuffer_init(qdev->ddev, &qfbdev->qfb, &mode_cmd, gobj);
+
+ fb = &qfbdev->qfb.base;
+
+ /* setup helper with fb data */
+ qfbdev->helper.fb = fb;
+ qfbdev->helper.fbdev = info;
+
+ // TODO: memset_io/memset causes 100k exits/sec for 60 seconds. wtf?
+ //memset_io(qbo->kptr, 0xff, size); // TODO - 0xff on purpose?
+
+ strcpy(info->fix.id, "qxldrmfb");
+
+ drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
+
+ info->flags = FBINFO_DEFAULT;
+ info->fbops = &qxlfb_ops;
+
+ // TODO: using gobj->size in various places in this function. Not sure
+ // what the difference between the different sizes is.
+// tmp = fb_gpuaddr - qdev->mc.vram_location;
+ // info->fix.smem_start = qdev->mc.aper_base + tmp;
+ info->fix.smem_start = qdev->vram_base; // TODO - correct?
+ info->fix.smem_len = gobj->size;
+ info->screen_base = qdev->surface0_shadow;
+ info->screen_size = gobj->size;
+
+ drm_fb_helper_fill_var(info, &qfbdev->helper, sizes->fb_width, sizes->fb_height);
+
+ /* setup aperture base/size for vesafb takeover */
+ info->apertures = alloc_apertures(1);
+ if (!info->apertures) {
+ ret = -ENOMEM;
+ goto out_unref;
+ }
+ info->apertures->ranges[0].base = qdev->ddev->mode_config.fb_base;
+ info->apertures->ranges[0].size = qdev->vram_size;
+
+ info->fix.mmio_start = 0;
+ info->fix.mmio_len = 0;
+ info->pixmap.size = 64*1024; /* TODO: not correct - must be taken from pci */
+ info->pixmap.buf_align = 8;
+ info->pixmap.access_align = 32;
+ info->pixmap.flags = FB_PIXMAP_SYSTEM;
+ info->pixmap.scan_align = 1;
+ if (info->screen_base == NULL) {
+ ret = -ENOSPC;
+ goto out_unref;
+ }
+
+ ret = fb_alloc_cmap(&info->cmap, 256, 0);
+ if (ret) {
+ ret = -ENOMEM;
+ goto out_unref;
+ }
+
+ info->fbdefio = &qxl_defio;
+ fb_deferred_io_init(info);
+
+ qdev->fbdev_info = info;
+ qdev->fbdev_qfb = &qfbdev->qfb;
+ DRM_INFO("fb mappable at 0x%lX\n", info->fix.smem_start);
+ // DRM_INFO("vram apper at 0x%lX\n", (unsigned long)qdev->mc.aper_base);
+ DRM_INFO("size %lu\n", (unsigned long)info->screen_size);
+ DRM_INFO("fb depth is %d\n", fb->depth);
+ DRM_INFO(" pitch is %d\n", fb->pitches[0]);
+ DRM_INFO(" width is %d\n", fb->width);
+ DRM_INFO(" height is %d\n", fb->height);
+
+ return 0;
+
+out_unref:
+ if (qbo) {
+ ret = qxl_bo_reserve(qbo, false);
+ if (likely(ret == 0)) {
+ qxl_bo_kunmap(qbo);
+ qxl_bo_unpin(qbo);
+ qxl_bo_unreserve(qbo);
+ }
+ }
+ if (fb && ret) {
+ drm_gem_object_unreference(gobj);
+ drm_framebuffer_cleanup(fb);
+ kfree(fb);
+ }
+ drm_gem_object_unreference(gobj);
+ return ret;
+}
+
+static int qxl_fb_find_or_create_single(struct drm_fb_helper *helper,
+ struct drm_fb_helper_surface_size *sizes)
+{
+ struct qxl_fbdev *qfbdev = (struct qxl_fbdev *)helper;
+ int new_fb = 0;
+ int ret;
+
+ if (!helper->fb) {
+ printk("%s:%d\n", __func__, __LINE__);
+ ret = qxlfb_create(qfbdev, sizes);
+ printk("%s:%d %d\n", __func__, __LINE__, ret);
+ if (ret)
+ return ret;
+ new_fb = 1;
+ }
+ return new_fb;
+}
+
+static char *mode_option;
+int qxl_parse_options(char *options)
+{
+ char *this_opt;
+
+ if (!options || !*options)
+ return 0;
+
+ while ((this_opt = strsep(&options, ",")) != NULL) {
+ if (!*this_opt)
+ continue;
+ mode_option = this_opt;
+ }
+ return 0;
+}
+
+static int qxl_fbdev_destroy(struct drm_device *dev, struct qxl_fbdev *qfbdev)
+{
+ struct fb_info *info;
+ struct qxl_framebuffer *qfb = &qfbdev->qfb;
+
+ if (qfbdev->helper.fbdev) {
+ info = qfbdev->helper.fbdev;
+
+ unregister_framebuffer(info);
+ framebuffer_release(info);
+ }
+ if (qfb->obj) {
+ qxlfb_destroy_pinned_object(qfb->obj);
+ qfb->obj = NULL;
+ }
+ drm_fb_helper_fini(&qfbdev->helper);
+ drm_framebuffer_cleanup(&qfb->base);
+
+ return 0;
+}
+
+static struct drm_fb_helper_funcs qxl_fb_helper_funcs = {
+ // .gamma_set = qxl_crtc_fb_gamma_set,
+ // .gamma_get = qxl_crtc_fb_gamma_get,
+ .fb_probe = qxl_fb_find_or_create_single,
+};
+
+/* TODO: defio
+ * How does that work? can I get the changed data to send as opaques?
+ * Otherwise I need to invalidate an area - counter to qxl phylosophy.
+ */
+
+int qxl_fbdev_init(struct qxl_device *qdev)
+{
+ struct qxl_fbdev *qfbdev;
+ int bpp_sel = 32; /* TODO: parameter from somewhere? */
+ int ret;
+
+ qfbdev = kzalloc(sizeof(struct qxl_fbdev), GFP_KERNEL);
+ if (!qfbdev)
+ return -ENOMEM;
+
+ qfbdev->qdev = qdev;
+ qdev->mode_info.qfbdev = qfbdev;
+ qfbdev->helper.funcs = &qxl_fb_helper_funcs;
+
+ printk("%s: 1\n", __func__);
+ ret = drm_fb_helper_init(qdev->ddev, &qfbdev->helper,
+ 1 /* num_crtc - QXL supports just 1 */,
+ QXLFB_CONN_LIMIT);
+ if (ret) {
+ kfree(qfbdev);
+ return ret;
+ }
+
+ drm_fb_helper_single_add_all_connectors(&qfbdev->helper);
+ drm_fb_helper_initial_config(&qfbdev->helper, bpp_sel);
+ return 0;
+}
+
+void qxl_fbdev_fini(struct qxl_device *qdev)
+{
+ if (!qdev->mode_info.qfbdev)
+ return;
+
+ qxl_fbdev_destroy(qdev->ddev, qdev->mode_info.qfbdev);
+ kfree(qdev->mode_info.qfbdev);
+ qdev->mode_info.qfbdev = NULL;
+}
+
+void qxl_fbdev_set_suspend(struct qxl_device *qdev, int state)
+{
+ fb_set_suspend(qdev->mode_info.qfbdev->helper.fbdev, state);
+}
+
+int qxl_fbdev_total_size(struct qxl_device *qdev)
+{
+ struct qxl_bo *robj;
+ int size = 0;
+
+ robj = gem_to_qxl_bo(qdev->mode_info.qfbdev->qfb.obj);
+ size += qxl_bo_size(robj);
+ return size;
+}
+
+bool qxl_fbdev_robj_is_fb(struct qxl_device *qdev, struct qxl_bo *robj)
+{
+ if (robj == gem_to_qxl_bo(qdev->mode_info.qfbdev->qfb.obj))
+ return true;
+ return false;
+}