summaryrefslogtreecommitdiff
path: root/linux/mga_clear.c
diff options
context:
space:
mode:
authorKevin E Martin <kem@kem.org>2000-02-22 15:43:59 +0000
committerKevin E Martin <kem@kem.org>2000-02-22 15:43:59 +0000
commit7a9b291ab55dd9dcfeb35217f6105ad57c94f433 (patch)
tree8941caba3a974912d8abc9f2f7b307c82ef5a54a /linux/mga_clear.c
parent9a1197da5cd84624f5b0741e0a20fee60eb8b4f1 (diff)
Import of XFree86 3.9.18X_3_9_18
Diffstat (limited to 'linux/mga_clear.c')
-rw-r--r--linux/mga_clear.c417
1 files changed, 417 insertions, 0 deletions
diff --git a/linux/mga_clear.c b/linux/mga_clear.c
new file mode 100644
index 00000000..05895415
--- /dev/null
+++ b/linux/mga_clear.c
@@ -0,0 +1,417 @@
+/* mga_state.c -- State support for mga g200/g400 -*- linux-c -*-
+ *
+ * Created: February 2000 by keithw@precisioninsight.com
+ *
+ * Copyright 1999 Precision Insight, Inc., Cedar Park, Texas.
+ * All Rights Reserved.
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
+ * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
+ * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
+ * DEALINGS IN THE SOFTWARE.
+ *
+ * Authors:
+ * Keith Whitwell <keithw@precisioninsight.com>
+ *
+ */
+
+#define __NO_VERSION__
+#include "drmP.h"
+#include "mga_drv.h"
+#include "mgareg_flags.h"
+#include "mga_dma.h"
+#include "mga_state.h"
+
+#define MGA_CLEAR_CMD (DC_opcod_trap | DC_arzero_enable | \
+ DC_sgnzero_enable | DC_shftzero_enable | \
+ (0xC << DC_bop_SHIFT) | DC_clipdis_enable | \
+ DC_solid_enable | DC_transc_enable)
+
+
+#define MGA_COPY_CMD (DC_opcod_bitblt | DC_atype_rpl | DC_linear_xy | \
+ DC_solid_disable | DC_arzero_disable | \
+ DC_sgnzero_enable | DC_shftzero_enable | \
+ (0xC << DC_bop_SHIFT) | DC_bltmod_bfcol | \
+ DC_pattern_disable | DC_transc_disable | \
+ DC_clipdis_enable) \
+
+
+
+/* Build and queue a TT_GENERAL secondary buffer to do the clears.
+ * With Jeff's ringbuffer idea, it might make sense if there are only
+ * one or two cliprects to emit straight to the primary buffer.
+ */
+static int mgaClearBuffers(drm_device_t *dev,
+ int clear_color,
+ int clear_depth,
+ int flags)
+{
+ int cmd, i;
+ drm_device_dma_t *dma = dev->dma;
+ drm_mga_private_t *dev_priv = (drm_mga_private_t *)dev->dev_private;
+ drm_mga_sarea_t *sarea_priv = dev_priv->sarea_priv;
+ xf86drmClipRectRec *pbox = sarea_priv->boxes;
+ int nbox = sarea_priv->nbox;
+ drm_buf_t *buf;
+ drm_dma_t d;
+ int order = 10; /* ??? what orders do we have ???*/
+ DMALOCALS;
+
+
+ if (!nbox)
+ return -EINVAL;
+
+ if ( dev_priv->sgram )
+ cmd = MGA_CLEAR_CMD | DC_atype_blk;
+ else
+ cmd = MGA_CLEAR_CMD | DC_atype_rstr;
+
+ buf = drm_freelist_get(&dma->bufs[order].freelist, _DRM_DMA_WAIT);
+
+
+ DMAGETPTR( buf );
+
+ for (i = 0 ; i < nbox ; i++) {
+ unsigned int height = pbox[i].y2 - pbox[i].y1;
+
+ /* Is it necessary to be this paranoid? I don't think so.
+ if (pbox[i].x1 > dev_priv->width) continue;
+ if (pbox[i].y1 > dev_priv->height) continue;
+ if (pbox[i].x2 > dev_priv->width) continue;
+ if (pbox[i].y2 > dev_priv->height) continue;
+ if (pbox[i].x2 <= pbox[i].x1) continue;
+ if (pbox[i].y2 <= pbox[i].x1) continue;
+ */
+
+ DMAOUTREG(MGAREG_YDSTLEN, (pbox[i].y1<<16)|height);
+ DMAOUTREG(MGAREG_FXBNDRY, (pbox[i].x2<<16)|pbox[i].x1);
+
+ if ( flags & MGA_CLEAR_FRONT ) {
+ DMAOUTREG(MGAREG_FCOL, clear_color);
+ DMAOUTREG(MGAREG_DSTORG, dev_priv->frontOrg);
+ DMAOUTREG(MGAREG_DWGCTL+MGAREG_MGA_EXEC, cmd );
+ }
+
+ if ( flags & MGA_CLEAR_BACK ) {
+ DMAOUTREG(MGAREG_FCOL, clear_color);
+ DMAOUTREG(MGAREG_DSTORG, dev_priv->backOrg);
+ DMAOUTREG(MGAREG_DWGCTL+MGAREG_MGA_EXEC, cmd );
+ }
+
+ if ( flags & MGA_CLEAR_DEPTH )
+ {
+ DMAOUTREG(MGAREG_FCOL, clear_depth);
+ DMAOUTREG(MGAREG_DSTORG, dev_priv->depthOrg);
+ DMAOUTREG(MGAREG_DWGCTL+MGAREG_MGA_EXEC, cmd );
+ }
+ }
+
+ DMAADVANCE( buf );
+
+ /* Make sure we restore the 3D state next time.
+ */
+ sarea_priv->dirty |= MGASAREA_NEW_CONTEXT;
+
+ ((drm_mga_buf_priv_t *)buf->dev_private)->dma_type = MGA_DMA_GENERAL;
+
+ d.context = DRM_KERNEL_CONTEXT;
+ d.send_count = 1;
+ d.send_indices = &buf->idx;
+ d.send_sizes = &buf->used;
+ d.flags = 0;
+ d.request_count = 0;
+ d.request_size = 0;
+ d.request_indices = NULL;
+ d.request_sizes = NULL;
+ d.granted_count = 0;
+
+ atomic_inc(&dev_priv->pending_bufs);
+ if((drm_dma_enqueue(dev, &d)) != 0)
+ atomic_dec(&dev_priv->pending_bufs);
+ mga_dma_schedule(dev, 1);
+ return 0;
+}
+
+int mgaSwapBuffers(drm_device_t *dev, int flags)
+{
+ drm_device_dma_t *dma = dev->dma;
+ drm_mga_private_t *dev_priv = (drm_mga_private_t *)dev->dev_private;
+ drm_mga_sarea_t *sarea_priv = dev_priv->sarea_priv;
+ xf86drmClipRectRec *pbox = sarea_priv->boxes;
+ int nbox = sarea_priv->nbox;
+ drm_buf_t *buf;
+ drm_dma_t d;
+ int order = 10; /* ??? */
+ int i;
+ DMALOCALS;
+
+ if (!nbox)
+ return -EINVAL;
+
+ buf = drm_freelist_get(&dma->bufs[order].freelist, _DRM_DMA_WAIT);
+
+ DMAGETPTR(buf);
+
+ DMAOUTREG(MGAREG_DSTORG, dev_priv->frontOrg);
+ DMAOUTREG(MGAREG_MACCESS, dev_priv->mAccess);
+ DMAOUTREG(MGAREG_SRCORG, dev_priv->backOrg);
+ DMAOUTREG(MGAREG_AR5, dev_priv->stride); /* unnecessary? */
+ DMAOUTREG(MGAREG_DWGCTL, MGA_COPY_CMD);
+
+ for (i = 0 ; i < nbox; i++) {
+ unsigned int h = pbox[i].y2 - pbox[i].y1;
+ unsigned int start = pbox[i].y1 * dev_priv->stride;
+
+ /*
+ if (pbox[i].x1 > dev_priv->width) continue;
+ if (pbox[i].y1 > dev_priv->height) continue;
+ if (pbox[i].x2 > dev_priv->width) continue;
+ if (pbox[i].y2 > dev_priv->height) continue;
+ if (pbox[i].x2 <= pbox[i].x1) continue;
+ if (pbox[i].y2 <= pbox[i].x1) continue;
+ */
+
+ DMAOUTREG(MGAREG_AR0, start + pbox[i].x2 - 1);
+ DMAOUTREG(MGAREG_AR3, start + pbox[i].x1);
+ DMAOUTREG(MGAREG_FXBNDRY, pbox[i].x1|((pbox[i].x2 - 1)<<16));
+ DMAOUTREG(MGAREG_YDSTLEN+MGAREG_MGA_EXEC, (pbox[i].y1<<16)|h);
+ }
+
+ DMAOUTREG(MGAREG_SRCORG, 0);
+ DMAADVANCE( buf );
+
+ /* Make sure we restore the 3D state next time.
+ */
+ sarea_priv->dirty |= MGASAREA_NEW_CONTEXT;
+
+ ((drm_mga_buf_priv_t *)buf->dev_private)->dma_type = MGA_DMA_GENERAL;
+
+ d.context = DRM_KERNEL_CONTEXT;
+ d.send_count = 1;
+ d.send_indices = &buf->idx;
+ d.send_sizes = &buf->used;
+ d.flags = 0;
+ d.request_count = 0;
+ d.request_size = 0;
+ d.request_indices = NULL;
+ d.request_sizes = NULL;
+ d.granted_count = 0;
+
+ atomic_inc(&dev_priv->pending_bufs);
+ if((drm_dma_enqueue(dev, &d)) != 0)
+ atomic_dec(&dev_priv->pending_bufs);
+ mga_dma_schedule(dev, 1);
+ return 0;
+}
+
+
+static int mgaIload(drm_device_t *dev, drm_mga_iload_t *args)
+{
+ drm_device_dma_t *dma = dev->dma;
+ drm_buf_t *buf = dma->buflist[ args->idx ];
+ drm_mga_buf_priv_t *buf_priv = (drm_mga_buf_priv_t *)buf->dev_private;
+ drm_mga_private_t *dev_priv = (drm_mga_private_t *)dev->dev_private;
+ drm_mga_sarea_t *sarea_priv = dev_priv->sarea_priv;
+ drm_dma_t d;
+ int pixperdword;
+
+ buf_priv->dma_type = MGA_DMA_ILOAD;
+ buf_priv->boxes[0].y1 = args->texture.y1;
+ buf_priv->boxes[0].y2 = args->texture.y2;
+ buf_priv->boxes[0].x1 = args->texture.x1;
+ buf_priv->boxes[0].x2 = args->texture.x2;
+ buf_priv->ContextState[MGA_CTXREG_DSTORG] = args->destOrg;
+ buf_priv->ContextState[MGA_CTXREG_MACCESS] = args->mAccess;
+ buf_priv->ServerState[MGA_2DREG_PITCH] = args->pitch;
+ buf_priv->nbox = 1;
+ sarea_priv->dirty |= (MGASAREA_NEW_CONTEXT | MGASAREA_NEW_2D);
+ switch((args->mAccess & 0x00000003)) {
+ case 0:
+ pixperdword = 4;
+ break;
+ case 1:
+ pixperdword = 2;
+ break;
+ case 2:
+ pixperdword = 1;
+ break;
+ default:
+ DRM_ERROR("Invalid maccess value passed"
+ " to mgaIload\n");
+ return -EINVAL;
+ }
+ buf->used = ((args->texture.y2 - args->texture.y1) *
+ (args->texture.x2 - args->texture.x1) /
+ pixperdword);
+ DRM_DEBUG("buf->used : %d\n", buf->used);
+ d.context = DRM_KERNEL_CONTEXT;
+ d.send_count = 1;
+ d.send_indices = &buf->idx;
+ d.send_sizes = &buf->used;
+ d.flags = 0;
+ d.request_count = 0;
+ d.request_size = 0;
+ d.request_indices = NULL;
+ d.request_sizes = NULL;
+ d.granted_count = 0;
+
+ atomic_inc(&dev_priv->pending_bufs);
+ if((drm_dma_enqueue(dev, &d)) != 0)
+ atomic_dec(&dev_priv->pending_bufs);
+ mga_dma_schedule(dev, 1);
+
+ return 0;
+}
+
+
+/* Necessary? Not necessary??
+ */
+static int check_lock(void)
+{
+ return 1;
+}
+
+int mga_clear_bufs(struct inode *inode, struct file *filp,
+ unsigned int cmd, unsigned long arg)
+{
+ drm_file_t *priv = filp->private_data;
+ drm_device_t *dev = priv->dev;
+ drm_mga_clear_t clear;
+ int retcode;
+
+ copy_from_user_ret(&clear, (drm_mga_clear_t *)arg,
+ sizeof(clear), -EFAULT);
+
+/* if (!check_lock( dev )) */
+/* return -EIEIO; */
+
+ retcode = mgaClearBuffers(dev, clear.clear_color,
+ clear.clear_depth,
+ clear.flags);
+
+ return retcode;
+}
+
+int mga_swap_bufs(struct inode *inode, struct file *filp,
+ unsigned int cmd, unsigned long arg)
+{
+ drm_file_t *priv = filp->private_data;
+ drm_device_t *dev = priv->dev;
+ drm_mga_swap_t swap;
+ int retcode = 0;
+
+/* if (!check_lock( dev )) */
+/* return -EIEIO; */
+
+ copy_from_user_ret(&swap, (drm_mga_swap_t *)arg,
+ sizeof(swap), -EFAULT);
+
+ retcode = mgaSwapBuffers(dev, swap.flags);
+
+ return retcode;
+}
+
+int mga_iload(struct inode *inode, struct file *filp,
+ unsigned int cmd, unsigned long arg)
+{
+ drm_file_t *priv = filp->private_data;
+ drm_device_t *dev = priv->dev;
+ drm_mga_iload_t iload;
+ int retcode = 0;
+
+/* if (!check_lock( dev )) */
+/* return -EIEIO; */
+
+ copy_from_user_ret(&iload, (drm_mga_iload_t *)arg,
+ sizeof(iload), -EFAULT);
+
+ retcode = mgaIload(dev, &iload);
+
+ return retcode;
+}
+
+
+int mga_dma(struct inode *inode, struct file *filp, unsigned int cmd,
+ unsigned long arg)
+{
+ drm_file_t *priv = filp->private_data;
+ drm_device_t *dev = priv->dev;
+ drm_device_dma_t *dma = dev->dma;
+ int retcode = 0;
+ drm_dma_t d;
+
+ copy_from_user_ret(&d, (drm_dma_t *)arg, sizeof(d), -EFAULT);
+ DRM_DEBUG("%d %d: %d send, %d req\n",
+ current->pid, d.context, d.send_count, d.request_count);
+
+ /* Per-context queues are unworkable if you are trying to do
+ * state management from the client.
+ */
+ d.context = DRM_KERNEL_CONTEXT;
+ d.flags &= ~_DRM_DMA_WHILE_LOCKED;
+
+ /* Maybe multiple buffers is useful for iload...
+ * But this ioctl is only for *despatching* vertex data...
+ */
+ if (d.send_count < 0 || d.send_count > 1) {
+ DRM_ERROR("Process %d trying to send %d buffers (max 1)\n",
+ current->pid, d.send_count);
+ return -EINVAL;
+ }
+
+
+ /* But it *is* used to request buffers for all types of dma:
+ */
+ if (d.request_count < 0 || d.request_count > dma->buf_count) {
+ DRM_ERROR("Process %d trying to get %d buffers (of %d max)\n",
+ current->pid, d.request_count, dma->buf_count);
+ return -EINVAL;
+ }
+
+ if (d.send_count) {
+ int idx = d.send_indices[0];
+ drm_mga_buf_priv_t *buf_priv = dma->buflist[ idx ]->dev_private;
+ drm_mga_private_t *dev_priv = dev->dev_private;
+
+ buf_priv->dma_type = MGA_DMA_VERTEX;
+
+/* if (!check_lock( dev )) */
+/* return -EIEIO; */
+
+ /* Snapshot the relevent bits of the sarea...
+ */
+ mgaCopyAndVerifyState( dev_priv, buf_priv );
+
+ atomic_inc(&dev_priv->pending_bufs);
+ retcode = drm_dma_enqueue(dev, &d);
+ if(retcode != 0)
+ atomic_dec(&dev_priv->pending_bufs);
+ mga_dma_schedule(dev, 1);
+ }
+
+ d.granted_count = 0;
+
+ if (!retcode && d.request_count) {
+ retcode = drm_dma_get_buffers(dev, &d);
+ }
+
+ DRM_DEBUG("%d returning, granted = %d\n",
+ current->pid, d.granted_count);
+ copy_to_user_ret((drm_dma_t *)arg, &d, sizeof(d), -EFAULT);
+
+ return retcode;
+}