1163 lines
30 KiB
C
1163 lines
30 KiB
C
/* savage_state.c -- State and drawing support for Savage
|
|
*
|
|
* Copyright 2004 Felix Kuehling
|
|
* All Rights Reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sub license,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the
|
|
* next paragraph) shall be included in all copies or substantial portions
|
|
* of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NON-INFRINGEMENT. IN NO EVENT SHALL FELIX KUEHLING BE LIABLE FOR
|
|
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF
|
|
* CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
#include <drm/drmP.h>
|
|
#include <drm/savage_drm.h>
|
|
#include "savage_drv.h"
|
|
|
|
void savage_emit_clip_rect_s3d(drm_savage_private_t * dev_priv,
|
|
const struct drm_clip_rect * pbox)
|
|
{
|
|
uint32_t scstart = dev_priv->state.s3d.new_scstart;
|
|
uint32_t scend = dev_priv->state.s3d.new_scend;
|
|
scstart = (scstart & ~SAVAGE_SCISSOR_MASK_S3D) |
|
|
((uint32_t) pbox->x1 & 0x000007ff) |
|
|
(((uint32_t) pbox->y1 << 16) & 0x07ff0000);
|
|
scend = (scend & ~SAVAGE_SCISSOR_MASK_S3D) |
|
|
(((uint32_t) pbox->x2 - 1) & 0x000007ff) |
|
|
((((uint32_t) pbox->y2 - 1) << 16) & 0x07ff0000);
|
|
if (scstart != dev_priv->state.s3d.scstart ||
|
|
scend != dev_priv->state.s3d.scend) {
|
|
DMA_LOCALS;
|
|
BEGIN_DMA(4);
|
|
DMA_WRITE(BCI_CMD_WAIT | BCI_CMD_WAIT_3D);
|
|
DMA_SET_REGISTERS(SAVAGE_SCSTART_S3D, 2);
|
|
DMA_WRITE(scstart);
|
|
DMA_WRITE(scend);
|
|
dev_priv->state.s3d.scstart = scstart;
|
|
dev_priv->state.s3d.scend = scend;
|
|
dev_priv->waiting = 1;
|
|
DMA_COMMIT();
|
|
}
|
|
}
|
|
|
|
void savage_emit_clip_rect_s4(drm_savage_private_t * dev_priv,
|
|
const struct drm_clip_rect * pbox)
|
|
{
|
|
uint32_t drawctrl0 = dev_priv->state.s4.new_drawctrl0;
|
|
uint32_t drawctrl1 = dev_priv->state.s4.new_drawctrl1;
|
|
drawctrl0 = (drawctrl0 & ~SAVAGE_SCISSOR_MASK_S4) |
|
|
((uint32_t) pbox->x1 & 0x000007ff) |
|
|
(((uint32_t) pbox->y1 << 12) & 0x00fff000);
|
|
drawctrl1 = (drawctrl1 & ~SAVAGE_SCISSOR_MASK_S4) |
|
|
(((uint32_t) pbox->x2 - 1) & 0x000007ff) |
|
|
((((uint32_t) pbox->y2 - 1) << 12) & 0x00fff000);
|
|
if (drawctrl0 != dev_priv->state.s4.drawctrl0 ||
|
|
drawctrl1 != dev_priv->state.s4.drawctrl1) {
|
|
DMA_LOCALS;
|
|
BEGIN_DMA(4);
|
|
DMA_WRITE(BCI_CMD_WAIT | BCI_CMD_WAIT_3D);
|
|
DMA_SET_REGISTERS(SAVAGE_DRAWCTRL0_S4, 2);
|
|
DMA_WRITE(drawctrl0);
|
|
DMA_WRITE(drawctrl1);
|
|
dev_priv->state.s4.drawctrl0 = drawctrl0;
|
|
dev_priv->state.s4.drawctrl1 = drawctrl1;
|
|
dev_priv->waiting = 1;
|
|
DMA_COMMIT();
|
|
}
|
|
}
|
|
|
|
static int savage_verify_texaddr(drm_savage_private_t * dev_priv, int unit,
|
|
uint32_t addr)
|
|
{
|
|
if ((addr & 6) != 2) { /* reserved bits */
|
|
DRM_ERROR("bad texAddr%d %08x (reserved bits)\n", unit, addr);
|
|
return -EINVAL;
|
|
}
|
|
if (!(addr & 1)) { /* local */
|
|
addr &= ~7;
|
|
if (addr < dev_priv->texture_offset ||
|
|
addr >= dev_priv->texture_offset + dev_priv->texture_size) {
|
|
DRM_ERROR
|
|
("bad texAddr%d %08x (local addr out of range)\n",
|
|
unit, addr);
|
|
return -EINVAL;
|
|
}
|
|
} else { /* AGP */
|
|
if (!dev_priv->agp_textures) {
|
|
DRM_ERROR("bad texAddr%d %08x (AGP not available)\n",
|
|
unit, addr);
|
|
return -EINVAL;
|
|
}
|
|
addr &= ~7;
|
|
if (addr < dev_priv->agp_textures->offset ||
|
|
addr >= (dev_priv->agp_textures->offset +
|
|
dev_priv->agp_textures->size)) {
|
|
DRM_ERROR
|
|
("bad texAddr%d %08x (AGP addr out of range)\n",
|
|
unit, addr);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#define SAVE_STATE(reg,where) \
|
|
if(start <= reg && start+count > reg) \
|
|
dev_priv->state.where = regs[reg - start]
|
|
#define SAVE_STATE_MASK(reg,where,mask) do { \
|
|
if(start <= reg && start+count > reg) { \
|
|
uint32_t tmp; \
|
|
tmp = regs[reg - start]; \
|
|
dev_priv->state.where = (tmp & (mask)) | \
|
|
(dev_priv->state.where & ~(mask)); \
|
|
} \
|
|
} while (0)
|
|
|
|
static int savage_verify_state_s3d(drm_savage_private_t * dev_priv,
|
|
unsigned int start, unsigned int count,
|
|
const uint32_t *regs)
|
|
{
|
|
if (start < SAVAGE_TEXPALADDR_S3D ||
|
|
start + count - 1 > SAVAGE_DESTTEXRWWATERMARK_S3D) {
|
|
DRM_ERROR("invalid register range (0x%04x-0x%04x)\n",
|
|
start, start + count - 1);
|
|
return -EINVAL;
|
|
}
|
|
|
|
SAVE_STATE_MASK(SAVAGE_SCSTART_S3D, s3d.new_scstart,
|
|
~SAVAGE_SCISSOR_MASK_S3D);
|
|
SAVE_STATE_MASK(SAVAGE_SCEND_S3D, s3d.new_scend,
|
|
~SAVAGE_SCISSOR_MASK_S3D);
|
|
|
|
/* if any texture regs were changed ... */
|
|
if (start <= SAVAGE_TEXCTRL_S3D &&
|
|
start + count > SAVAGE_TEXPALADDR_S3D) {
|
|
/* ... check texture state */
|
|
SAVE_STATE(SAVAGE_TEXCTRL_S3D, s3d.texctrl);
|
|
SAVE_STATE(SAVAGE_TEXADDR_S3D, s3d.texaddr);
|
|
if (dev_priv->state.s3d.texctrl & SAVAGE_TEXCTRL_TEXEN_MASK)
|
|
return savage_verify_texaddr(dev_priv, 0,
|
|
dev_priv->state.s3d.texaddr);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int savage_verify_state_s4(drm_savage_private_t * dev_priv,
|
|
unsigned int start, unsigned int count,
|
|
const uint32_t *regs)
|
|
{
|
|
int ret = 0;
|
|
|
|
if (start < SAVAGE_DRAWLOCALCTRL_S4 ||
|
|
start + count - 1 > SAVAGE_TEXBLENDCOLOR_S4) {
|
|
DRM_ERROR("invalid register range (0x%04x-0x%04x)\n",
|
|
start, start + count - 1);
|
|
return -EINVAL;
|
|
}
|
|
|
|
SAVE_STATE_MASK(SAVAGE_DRAWCTRL0_S4, s4.new_drawctrl0,
|
|
~SAVAGE_SCISSOR_MASK_S4);
|
|
SAVE_STATE_MASK(SAVAGE_DRAWCTRL1_S4, s4.new_drawctrl1,
|
|
~SAVAGE_SCISSOR_MASK_S4);
|
|
|
|
/* if any texture regs were changed ... */
|
|
if (start <= SAVAGE_TEXDESCR_S4 &&
|
|
start + count > SAVAGE_TEXPALADDR_S4) {
|
|
/* ... check texture state */
|
|
SAVE_STATE(SAVAGE_TEXDESCR_S4, s4.texdescr);
|
|
SAVE_STATE(SAVAGE_TEXADDR0_S4, s4.texaddr0);
|
|
SAVE_STATE(SAVAGE_TEXADDR1_S4, s4.texaddr1);
|
|
if (dev_priv->state.s4.texdescr & SAVAGE_TEXDESCR_TEX0EN_MASK)
|
|
ret |= savage_verify_texaddr(dev_priv, 0,
|
|
dev_priv->state.s4.texaddr0);
|
|
if (dev_priv->state.s4.texdescr & SAVAGE_TEXDESCR_TEX1EN_MASK)
|
|
ret |= savage_verify_texaddr(dev_priv, 1,
|
|
dev_priv->state.s4.texaddr1);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
#undef SAVE_STATE
|
|
#undef SAVE_STATE_MASK
|
|
|
|
static int savage_dispatch_state(drm_savage_private_t * dev_priv,
|
|
const drm_savage_cmd_header_t * cmd_header,
|
|
const uint32_t *regs)
|
|
{
|
|
unsigned int count = cmd_header->state.count;
|
|
unsigned int start = cmd_header->state.start;
|
|
unsigned int count2 = 0;
|
|
unsigned int bci_size;
|
|
int ret;
|
|
DMA_LOCALS;
|
|
|
|
if (!count)
|
|
return 0;
|
|
|
|
if (S3_SAVAGE3D_SERIES(dev_priv->chipset)) {
|
|
ret = savage_verify_state_s3d(dev_priv, start, count, regs);
|
|
if (ret != 0)
|
|
return ret;
|
|
/* scissor regs are emitted in savage_dispatch_draw */
|
|
if (start < SAVAGE_SCSTART_S3D) {
|
|
if (start + count > SAVAGE_SCEND_S3D + 1)
|
|
count2 = count - (SAVAGE_SCEND_S3D + 1 - start);
|
|
if (start + count > SAVAGE_SCSTART_S3D)
|
|
count = SAVAGE_SCSTART_S3D - start;
|
|
} else if (start <= SAVAGE_SCEND_S3D) {
|
|
if (start + count > SAVAGE_SCEND_S3D + 1) {
|
|
count -= SAVAGE_SCEND_S3D + 1 - start;
|
|
start = SAVAGE_SCEND_S3D + 1;
|
|
} else
|
|
return 0;
|
|
}
|
|
} else {
|
|
ret = savage_verify_state_s4(dev_priv, start, count, regs);
|
|
if (ret != 0)
|
|
return ret;
|
|
/* scissor regs are emitted in savage_dispatch_draw */
|
|
if (start < SAVAGE_DRAWCTRL0_S4) {
|
|
if (start + count > SAVAGE_DRAWCTRL1_S4 + 1)
|
|
count2 = count -
|
|
(SAVAGE_DRAWCTRL1_S4 + 1 - start);
|
|
if (start + count > SAVAGE_DRAWCTRL0_S4)
|
|
count = SAVAGE_DRAWCTRL0_S4 - start;
|
|
} else if (start <= SAVAGE_DRAWCTRL1_S4) {
|
|
if (start + count > SAVAGE_DRAWCTRL1_S4 + 1) {
|
|
count -= SAVAGE_DRAWCTRL1_S4 + 1 - start;
|
|
start = SAVAGE_DRAWCTRL1_S4 + 1;
|
|
} else
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
bci_size = count + (count + 254) / 255 + count2 + (count2 + 254) / 255;
|
|
|
|
if (cmd_header->state.global) {
|
|
BEGIN_DMA(bci_size + 1);
|
|
DMA_WRITE(BCI_CMD_WAIT | BCI_CMD_WAIT_3D);
|
|
dev_priv->waiting = 1;
|
|
} else {
|
|
BEGIN_DMA(bci_size);
|
|
}
|
|
|
|
do {
|
|
while (count > 0) {
|
|
unsigned int n = count < 255 ? count : 255;
|
|
DMA_SET_REGISTERS(start, n);
|
|
DMA_COPY(regs, n);
|
|
count -= n;
|
|
start += n;
|
|
regs += n;
|
|
}
|
|
start += 2;
|
|
regs += 2;
|
|
count = count2;
|
|
count2 = 0;
|
|
} while (count);
|
|
|
|
DMA_COMMIT();
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int savage_dispatch_dma_prim(drm_savage_private_t * dev_priv,
|
|
const drm_savage_cmd_header_t * cmd_header,
|
|
const struct drm_buf * dmabuf)
|
|
{
|
|
unsigned char reorder = 0;
|
|
unsigned int prim = cmd_header->prim.prim;
|
|
unsigned int skip = cmd_header->prim.skip;
|
|
unsigned int n = cmd_header->prim.count;
|
|
unsigned int start = cmd_header->prim.start;
|
|
unsigned int i;
|
|
BCI_LOCALS;
|
|
|
|
if (!dmabuf) {
|
|
DRM_ERROR("called without dma buffers!\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!n)
|
|
return 0;
|
|
|
|
switch (prim) {
|
|
case SAVAGE_PRIM_TRILIST_201:
|
|
reorder = 1;
|
|
prim = SAVAGE_PRIM_TRILIST;
|
|
case SAVAGE_PRIM_TRILIST:
|
|
if (n % 3 != 0) {
|
|
DRM_ERROR("wrong number of vertices %u in TRILIST\n",
|
|
n);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
case SAVAGE_PRIM_TRISTRIP:
|
|
case SAVAGE_PRIM_TRIFAN:
|
|
if (n < 3) {
|
|
DRM_ERROR
|
|
("wrong number of vertices %u in TRIFAN/STRIP\n",
|
|
n);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
default:
|
|
DRM_ERROR("invalid primitive type %u\n", prim);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (S3_SAVAGE3D_SERIES(dev_priv->chipset)) {
|
|
if (skip != 0) {
|
|
DRM_ERROR("invalid skip flags 0x%04x for DMA\n", skip);
|
|
return -EINVAL;
|
|
}
|
|
} else {
|
|
unsigned int size = 10 - (skip & 1) - (skip >> 1 & 1) -
|
|
(skip >> 2 & 1) - (skip >> 3 & 1) - (skip >> 4 & 1) -
|
|
(skip >> 5 & 1) - (skip >> 6 & 1) - (skip >> 7 & 1);
|
|
if (skip > SAVAGE_SKIP_ALL_S4 || size != 8) {
|
|
DRM_ERROR("invalid skip flags 0x%04x for DMA\n", skip);
|
|
return -EINVAL;
|
|
}
|
|
if (reorder) {
|
|
DRM_ERROR("TRILIST_201 used on Savage4 hardware\n");
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
if (start + n > dmabuf->total / 32) {
|
|
DRM_ERROR("vertex indices (%u-%u) out of range (0-%u)\n",
|
|
start, start + n - 1, dmabuf->total / 32);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Vertex DMA doesn't work with command DMA at the same time,
|
|
* so we use BCI_... to submit commands here. Flush buffered
|
|
* faked DMA first. */
|
|
DMA_FLUSH();
|
|
|
|
if (dmabuf->bus_address != dev_priv->state.common.vbaddr) {
|
|
BEGIN_BCI(2);
|
|
BCI_SET_REGISTERS(SAVAGE_VERTBUFADDR, 1);
|
|
BCI_WRITE(dmabuf->bus_address | dev_priv->dma_type);
|
|
dev_priv->state.common.vbaddr = dmabuf->bus_address;
|
|
}
|
|
if (S3_SAVAGE3D_SERIES(dev_priv->chipset) && dev_priv->waiting) {
|
|
/* Workaround for what looks like a hardware bug. If a
|
|
* WAIT_3D_IDLE was emitted some time before the
|
|
* indexed drawing command then the engine will lock
|
|
* up. There are two known workarounds:
|
|
* WAIT_IDLE_EMPTY or emit at least 63 NOPs. */
|
|
BEGIN_BCI(63);
|
|
for (i = 0; i < 63; ++i)
|
|
BCI_WRITE(BCI_CMD_WAIT);
|
|
dev_priv->waiting = 0;
|
|
}
|
|
|
|
prim <<= 25;
|
|
while (n != 0) {
|
|
/* Can emit up to 255 indices (85 triangles) at once. */
|
|
unsigned int count = n > 255 ? 255 : n;
|
|
if (reorder) {
|
|
/* Need to reorder indices for correct flat
|
|
* shading while preserving the clock sense
|
|
* for correct culling. Only on Savage3D. */
|
|
int reorder[3] = { -1, -1, -1 };
|
|
reorder[start % 3] = 2;
|
|
|
|
BEGIN_BCI((count + 1 + 1) / 2);
|
|
BCI_DRAW_INDICES_S3D(count, prim, start + 2);
|
|
|
|
for (i = start + 1; i + 1 < start + count; i += 2)
|
|
BCI_WRITE((i + reorder[i % 3]) |
|
|
((i + 1 +
|
|
reorder[(i + 1) % 3]) << 16));
|
|
if (i < start + count)
|
|
BCI_WRITE(i + reorder[i % 3]);
|
|
} else if (S3_SAVAGE3D_SERIES(dev_priv->chipset)) {
|
|
BEGIN_BCI((count + 1 + 1) / 2);
|
|
BCI_DRAW_INDICES_S3D(count, prim, start);
|
|
|
|
for (i = start + 1; i + 1 < start + count; i += 2)
|
|
BCI_WRITE(i | ((i + 1) << 16));
|
|
if (i < start + count)
|
|
BCI_WRITE(i);
|
|
} else {
|
|
BEGIN_BCI((count + 2 + 1) / 2);
|
|
BCI_DRAW_INDICES_S4(count, prim, skip);
|
|
|
|
for (i = start; i + 1 < start + count; i += 2)
|
|
BCI_WRITE(i | ((i + 1) << 16));
|
|
if (i < start + count)
|
|
BCI_WRITE(i);
|
|
}
|
|
|
|
start += count;
|
|
n -= count;
|
|
|
|
prim |= BCI_CMD_DRAW_CONT;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int savage_dispatch_vb_prim(drm_savage_private_t * dev_priv,
|
|
const drm_savage_cmd_header_t * cmd_header,
|
|
const uint32_t *vtxbuf, unsigned int vb_size,
|
|
unsigned int vb_stride)
|
|
{
|
|
unsigned char reorder = 0;
|
|
unsigned int prim = cmd_header->prim.prim;
|
|
unsigned int skip = cmd_header->prim.skip;
|
|
unsigned int n = cmd_header->prim.count;
|
|
unsigned int start = cmd_header->prim.start;
|
|
unsigned int vtx_size;
|
|
unsigned int i;
|
|
DMA_LOCALS;
|
|
|
|
if (!n)
|
|
return 0;
|
|
|
|
switch (prim) {
|
|
case SAVAGE_PRIM_TRILIST_201:
|
|
reorder = 1;
|
|
prim = SAVAGE_PRIM_TRILIST;
|
|
case SAVAGE_PRIM_TRILIST:
|
|
if (n % 3 != 0) {
|
|
DRM_ERROR("wrong number of vertices %u in TRILIST\n",
|
|
n);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
case SAVAGE_PRIM_TRISTRIP:
|
|
case SAVAGE_PRIM_TRIFAN:
|
|
if (n < 3) {
|
|
DRM_ERROR
|
|
("wrong number of vertices %u in TRIFAN/STRIP\n",
|
|
n);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
default:
|
|
DRM_ERROR("invalid primitive type %u\n", prim);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (S3_SAVAGE3D_SERIES(dev_priv->chipset)) {
|
|
if (skip > SAVAGE_SKIP_ALL_S3D) {
|
|
DRM_ERROR("invalid skip flags 0x%04x\n", skip);
|
|
return -EINVAL;
|
|
}
|
|
vtx_size = 8; /* full vertex */
|
|
} else {
|
|
if (skip > SAVAGE_SKIP_ALL_S4) {
|
|
DRM_ERROR("invalid skip flags 0x%04x\n", skip);
|
|
return -EINVAL;
|
|
}
|
|
vtx_size = 10; /* full vertex */
|
|
}
|
|
|
|
vtx_size -= (skip & 1) + (skip >> 1 & 1) +
|
|
(skip >> 2 & 1) + (skip >> 3 & 1) + (skip >> 4 & 1) +
|
|
(skip >> 5 & 1) + (skip >> 6 & 1) + (skip >> 7 & 1);
|
|
|
|
if (vtx_size > vb_stride) {
|
|
DRM_ERROR("vertex size greater than vb stride (%u > %u)\n",
|
|
vtx_size, vb_stride);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (start + n > vb_size / (vb_stride * 4)) {
|
|
DRM_ERROR("vertex indices (%u-%u) out of range (0-%u)\n",
|
|
start, start + n - 1, vb_size / (vb_stride * 4));
|
|
return -EINVAL;
|
|
}
|
|
|
|
prim <<= 25;
|
|
while (n != 0) {
|
|
/* Can emit up to 255 vertices (85 triangles) at once. */
|
|
unsigned int count = n > 255 ? 255 : n;
|
|
if (reorder) {
|
|
/* Need to reorder vertices for correct flat
|
|
* shading while preserving the clock sense
|
|
* for correct culling. Only on Savage3D. */
|
|
int reorder[3] = { -1, -1, -1 };
|
|
reorder[start % 3] = 2;
|
|
|
|
BEGIN_DMA(count * vtx_size + 1);
|
|
DMA_DRAW_PRIMITIVE(count, prim, skip);
|
|
|
|
for (i = start; i < start + count; ++i) {
|
|
unsigned int j = i + reorder[i % 3];
|
|
DMA_COPY(&vtxbuf[vb_stride * j], vtx_size);
|
|
}
|
|
|
|
DMA_COMMIT();
|
|
} else {
|
|
BEGIN_DMA(count * vtx_size + 1);
|
|
DMA_DRAW_PRIMITIVE(count, prim, skip);
|
|
|
|
if (vb_stride == vtx_size) {
|
|
DMA_COPY(&vtxbuf[vb_stride * start],
|
|
vtx_size * count);
|
|
} else {
|
|
for (i = start; i < start + count; ++i) {
|
|
DMA_COPY(&vtxbuf [vb_stride * i],
|
|
vtx_size);
|
|
}
|
|
}
|
|
|
|
DMA_COMMIT();
|
|
}
|
|
|
|
start += count;
|
|
n -= count;
|
|
|
|
prim |= BCI_CMD_DRAW_CONT;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int savage_dispatch_dma_idx(drm_savage_private_t * dev_priv,
|
|
const drm_savage_cmd_header_t * cmd_header,
|
|
const uint16_t *idx,
|
|
const struct drm_buf * dmabuf)
|
|
{
|
|
unsigned char reorder = 0;
|
|
unsigned int prim = cmd_header->idx.prim;
|
|
unsigned int skip = cmd_header->idx.skip;
|
|
unsigned int n = cmd_header->idx.count;
|
|
unsigned int i;
|
|
BCI_LOCALS;
|
|
|
|
if (!dmabuf) {
|
|
DRM_ERROR("called without dma buffers!\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!n)
|
|
return 0;
|
|
|
|
switch (prim) {
|
|
case SAVAGE_PRIM_TRILIST_201:
|
|
reorder = 1;
|
|
prim = SAVAGE_PRIM_TRILIST;
|
|
case SAVAGE_PRIM_TRILIST:
|
|
if (n % 3 != 0) {
|
|
DRM_ERROR("wrong number of indices %u in TRILIST\n", n);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
case SAVAGE_PRIM_TRISTRIP:
|
|
case SAVAGE_PRIM_TRIFAN:
|
|
if (n < 3) {
|
|
DRM_ERROR
|
|
("wrong number of indices %u in TRIFAN/STRIP\n", n);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
default:
|
|
DRM_ERROR("invalid primitive type %u\n", prim);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (S3_SAVAGE3D_SERIES(dev_priv->chipset)) {
|
|
if (skip != 0) {
|
|
DRM_ERROR("invalid skip flags 0x%04x for DMA\n", skip);
|
|
return -EINVAL;
|
|
}
|
|
} else {
|
|
unsigned int size = 10 - (skip & 1) - (skip >> 1 & 1) -
|
|
(skip >> 2 & 1) - (skip >> 3 & 1) - (skip >> 4 & 1) -
|
|
(skip >> 5 & 1) - (skip >> 6 & 1) - (skip >> 7 & 1);
|
|
if (skip > SAVAGE_SKIP_ALL_S4 || size != 8) {
|
|
DRM_ERROR("invalid skip flags 0x%04x for DMA\n", skip);
|
|
return -EINVAL;
|
|
}
|
|
if (reorder) {
|
|
DRM_ERROR("TRILIST_201 used on Savage4 hardware\n");
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
/* Vertex DMA doesn't work with command DMA at the same time,
|
|
* so we use BCI_... to submit commands here. Flush buffered
|
|
* faked DMA first. */
|
|
DMA_FLUSH();
|
|
|
|
if (dmabuf->bus_address != dev_priv->state.common.vbaddr) {
|
|
BEGIN_BCI(2);
|
|
BCI_SET_REGISTERS(SAVAGE_VERTBUFADDR, 1);
|
|
BCI_WRITE(dmabuf->bus_address | dev_priv->dma_type);
|
|
dev_priv->state.common.vbaddr = dmabuf->bus_address;
|
|
}
|
|
if (S3_SAVAGE3D_SERIES(dev_priv->chipset) && dev_priv->waiting) {
|
|
/* Workaround for what looks like a hardware bug. If a
|
|
* WAIT_3D_IDLE was emitted some time before the
|
|
* indexed drawing command then the engine will lock
|
|
* up. There are two known workarounds:
|
|
* WAIT_IDLE_EMPTY or emit at least 63 NOPs. */
|
|
BEGIN_BCI(63);
|
|
for (i = 0; i < 63; ++i)
|
|
BCI_WRITE(BCI_CMD_WAIT);
|
|
dev_priv->waiting = 0;
|
|
}
|
|
|
|
prim <<= 25;
|
|
while (n != 0) {
|
|
/* Can emit up to 255 indices (85 triangles) at once. */
|
|
unsigned int count = n > 255 ? 255 : n;
|
|
|
|
/* check indices */
|
|
for (i = 0; i < count; ++i) {
|
|
if (idx[i] > dmabuf->total / 32) {
|
|
DRM_ERROR("idx[%u]=%u out of range (0-%u)\n",
|
|
i, idx[i], dmabuf->total / 32);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
if (reorder) {
|
|
/* Need to reorder indices for correct flat
|
|
* shading while preserving the clock sense
|
|
* for correct culling. Only on Savage3D. */
|
|
int reorder[3] = { 2, -1, -1 };
|
|
|
|
BEGIN_BCI((count + 1 + 1) / 2);
|
|
BCI_DRAW_INDICES_S3D(count, prim, idx[2]);
|
|
|
|
for (i = 1; i + 1 < count; i += 2)
|
|
BCI_WRITE(idx[i + reorder[i % 3]] |
|
|
(idx[i + 1 +
|
|
reorder[(i + 1) % 3]] << 16));
|
|
if (i < count)
|
|
BCI_WRITE(idx[i + reorder[i % 3]]);
|
|
} else if (S3_SAVAGE3D_SERIES(dev_priv->chipset)) {
|
|
BEGIN_BCI((count + 1 + 1) / 2);
|
|
BCI_DRAW_INDICES_S3D(count, prim, idx[0]);
|
|
|
|
for (i = 1; i + 1 < count; i += 2)
|
|
BCI_WRITE(idx[i] | (idx[i + 1] << 16));
|
|
if (i < count)
|
|
BCI_WRITE(idx[i]);
|
|
} else {
|
|
BEGIN_BCI((count + 2 + 1) / 2);
|
|
BCI_DRAW_INDICES_S4(count, prim, skip);
|
|
|
|
for (i = 0; i + 1 < count; i += 2)
|
|
BCI_WRITE(idx[i] | (idx[i + 1] << 16));
|
|
if (i < count)
|
|
BCI_WRITE(idx[i]);
|
|
}
|
|
|
|
idx += count;
|
|
n -= count;
|
|
|
|
prim |= BCI_CMD_DRAW_CONT;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int savage_dispatch_vb_idx(drm_savage_private_t * dev_priv,
|
|
const drm_savage_cmd_header_t * cmd_header,
|
|
const uint16_t *idx,
|
|
const uint32_t *vtxbuf,
|
|
unsigned int vb_size, unsigned int vb_stride)
|
|
{
|
|
unsigned char reorder = 0;
|
|
unsigned int prim = cmd_header->idx.prim;
|
|
unsigned int skip = cmd_header->idx.skip;
|
|
unsigned int n = cmd_header->idx.count;
|
|
unsigned int vtx_size;
|
|
unsigned int i;
|
|
DMA_LOCALS;
|
|
|
|
if (!n)
|
|
return 0;
|
|
|
|
switch (prim) {
|
|
case SAVAGE_PRIM_TRILIST_201:
|
|
reorder = 1;
|
|
prim = SAVAGE_PRIM_TRILIST;
|
|
case SAVAGE_PRIM_TRILIST:
|
|
if (n % 3 != 0) {
|
|
DRM_ERROR("wrong number of indices %u in TRILIST\n", n);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
case SAVAGE_PRIM_TRISTRIP:
|
|
case SAVAGE_PRIM_TRIFAN:
|
|
if (n < 3) {
|
|
DRM_ERROR
|
|
("wrong number of indices %u in TRIFAN/STRIP\n", n);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
default:
|
|
DRM_ERROR("invalid primitive type %u\n", prim);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (S3_SAVAGE3D_SERIES(dev_priv->chipset)) {
|
|
if (skip > SAVAGE_SKIP_ALL_S3D) {
|
|
DRM_ERROR("invalid skip flags 0x%04x\n", skip);
|
|
return -EINVAL;
|
|
}
|
|
vtx_size = 8; /* full vertex */
|
|
} else {
|
|
if (skip > SAVAGE_SKIP_ALL_S4) {
|
|
DRM_ERROR("invalid skip flags 0x%04x\n", skip);
|
|
return -EINVAL;
|
|
}
|
|
vtx_size = 10; /* full vertex */
|
|
}
|
|
|
|
vtx_size -= (skip & 1) + (skip >> 1 & 1) +
|
|
(skip >> 2 & 1) + (skip >> 3 & 1) + (skip >> 4 & 1) +
|
|
(skip >> 5 & 1) + (skip >> 6 & 1) + (skip >> 7 & 1);
|
|
|
|
if (vtx_size > vb_stride) {
|
|
DRM_ERROR("vertex size greater than vb stride (%u > %u)\n",
|
|
vtx_size, vb_stride);
|
|
return -EINVAL;
|
|
}
|
|
|
|
prim <<= 25;
|
|
while (n != 0) {
|
|
/* Can emit up to 255 vertices (85 triangles) at once. */
|
|
unsigned int count = n > 255 ? 255 : n;
|
|
|
|
/* Check indices */
|
|
for (i = 0; i < count; ++i) {
|
|
if (idx[i] > vb_size / (vb_stride * 4)) {
|
|
DRM_ERROR("idx[%u]=%u out of range (0-%u)\n",
|
|
i, idx[i], vb_size / (vb_stride * 4));
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
if (reorder) {
|
|
/* Need to reorder vertices for correct flat
|
|
* shading while preserving the clock sense
|
|
* for correct culling. Only on Savage3D. */
|
|
int reorder[3] = { 2, -1, -1 };
|
|
|
|
BEGIN_DMA(count * vtx_size + 1);
|
|
DMA_DRAW_PRIMITIVE(count, prim, skip);
|
|
|
|
for (i = 0; i < count; ++i) {
|
|
unsigned int j = idx[i + reorder[i % 3]];
|
|
DMA_COPY(&vtxbuf[vb_stride * j], vtx_size);
|
|
}
|
|
|
|
DMA_COMMIT();
|
|
} else {
|
|
BEGIN_DMA(count * vtx_size + 1);
|
|
DMA_DRAW_PRIMITIVE(count, prim, skip);
|
|
|
|
for (i = 0; i < count; ++i) {
|
|
unsigned int j = idx[i];
|
|
DMA_COPY(&vtxbuf[vb_stride * j], vtx_size);
|
|
}
|
|
|
|
DMA_COMMIT();
|
|
}
|
|
|
|
idx += count;
|
|
n -= count;
|
|
|
|
prim |= BCI_CMD_DRAW_CONT;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int savage_dispatch_clear(drm_savage_private_t * dev_priv,
|
|
const drm_savage_cmd_header_t * cmd_header,
|
|
const drm_savage_cmd_header_t *data,
|
|
unsigned int nbox,
|
|
const struct drm_clip_rect *boxes)
|
|
{
|
|
unsigned int flags = cmd_header->clear0.flags;
|
|
unsigned int clear_cmd;
|
|
unsigned int i, nbufs;
|
|
DMA_LOCALS;
|
|
|
|
if (nbox == 0)
|
|
return 0;
|
|
|
|
clear_cmd = BCI_CMD_RECT | BCI_CMD_RECT_XP | BCI_CMD_RECT_YP |
|
|
BCI_CMD_SEND_COLOR | BCI_CMD_DEST_PBD_NEW;
|
|
BCI_CMD_SET_ROP(clear_cmd, 0xCC);
|
|
|
|
nbufs = ((flags & SAVAGE_FRONT) ? 1 : 0) +
|
|
((flags & SAVAGE_BACK) ? 1 : 0) + ((flags & SAVAGE_DEPTH) ? 1 : 0);
|
|
if (nbufs == 0)
|
|
return 0;
|
|
|
|
if (data->clear1.mask != 0xffffffff) {
|
|
/* set mask */
|
|
BEGIN_DMA(2);
|
|
DMA_SET_REGISTERS(SAVAGE_BITPLANEWTMASK, 1);
|
|
DMA_WRITE(data->clear1.mask);
|
|
DMA_COMMIT();
|
|
}
|
|
for (i = 0; i < nbox; ++i) {
|
|
unsigned int x, y, w, h;
|
|
unsigned int buf;
|
|
x = boxes[i].x1, y = boxes[i].y1;
|
|
w = boxes[i].x2 - boxes[i].x1;
|
|
h = boxes[i].y2 - boxes[i].y1;
|
|
BEGIN_DMA(nbufs * 6);
|
|
for (buf = SAVAGE_FRONT; buf <= SAVAGE_DEPTH; buf <<= 1) {
|
|
if (!(flags & buf))
|
|
continue;
|
|
DMA_WRITE(clear_cmd);
|
|
switch (buf) {
|
|
case SAVAGE_FRONT:
|
|
DMA_WRITE(dev_priv->front_offset);
|
|
DMA_WRITE(dev_priv->front_bd);
|
|
break;
|
|
case SAVAGE_BACK:
|
|
DMA_WRITE(dev_priv->back_offset);
|
|
DMA_WRITE(dev_priv->back_bd);
|
|
break;
|
|
case SAVAGE_DEPTH:
|
|
DMA_WRITE(dev_priv->depth_offset);
|
|
DMA_WRITE(dev_priv->depth_bd);
|
|
break;
|
|
}
|
|
DMA_WRITE(data->clear1.value);
|
|
DMA_WRITE(BCI_X_Y(x, y));
|
|
DMA_WRITE(BCI_W_H(w, h));
|
|
}
|
|
DMA_COMMIT();
|
|
}
|
|
if (data->clear1.mask != 0xffffffff) {
|
|
/* reset mask */
|
|
BEGIN_DMA(2);
|
|
DMA_SET_REGISTERS(SAVAGE_BITPLANEWTMASK, 1);
|
|
DMA_WRITE(0xffffffff);
|
|
DMA_COMMIT();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int savage_dispatch_swap(drm_savage_private_t * dev_priv,
|
|
unsigned int nbox, const struct drm_clip_rect *boxes)
|
|
{
|
|
unsigned int swap_cmd;
|
|
unsigned int i;
|
|
DMA_LOCALS;
|
|
|
|
if (nbox == 0)
|
|
return 0;
|
|
|
|
swap_cmd = BCI_CMD_RECT | BCI_CMD_RECT_XP | BCI_CMD_RECT_YP |
|
|
BCI_CMD_SRC_PBD_COLOR_NEW | BCI_CMD_DEST_GBD;
|
|
BCI_CMD_SET_ROP(swap_cmd, 0xCC);
|
|
|
|
for (i = 0; i < nbox; ++i) {
|
|
BEGIN_DMA(6);
|
|
DMA_WRITE(swap_cmd);
|
|
DMA_WRITE(dev_priv->back_offset);
|
|
DMA_WRITE(dev_priv->back_bd);
|
|
DMA_WRITE(BCI_X_Y(boxes[i].x1, boxes[i].y1));
|
|
DMA_WRITE(BCI_X_Y(boxes[i].x1, boxes[i].y1));
|
|
DMA_WRITE(BCI_W_H(boxes[i].x2 - boxes[i].x1,
|
|
boxes[i].y2 - boxes[i].y1));
|
|
DMA_COMMIT();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int savage_dispatch_draw(drm_savage_private_t * dev_priv,
|
|
const drm_savage_cmd_header_t *start,
|
|
const drm_savage_cmd_header_t *end,
|
|
const struct drm_buf * dmabuf,
|
|
const unsigned int *vtxbuf,
|
|
unsigned int vb_size, unsigned int vb_stride,
|
|
unsigned int nbox,
|
|
const struct drm_clip_rect *boxes)
|
|
{
|
|
unsigned int i, j;
|
|
int ret;
|
|
|
|
for (i = 0; i < nbox; ++i) {
|
|
const drm_savage_cmd_header_t *cmdbuf;
|
|
dev_priv->emit_clip_rect(dev_priv, &boxes[i]);
|
|
|
|
cmdbuf = start;
|
|
while (cmdbuf < end) {
|
|
drm_savage_cmd_header_t cmd_header;
|
|
cmd_header = *cmdbuf;
|
|
cmdbuf++;
|
|
switch (cmd_header.cmd.cmd) {
|
|
case SAVAGE_CMD_DMA_PRIM:
|
|
ret = savage_dispatch_dma_prim(
|
|
dev_priv, &cmd_header, dmabuf);
|
|
break;
|
|
case SAVAGE_CMD_VB_PRIM:
|
|
ret = savage_dispatch_vb_prim(
|
|
dev_priv, &cmd_header,
|
|
vtxbuf, vb_size, vb_stride);
|
|
break;
|
|
case SAVAGE_CMD_DMA_IDX:
|
|
j = (cmd_header.idx.count + 3) / 4;
|
|
/* j was check in savage_bci_cmdbuf */
|
|
ret = savage_dispatch_dma_idx(dev_priv,
|
|
&cmd_header, (const uint16_t *)cmdbuf,
|
|
dmabuf);
|
|
cmdbuf += j;
|
|
break;
|
|
case SAVAGE_CMD_VB_IDX:
|
|
j = (cmd_header.idx.count + 3) / 4;
|
|
/* j was check in savage_bci_cmdbuf */
|
|
ret = savage_dispatch_vb_idx(dev_priv,
|
|
&cmd_header, (const uint16_t *)cmdbuf,
|
|
(const uint32_t *)vtxbuf, vb_size,
|
|
vb_stride);
|
|
cmdbuf += j;
|
|
break;
|
|
default:
|
|
/* What's the best return code? EFAULT? */
|
|
DRM_ERROR("IMPLEMENTATION ERROR: "
|
|
"non-drawing-command %d\n",
|
|
cmd_header.cmd.cmd);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (ret != 0)
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int savage_bci_cmdbuf(struct drm_device *dev, void *data, struct drm_file *file_priv)
|
|
{
|
|
drm_savage_private_t *dev_priv = dev->dev_private;
|
|
struct drm_device_dma *dma = dev->dma;
|
|
struct drm_buf *dmabuf;
|
|
drm_savage_cmdbuf_t *cmdbuf = data;
|
|
drm_savage_cmd_header_t *kcmd_addr = NULL;
|
|
drm_savage_cmd_header_t *first_draw_cmd;
|
|
unsigned int *kvb_addr = NULL;
|
|
struct drm_clip_rect *kbox_addr = NULL;
|
|
unsigned int i, j;
|
|
int ret = 0;
|
|
|
|
DRM_DEBUG("\n");
|
|
|
|
LOCK_TEST_WITH_RETURN(dev, file_priv);
|
|
|
|
if (dma && dma->buflist) {
|
|
if (cmdbuf->dma_idx > dma->buf_count) {
|
|
DRM_ERROR
|
|
("vertex buffer index %u out of range (0-%u)\n",
|
|
cmdbuf->dma_idx, dma->buf_count - 1);
|
|
return -EINVAL;
|
|
}
|
|
dmabuf = dma->buflist[cmdbuf->dma_idx];
|
|
} else {
|
|
dmabuf = NULL;
|
|
}
|
|
|
|
/* Copy the user buffers into kernel temporary areas. This hasn't been
|
|
* a performance loss compared to VERIFYAREA_READ/
|
|
* COPY_FROM_USER_UNCHECKED when done in other drivers, and is correct
|
|
* for locking on FreeBSD.
|
|
*/
|
|
if (cmdbuf->size) {
|
|
kcmd_addr = kmalloc_array(cmdbuf->size, 8, GFP_KERNEL);
|
|
if (kcmd_addr == NULL)
|
|
return -ENOMEM;
|
|
|
|
if (copy_from_user(kcmd_addr, cmdbuf->cmd_addr,
|
|
cmdbuf->size * 8))
|
|
{
|
|
kfree(kcmd_addr);
|
|
return -EFAULT;
|
|
}
|
|
cmdbuf->cmd_addr = kcmd_addr;
|
|
}
|
|
if (cmdbuf->vb_size) {
|
|
kvb_addr = kmalloc(cmdbuf->vb_size, GFP_KERNEL);
|
|
if (kvb_addr == NULL) {
|
|
ret = -ENOMEM;
|
|
goto done;
|
|
}
|
|
|
|
if (copy_from_user(kvb_addr, cmdbuf->vb_addr,
|
|
cmdbuf->vb_size)) {
|
|
ret = -EFAULT;
|
|
goto done;
|
|
}
|
|
cmdbuf->vb_addr = kvb_addr;
|
|
}
|
|
if (cmdbuf->nbox) {
|
|
kbox_addr = kmalloc_array(cmdbuf->nbox, sizeof(struct drm_clip_rect),
|
|
GFP_KERNEL);
|
|
if (kbox_addr == NULL) {
|
|
ret = -ENOMEM;
|
|
goto done;
|
|
}
|
|
|
|
if (copy_from_user(kbox_addr, cmdbuf->box_addr,
|
|
cmdbuf->nbox * sizeof(struct drm_clip_rect))) {
|
|
ret = -EFAULT;
|
|
goto done;
|
|
}
|
|
cmdbuf->box_addr = kbox_addr;
|
|
}
|
|
|
|
/* Make sure writes to DMA buffers are finished before sending
|
|
* DMA commands to the graphics hardware. */
|
|
mb();
|
|
|
|
/* Coming from user space. Don't know if the Xserver has
|
|
* emitted wait commands. Assuming the worst. */
|
|
dev_priv->waiting = 1;
|
|
|
|
i = 0;
|
|
first_draw_cmd = NULL;
|
|
while (i < cmdbuf->size) {
|
|
drm_savage_cmd_header_t cmd_header;
|
|
cmd_header = *(drm_savage_cmd_header_t *)cmdbuf->cmd_addr;
|
|
cmdbuf->cmd_addr++;
|
|
i++;
|
|
|
|
/* Group drawing commands with same state to minimize
|
|
* iterations over clip rects. */
|
|
j = 0;
|
|
switch (cmd_header.cmd.cmd) {
|
|
case SAVAGE_CMD_DMA_IDX:
|
|
case SAVAGE_CMD_VB_IDX:
|
|
j = (cmd_header.idx.count + 3) / 4;
|
|
if (i + j > cmdbuf->size) {
|
|
DRM_ERROR("indexed drawing command extends "
|
|
"beyond end of command buffer\n");
|
|
DMA_FLUSH();
|
|
ret = -EINVAL;
|
|
goto done;
|
|
}
|
|
/* fall through */
|
|
case SAVAGE_CMD_DMA_PRIM:
|
|
case SAVAGE_CMD_VB_PRIM:
|
|
if (!first_draw_cmd)
|
|
first_draw_cmd = cmdbuf->cmd_addr - 1;
|
|
cmdbuf->cmd_addr += j;
|
|
i += j;
|
|
break;
|
|
default:
|
|
if (first_draw_cmd) {
|
|
ret = savage_dispatch_draw(
|
|
dev_priv, first_draw_cmd,
|
|
cmdbuf->cmd_addr - 1,
|
|
dmabuf, cmdbuf->vb_addr, cmdbuf->vb_size,
|
|
cmdbuf->vb_stride,
|
|
cmdbuf->nbox, cmdbuf->box_addr);
|
|
if (ret != 0)
|
|
goto done;
|
|
first_draw_cmd = NULL;
|
|
}
|
|
}
|
|
if (first_draw_cmd)
|
|
continue;
|
|
|
|
switch (cmd_header.cmd.cmd) {
|
|
case SAVAGE_CMD_STATE:
|
|
j = (cmd_header.state.count + 1) / 2;
|
|
if (i + j > cmdbuf->size) {
|
|
DRM_ERROR("command SAVAGE_CMD_STATE extends "
|
|
"beyond end of command buffer\n");
|
|
DMA_FLUSH();
|
|
ret = -EINVAL;
|
|
goto done;
|
|
}
|
|
ret = savage_dispatch_state(dev_priv, &cmd_header,
|
|
(const uint32_t *)cmdbuf->cmd_addr);
|
|
cmdbuf->cmd_addr += j;
|
|
i += j;
|
|
break;
|
|
case SAVAGE_CMD_CLEAR:
|
|
if (i + 1 > cmdbuf->size) {
|
|
DRM_ERROR("command SAVAGE_CMD_CLEAR extends "
|
|
"beyond end of command buffer\n");
|
|
DMA_FLUSH();
|
|
ret = -EINVAL;
|
|
goto done;
|
|
}
|
|
ret = savage_dispatch_clear(dev_priv, &cmd_header,
|
|
cmdbuf->cmd_addr,
|
|
cmdbuf->nbox,
|
|
cmdbuf->box_addr);
|
|
cmdbuf->cmd_addr++;
|
|
i++;
|
|
break;
|
|
case SAVAGE_CMD_SWAP:
|
|
ret = savage_dispatch_swap(dev_priv, cmdbuf->nbox,
|
|
cmdbuf->box_addr);
|
|
break;
|
|
default:
|
|
DRM_ERROR("invalid command 0x%x\n",
|
|
cmd_header.cmd.cmd);
|
|
DMA_FLUSH();
|
|
ret = -EINVAL;
|
|
goto done;
|
|
}
|
|
|
|
if (ret != 0) {
|
|
DMA_FLUSH();
|
|
goto done;
|
|
}
|
|
}
|
|
|
|
if (first_draw_cmd) {
|
|
ret = savage_dispatch_draw (
|
|
dev_priv, first_draw_cmd, cmdbuf->cmd_addr, dmabuf,
|
|
cmdbuf->vb_addr, cmdbuf->vb_size, cmdbuf->vb_stride,
|
|
cmdbuf->nbox, cmdbuf->box_addr);
|
|
if (ret != 0) {
|
|
DMA_FLUSH();
|
|
goto done;
|
|
}
|
|
}
|
|
|
|
DMA_FLUSH();
|
|
|
|
if (dmabuf && cmdbuf->discard) {
|
|
drm_savage_buf_priv_t *buf_priv = dmabuf->dev_private;
|
|
uint16_t event;
|
|
event = savage_bci_emit_event(dev_priv, SAVAGE_WAIT_3D);
|
|
SET_AGE(&buf_priv->age, event, dev_priv->event_wrap);
|
|
savage_freelist_put(dev, dmabuf);
|
|
}
|
|
|
|
done:
|
|
/* If we didn't need to allocate them, these'll be NULL */
|
|
kfree(kcmd_addr);
|
|
kfree(kvb_addr);
|
|
kfree(kbox_addr);
|
|
|
|
return ret;
|
|
}
|