diff options
Diffstat (limited to 'xc/lib/GL/mesa/src/drv/r128/r128_context.c')
-rw-r--r-- | xc/lib/GL/mesa/src/drv/r128/r128_context.c | 18 |
1 files changed, 8 insertions, 10 deletions
diff --git a/xc/lib/GL/mesa/src/drv/r128/r128_context.c b/xc/lib/GL/mesa/src/drv/r128/r128_context.c index 3d9ba4ddd..9bbab1dc5 100644 --- a/xc/lib/GL/mesa/src/drv/r128/r128_context.c +++ b/xc/lib/GL/mesa/src/drv/r128/r128_context.c @@ -84,12 +84,7 @@ GLboolean r128CreateContext(Display *dpy, GLvisual *glVisual, r128ctx->useFastPath = GL_FALSE; - r128ctx->vb.start = 0; - r128ctx->vb.count = 0; - r128ctx->vb.size = 0; - r128ctx->vb.index = 0; - r128ctx->vb.buf = NULL; - r128ctx->vb.done = GL_TRUE; + r128ctx->vert_buf = NULL; r128ctx->CCEbuf = (CARD32 *)malloc(sizeof(*r128ctx->CCEbuf) * r128scrn->ringEntries); @@ -118,7 +113,10 @@ GLboolean r128CreateContext(Display *dpy, GLvisual *glVisual, r128ctx->Fallback |= R128_FALLBACK_SWONLY; /* KW: No vertex buffer code for PCI cards -- for now fallback to - * software always. + * software always. + * GTH: There is support there, but I'm seeing strange behaviour + * with it enabled. I'll leave the software fallbacks in place + * for now. */ if (r128scrn->IsPCI) r128ctx->Fallback |= R128_FALLBACK_SWONLY; @@ -139,17 +137,17 @@ GLboolean r128CreateContext(Display *dpy, GLvisual *glVisual, /* KW: Set the maximum texture size small enough that we can * guarentee that both texture units can bind a maximal texture * and have them both in on-card memory at once. (Kevin or - * Gareth: Please check these numbers are OK) + * Gareth: Please check these numbers are OK) */ if (r128scrn->texSize[0] < 2*1024*1024) { glCtx->Const.MaxTextureLevels = 9; glCtx->Const.MaxTextureSize = 1<<8; } else if (r128scrn->texSize[0] < 8*1024*1024) { glCtx->Const.MaxTextureLevels = 10; - glCtx->Const.MaxTextureSize = 1<<9; + glCtx->Const.MaxTextureSize = 1<<9; } else { glCtx->Const.MaxTextureLevels = 11; - glCtx->Const.MaxTextureSize = 1<<10; + glCtx->Const.MaxTextureSize = 1<<10; } glCtx->Const.MaxTextureUnits = 2; |