mirror of
https://gitlab.freedesktop.org/mesa/mesa.git
synced 2026-02-03 00:00:25 +01:00
[intel] Clean up cliprect handling in intel drivers.
In particular, batch buffers are no longer flushed when switching from CLIPRECTS to NO_CLIPRECTS or vice versa, and 965 just uses DRM cliprect handling for primitives instead of trying to sneak in its own to avoid the DRM stuff. The disadvantage is that we will re-execute state updates per cliprect, but the advantage is that we will be able to accumulate larger batch buffers, which were proving to be a major overhead.
This commit is contained in:
parent
7ce12b0863
commit
beddf653a9
5 changed files with 57 additions and 73 deletions
|
|
@ -129,11 +129,11 @@ void
|
|||
intelWrapInlinePrimitive(struct intel_context *intel)
|
||||
{
|
||||
GLuint prim = intel->prim.primitive;
|
||||
GLuint batchflags = intel->batch->flags;
|
||||
GLuint cliprects_enable = intel->batch->cliprects_enable;
|
||||
|
||||
intel_flush_inline_primitive(intel);
|
||||
intel_batchbuffer_flush(intel->batch);
|
||||
intelStartInlinePrimitive(intel, prim, batchflags); /* ??? */
|
||||
intelStartInlinePrimitive(intel, prim, cliprects_enable); /* ??? */
|
||||
}
|
||||
|
||||
GLuint *
|
||||
|
|
|
|||
|
|
@ -123,25 +123,6 @@ static GLuint trim(GLenum prim, GLuint length)
|
|||
}
|
||||
|
||||
|
||||
static void brw_emit_cliprect( struct brw_context *brw,
|
||||
const drm_clip_rect_t *rect )
|
||||
{
|
||||
struct brw_drawrect bdr;
|
||||
|
||||
bdr.header.opcode = CMD_DRAW_RECT;
|
||||
bdr.header.length = sizeof(bdr)/4 - 2;
|
||||
bdr.xmin = rect->x1;
|
||||
bdr.xmax = rect->x2 - 1;
|
||||
bdr.ymin = rect->y1;
|
||||
bdr.ymax = rect->y2 - 1;
|
||||
bdr.xorg = brw->intel.drawX;
|
||||
bdr.yorg = brw->intel.drawY;
|
||||
|
||||
intel_batchbuffer_data( brw->intel.batch, &bdr, sizeof(bdr),
|
||||
INTEL_BATCH_NO_CLIPRECTS);
|
||||
}
|
||||
|
||||
|
||||
static void brw_emit_prim( struct brw_context *brw,
|
||||
const struct _mesa_prim *prim )
|
||||
|
||||
|
|
@ -165,8 +146,8 @@ static void brw_emit_prim( struct brw_context *brw,
|
|||
prim_packet.base_vert_location = 0;
|
||||
|
||||
if (prim_packet.verts_per_instance) {
|
||||
intel_batchbuffer_data( brw->intel.batch, &prim_packet, sizeof(prim_packet),
|
||||
INTEL_BATCH_NO_CLIPRECTS);
|
||||
intel_batchbuffer_data( brw->intel.batch, &prim_packet,
|
||||
sizeof(prim_packet), INTEL_BATCH_CLIPRECTS);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
@ -270,7 +251,7 @@ static GLboolean brw_try_draw_prims( GLcontext *ctx,
|
|||
struct intel_context *intel = intel_context(ctx);
|
||||
struct brw_context *brw = brw_context(ctx);
|
||||
GLboolean retval = GL_FALSE;
|
||||
GLuint i, j;
|
||||
GLuint i;
|
||||
|
||||
if (ctx->NewState)
|
||||
_mesa_update_state( ctx );
|
||||
|
|
@ -320,31 +301,8 @@ static GLboolean brw_try_draw_prims( GLcontext *ctx,
|
|||
goto out;
|
||||
}
|
||||
|
||||
/* For single cliprect, state is already emitted:
|
||||
*/
|
||||
if (brw->intel.numClipRects == 1) {
|
||||
for (i = 0; i < nr_prims; i++) {
|
||||
brw_emit_prim(brw, &prim[i]);
|
||||
}
|
||||
}
|
||||
else {
|
||||
/* Otherwise, explicitly do the cliprects at this point:
|
||||
*/
|
||||
GLuint nprims = 0;
|
||||
for (j = 0; j < brw->intel.numClipRects; j++) {
|
||||
brw_emit_cliprect(brw, &brw->intel.pClipRects[j]);
|
||||
|
||||
/* Emit prims to batchbuffer:
|
||||
*/
|
||||
for (i = 0; i < nr_prims; i++) {
|
||||
brw_emit_prim(brw, &prim[i]);
|
||||
|
||||
if (++nprims == VBO_MAX_PRIM) {
|
||||
intel_batchbuffer_flush(brw->intel.batch);
|
||||
nprims = 0;
|
||||
}
|
||||
}
|
||||
}
|
||||
for (i = 0; i < nr_prims; i++) {
|
||||
brw_emit_prim(brw, &prim[i]);
|
||||
}
|
||||
|
||||
retval = GL_TRUE;
|
||||
|
|
|
|||
|
|
@ -87,6 +87,7 @@ intel_batchbuffer_reset(struct intel_batchbuffer *batch)
|
|||
batch->ptr = batch->map;
|
||||
batch->dirty_state = ~0;
|
||||
batch->id = batch->intel->batch_id++;
|
||||
batch->cliprects_enable = INTEL_BATCH_NO_CLIPRECTS;
|
||||
}
|
||||
|
||||
struct intel_batchbuffer *
|
||||
|
|
@ -124,8 +125,7 @@ intel_batchbuffer_free(struct intel_batchbuffer *batch)
|
|||
*/
|
||||
static void
|
||||
do_flush_locked(struct intel_batchbuffer *batch,
|
||||
GLuint used,
|
||||
GLboolean ignore_cliprects, GLboolean allow_unlock)
|
||||
GLuint used, GLboolean allow_unlock)
|
||||
{
|
||||
struct intel_context *intel = batch->intel;
|
||||
void *start;
|
||||
|
|
@ -136,28 +136,33 @@ do_flush_locked(struct intel_batchbuffer *batch,
|
|||
|
||||
batch->map = NULL;
|
||||
batch->ptr = NULL;
|
||||
batch->flags = 0;
|
||||
|
||||
/* Throw away non-effective packets. Won't work once we have
|
||||
* hardware contexts which would preserve statechanges beyond a
|
||||
* single buffer.
|
||||
*/
|
||||
|
||||
if (!(intel->numClipRects == 0 && !ignore_cliprects)) {
|
||||
if (!(intel->numClipRects == 0 &&
|
||||
batch->cliprects_enable == INTEL_BATCH_CLIPRECTS)) {
|
||||
if (intel->ttm == GL_TRUE) {
|
||||
intel_exec_ioctl(batch->intel,
|
||||
used, ignore_cliprects, allow_unlock,
|
||||
used,
|
||||
batch->cliprects_enable == INTEL_BATCH_NO_CLIPRECTS,
|
||||
allow_unlock,
|
||||
start, count, &batch->last_fence);
|
||||
} else {
|
||||
intel_batch_ioctl(batch->intel,
|
||||
batch->buf->offset,
|
||||
used, ignore_cliprects, allow_unlock);
|
||||
used,
|
||||
batch->cliprects_enable == INTEL_BATCH_NO_CLIPRECTS,
|
||||
allow_unlock);
|
||||
}
|
||||
}
|
||||
|
||||
dri_post_submit(batch->buf, &batch->last_fence);
|
||||
|
||||
if (intel->numClipRects == 0 && !ignore_cliprects) {
|
||||
if (intel->numClipRects == 0 &&
|
||||
batch->cliprects_enable == INTEL_BATCH_CLIPRECTS) {
|
||||
if (allow_unlock) {
|
||||
/* If we are not doing any actual user-visible rendering,
|
||||
* do a sched_yield to keep the app from pegging the cpu while
|
||||
|
|
@ -212,9 +217,8 @@ intel_batchbuffer_flush(struct intel_batchbuffer *batch)
|
|||
if (!was_locked)
|
||||
LOCK_HARDWARE(intel);
|
||||
|
||||
do_flush_locked(batch, used, !(batch->flags & INTEL_BATCH_CLIPRECTS),
|
||||
GL_FALSE);
|
||||
|
||||
do_flush_locked(batch, used, GL_FALSE);
|
||||
|
||||
if (!was_locked)
|
||||
UNLOCK_HARDWARE(intel);
|
||||
|
||||
|
|
@ -258,10 +262,11 @@ intel_batchbuffer_emit_reloc(struct intel_batchbuffer *batch,
|
|||
|
||||
void
|
||||
intel_batchbuffer_data(struct intel_batchbuffer *batch,
|
||||
const void *data, GLuint bytes, GLuint flags)
|
||||
const void *data, GLuint bytes,
|
||||
enum cliprects_enable cliprects_enable)
|
||||
{
|
||||
assert((bytes & 3) == 0);
|
||||
intel_batchbuffer_require_space(batch, bytes, flags);
|
||||
intel_batchbuffer_require_space(batch, bytes, cliprects_enable);
|
||||
__memcpy(batch->ptr, data, bytes);
|
||||
batch->ptr += bytes;
|
||||
}
|
||||
|
|
|
|||
|
|
@ -10,8 +10,10 @@ struct intel_context;
|
|||
#define BATCH_SZ 16384
|
||||
#define BATCH_RESERVED 16
|
||||
|
||||
#define INTEL_BATCH_NO_CLIPRECTS 0x1
|
||||
#define INTEL_BATCH_CLIPRECTS 0x2
|
||||
enum cliprects_enable {
|
||||
INTEL_BATCH_CLIPRECTS = 0,
|
||||
INTEL_BATCH_NO_CLIPRECTS = 1
|
||||
};
|
||||
|
||||
struct intel_batchbuffer
|
||||
{
|
||||
|
|
@ -19,11 +21,12 @@ struct intel_batchbuffer
|
|||
|
||||
dri_bo *buf;
|
||||
dri_fence *last_fence;
|
||||
GLuint flags;
|
||||
|
||||
GLubyte *map;
|
||||
GLubyte *ptr;
|
||||
|
||||
enum cliprects_enable cliprects_enable;
|
||||
|
||||
GLuint size;
|
||||
|
||||
GLuint dirty_state;
|
||||
|
|
@ -48,7 +51,8 @@ void intel_batchbuffer_reset(struct intel_batchbuffer *batch);
|
|||
* intel_buffer_dword() calls.
|
||||
*/
|
||||
void intel_batchbuffer_data(struct intel_batchbuffer *batch,
|
||||
const void *data, GLuint bytes, GLuint flags);
|
||||
const void *data, GLuint bytes,
|
||||
enum cliprects_enable cliprects_enable);
|
||||
|
||||
void intel_batchbuffer_release_space(struct intel_batchbuffer *batch,
|
||||
GLuint bytes);
|
||||
|
|
@ -80,29 +84,37 @@ intel_batchbuffer_emit_dword(struct intel_batchbuffer *batch, GLuint dword)
|
|||
|
||||
static INLINE void
|
||||
intel_batchbuffer_require_space(struct intel_batchbuffer *batch,
|
||||
GLuint sz, GLuint flags)
|
||||
GLuint sz,
|
||||
enum cliprects_enable cliprects_enable)
|
||||
{
|
||||
assert(sz < batch->size - 8);
|
||||
if (intel_batchbuffer_space(batch) < sz ||
|
||||
(batch->flags != 0 && flags != 0 && batch->flags != flags))
|
||||
if (intel_batchbuffer_space(batch) < sz)
|
||||
intel_batchbuffer_flush(batch);
|
||||
|
||||
batch->flags |= flags;
|
||||
/* Upgrade the buffer to being looped over per cliprect if this batch
|
||||
* emit needs it. The code used to emit a batch whenever the
|
||||
* cliprects_enable was changed, but reducing the overhead of frequent
|
||||
* batch flushing is more important than reducing state parsing,
|
||||
* particularly as we move towards private backbuffers and number
|
||||
* cliprects always being 1 except at swap.
|
||||
*/
|
||||
if (cliprects_enable == INTEL_BATCH_CLIPRECTS)
|
||||
batch->cliprects_enable = INTEL_BATCH_CLIPRECTS;
|
||||
}
|
||||
|
||||
/* Here are the crusty old macros, to be removed:
|
||||
*/
|
||||
#define BATCH_LOCALS
|
||||
|
||||
#define BEGIN_BATCH(n, flags) do { \
|
||||
intel_batchbuffer_require_space(intel->batch, (n)*4, flags); \
|
||||
#define BEGIN_BATCH(n, cliprects_enable) do { \
|
||||
intel_batchbuffer_require_space(intel->batch, (n)*4, cliprects_enable); \
|
||||
} while (0)
|
||||
|
||||
#define OUT_BATCH(d) intel_batchbuffer_emit_dword(intel->batch, d)
|
||||
|
||||
#define OUT_RELOC(buf, flags, delta) do { \
|
||||
#define OUT_RELOC(buf, cliprects_enable, delta) do { \
|
||||
assert((delta) >= 0); \
|
||||
intel_batchbuffer_emit_reloc(intel->batch, buf, flags, delta); \
|
||||
intel_batchbuffer_emit_reloc(intel->batch, buf, cliprects_enable, delta); \
|
||||
} while (0)
|
||||
|
||||
#define ADVANCE_BATCH() do { } while(0)
|
||||
|
|
|
|||
|
|
@ -135,6 +135,9 @@ intel_readbuf_region(struct intel_context *intel)
|
|||
static void
|
||||
intelSetRenderbufferClipRects(struct intel_context *intel)
|
||||
{
|
||||
/* flush batch since pClipRects may change */
|
||||
intel_batchbuffer_flush(intel->batch);
|
||||
|
||||
assert(intel->ctx.DrawBuffer->Width > 0);
|
||||
assert(intel->ctx.DrawBuffer->Height > 0);
|
||||
intel->fboRect.x1 = 0;
|
||||
|
|
@ -160,6 +163,9 @@ intelSetFrontClipRects(struct intel_context *intel)
|
|||
if (!dPriv)
|
||||
return;
|
||||
|
||||
/* flush batch since pClipRects may change */
|
||||
intel_batchbuffer_flush(intel->batch);
|
||||
|
||||
intel->numClipRects = dPriv->numClipRects;
|
||||
intel->pClipRects = dPriv->pClipRects;
|
||||
intel->drawX = dPriv->x;
|
||||
|
|
@ -179,6 +185,9 @@ intelSetBackClipRects(struct intel_context *intel)
|
|||
if (!dPriv)
|
||||
return;
|
||||
|
||||
/* flush batch since pClipRects may change */
|
||||
intel_batchbuffer_flush(intel->batch);
|
||||
|
||||
intel_fb = dPriv->driverPrivate;
|
||||
|
||||
if (intel_fb->pf_active || dPriv->numBackClipRects == 0) {
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue