util: import a new vertex buffer manager

This code has originally matured in r300g and was ported to r600g several
times. It was obvious it's a code duplication.

See also comments in the header file.
This commit is contained in:
Marek Olšák 2011-01-29 20:53:57 +01:00
parent 1c2a4f0820
commit 975320ab76
4 changed files with 722 additions and 1 deletions

View file

@ -142,7 +142,8 @@ C_SOURCES = \
util/u_tile.c \
util/u_transfer.c \
util/u_resource.c \
util/u_upload_mgr.c
util/u_upload_mgr.c \
util/u_vbuf_mgr.c
# Disabling until pipe-video branch gets merged in
#vl/vl_bitstream_parser.c \

View file

@ -190,6 +190,7 @@ source = [
'util/u_tile.c',
'util/u_transfer.c',
'util/u_upload_mgr.c',
'util/u_vbuf_mgr.c',
# Disabling until pipe-video branch gets merged in
#'vl/vl_bitstream_parser.c',
#'vl/vl_mpeg12_mc_renderer.c',

View file

@ -0,0 +1,606 @@
/**************************************************************************
*
* Copyright 2011 Marek Olšák <maraeo@gmail.com>
* All Rights Reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the
* "Software"), to deal in the Software without restriction, including
* without limitation the rights to use, copy, modify, merge, publish,
* distribute, sub license, and/or sell copies of the Software, and to
* permit persons to whom the Software is furnished to do so, subject to
* the following conditions:
*
* The above copyright notice and this permission notice (including the
* next paragraph) shall be included in all copies or substantial portions
* of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
* OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
* IN NO EVENT SHALL AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*
**************************************************************************/
#include "util/u_vbuf_mgr.h"
#include "util/u_format.h"
#include "util/u_inlines.h"
#include "util/u_memory.h"
#include "util/u_upload_mgr.h"
#include "translate/translate.h"
#include "translate/translate_cache.h"
/* Hardware vertex fetcher limitations can be described by this structure. */
struct u_vbuf_caps {
/* Vertex format CAPs. */
/* TRUE if hardware supports it. */
unsigned format_fixed32:1; /* PIPE_FORMAT_*32*_FIXED */
unsigned format_float16:1; /* PIPE_FORMAT_*16*_FLOAT */
unsigned format_float64:1; /* PIPE_FORMAT_*64*_FLOAT */
unsigned format_norm32:1; /* PIPE_FORMAT_*32*NORM */
unsigned format_scaled32:1; /* PIPE_FORMAT_*32*SCALED */
/* Whether vertex fetches don't have to be dword-aligned. */
/* TRUE if hardware supports it. */
unsigned fetch_dword_unaligned:1;
};
struct u_vbuf_mgr_elements {
unsigned count;
struct pipe_vertex_element ve[PIPE_MAX_ATTRIBS];
/* If (velem[i].src_format != real_format[i]), the vertex buffer
* referenced by the vertex element cannot be used for rendering and
* its vertex data must be translated to real_format[i]. */
enum pipe_format native_format[PIPE_MAX_ATTRIBS];
unsigned native_format_size[PIPE_MAX_ATTRIBS];
/* This might mean two things:
* - src_format != native_format, as discussed above.
* - src_offset % 4 != 0 (if the caps don't allow such an offset). */
boolean incompatible_layout;
};
struct u_vbuf_mgr_priv {
struct u_vbuf_mgr b;
struct u_vbuf_caps caps;
struct pipe_context *pipe;
struct translate_cache *translate_cache;
unsigned translate_vb_slot;
struct u_upload_mgr *uploader;
struct u_vbuf_mgr_elements *ve;
void *saved_ve, *fallback_ve;
boolean ve_binding_lock;
boolean any_user_vbs;
boolean incompatible_vb_layout;
};
static void u_vbuf_mgr_init_format_caps(struct u_vbuf_mgr_priv *mgr)
{
struct pipe_screen *screen = mgr->pipe->screen;
mgr->caps.format_fixed32 =
screen->is_format_supported(screen, PIPE_FORMAT_R32_FIXED, PIPE_BUFFER,
0, PIPE_BIND_VERTEX_BUFFER, 0);
mgr->caps.format_float16 =
screen->is_format_supported(screen, PIPE_FORMAT_R16_FLOAT, PIPE_BUFFER,
0, PIPE_BIND_VERTEX_BUFFER, 0);
mgr->caps.format_float64 =
screen->is_format_supported(screen, PIPE_FORMAT_R64_FLOAT, PIPE_BUFFER,
0, PIPE_BIND_VERTEX_BUFFER, 0);
mgr->caps.format_norm32 =
screen->is_format_supported(screen, PIPE_FORMAT_R32_UNORM, PIPE_BUFFER,
0, PIPE_BIND_VERTEX_BUFFER, 0) &&
screen->is_format_supported(screen, PIPE_FORMAT_R32_SNORM, PIPE_BUFFER,
0, PIPE_BIND_VERTEX_BUFFER, 0);
mgr->caps.format_scaled32 =
screen->is_format_supported(screen, PIPE_FORMAT_R32_USCALED, PIPE_BUFFER,
0, PIPE_BIND_VERTEX_BUFFER, 0) &&
screen->is_format_supported(screen, PIPE_FORMAT_R32_SSCALED, PIPE_BUFFER,
0, PIPE_BIND_VERTEX_BUFFER, 0);
}
struct u_vbuf_mgr *
u_vbuf_mgr_create(struct pipe_context *pipe,
unsigned upload_buffer_size,
unsigned upload_buffer_alignment,
enum u_fetch_alignment fetch_alignment)
{
struct u_vbuf_mgr_priv *mgr = CALLOC_STRUCT(u_vbuf_mgr_priv);
mgr->pipe = pipe;
mgr->translate_cache = translate_cache_create();
mgr->uploader = u_upload_create(pipe, upload_buffer_size,
upload_buffer_alignment,
PIPE_BIND_VERTEX_BUFFER);
mgr->caps.fetch_dword_unaligned =
fetch_alignment == U_VERTEX_FETCH_BYTE_ALIGNED;
u_vbuf_mgr_init_format_caps(mgr);
return &mgr->b;
}
void u_vbuf_mgr_destroy(struct u_vbuf_mgr *mgrb)
{
struct u_vbuf_mgr_priv *mgr = (struct u_vbuf_mgr_priv*)mgrb;
unsigned i;
for (i = 0; i < mgr->b.nr_real_vertex_buffers; i++) {
pipe_resource_reference(&mgr->b.vertex_buffer[i].buffer, NULL);
pipe_resource_reference(&mgr->b.real_vertex_buffer[i], NULL);
}
translate_cache_destroy(mgr->translate_cache);
u_upload_destroy(mgr->uploader);
FREE(mgr);
}
static void u_vbuf_translate_begin(struct u_vbuf_mgr_priv *mgr,
int min_index, int max_index,
boolean *upload_flushed)
{
struct translate_key key = {0};
struct translate_element *te;
unsigned tr_elem_index[PIPE_MAX_ATTRIBS] = {0};
struct translate *tr;
boolean vb_translated[PIPE_MAX_ATTRIBS] = {0};
uint8_t *vb_map[PIPE_MAX_ATTRIBS] = {0}, *out_map;
struct pipe_transfer *vb_transfer[PIPE_MAX_ATTRIBS] = {0};
struct pipe_resource *out_buffer = NULL;
unsigned i, num_verts, out_offset;
struct pipe_vertex_element new_velems[PIPE_MAX_ATTRIBS];
/* Initialize the translate key, i.e. the recipe how vertices should be
* translated. */
for (i = 0; i < mgr->ve->count; i++) {
struct pipe_vertex_buffer *vb =
&mgr->b.vertex_buffer[mgr->ve->ve[i].vertex_buffer_index];
enum pipe_format output_format = mgr->ve->native_format[i];
unsigned output_format_size = mgr->ve->native_format_size[i];
/* Check for support. */
if (mgr->ve->ve[i].src_format == mgr->ve->native_format[i] &&
(mgr->caps.fetch_dword_unaligned ||
(vb->buffer_offset % 4 == 0 &&
vb->stride % 4 == 0 &&
mgr->ve->ve[i].src_offset % 4 == 0))) {
continue;
}
/* Workaround for translate: output floats instead of halfs. */
switch (output_format) {
case PIPE_FORMAT_R16_FLOAT:
output_format = PIPE_FORMAT_R32_FLOAT;
output_format_size = 4;
break;
case PIPE_FORMAT_R16G16_FLOAT:
output_format = PIPE_FORMAT_R32G32_FLOAT;
output_format_size = 8;
break;
case PIPE_FORMAT_R16G16B16_FLOAT:
output_format = PIPE_FORMAT_R32G32B32_FLOAT;
output_format_size = 12;
break;
case PIPE_FORMAT_R16G16B16A16_FLOAT:
output_format = PIPE_FORMAT_R32G32B32A32_FLOAT;
output_format_size = 16;
break;
default:;
}
/* Add this vertex element. */
te = &key.element[key.nr_elements];
/*te->type;
te->instance_divisor;*/
te->input_buffer = mgr->ve->ve[i].vertex_buffer_index;
te->input_format = mgr->ve->ve[i].src_format;
te->input_offset = mgr->ve->ve[i].src_offset;
te->output_format = output_format;
te->output_offset = key.output_stride;
key.output_stride += output_format_size;
vb_translated[mgr->ve->ve[i].vertex_buffer_index] = TRUE;
tr_elem_index[i] = key.nr_elements;
key.nr_elements++;
}
/* Get a translate object. */
tr = translate_cache_find(mgr->translate_cache, &key);
/* Map buffers we want to translate. */
for (i = 0; i < mgr->b.nr_vertex_buffers; i++) {
if (vb_translated[i]) {
struct pipe_vertex_buffer *vb = &mgr->b.vertex_buffer[i];
vb_map[i] = pipe_buffer_map(mgr->pipe, vb->buffer,
PIPE_TRANSFER_READ, &vb_transfer[i]);
tr->set_buffer(tr, i,
vb_map[i] + vb->buffer_offset + vb->stride * min_index,
vb->stride, ~0);
}
}
/* Create and map the output buffer. */
num_verts = max_index + 1 - min_index;
u_upload_alloc(mgr->uploader,
key.output_stride * min_index,
key.output_stride * num_verts,
&out_offset, &out_buffer, upload_flushed,
(void**)&out_map);
out_offset -= key.output_stride * min_index;
/* Translate. */
tr->run(tr, 0, num_verts, 0, out_map);
/* Unmap all buffers. */
for (i = 0; i < mgr->b.nr_vertex_buffers; i++) {
if (vb_translated[i]) {
pipe_buffer_unmap(mgr->pipe, vb_transfer[i]);
}
}
/* Setup the new vertex buffer in the first free slot. */
mgr->translate_vb_slot = ~0;
for (i = 0; i < PIPE_MAX_ATTRIBS; i++) {
if (!mgr->b.vertex_buffer[i].buffer) {
mgr->translate_vb_slot = i;
if (i >= mgr->b.nr_vertex_buffers) {
mgr->b.nr_real_vertex_buffers = i+1;
}
break;
}
}
if (mgr->translate_vb_slot != ~0) {
/* Setup the new vertex buffer. */
pipe_resource_reference(
&mgr->b.real_vertex_buffer[mgr->translate_vb_slot], out_buffer);
mgr->b.vertex_buffer[mgr->translate_vb_slot].buffer_offset = out_offset;
mgr->b.vertex_buffer[mgr->translate_vb_slot].stride = key.output_stride;
/* Setup new vertex elements. */
for (i = 0; i < mgr->ve->count; i++) {
if (vb_translated[mgr->ve->ve[i].vertex_buffer_index]) {
te = &key.element[tr_elem_index[i]];
new_velems[i].instance_divisor = mgr->ve->ve[i].instance_divisor;
new_velems[i].src_format = te->output_format;
new_velems[i].src_offset = te->output_offset;
new_velems[i].vertex_buffer_index = mgr->translate_vb_slot;
} else {
memcpy(&new_velems[i], &mgr->ve->ve[i],
sizeof(struct pipe_vertex_element));
}
}
mgr->fallback_ve =
mgr->pipe->create_vertex_elements_state(mgr->pipe, mgr->ve->count,
new_velems);
/* Preserve saved_ve. */
mgr->ve_binding_lock = TRUE;
mgr->pipe->bind_vertex_elements_state(mgr->pipe, mgr->fallback_ve);
mgr->ve_binding_lock = FALSE;
}
pipe_resource_reference(&out_buffer, NULL);
}
static void u_vbuf_translate_end(struct u_vbuf_mgr_priv *mgr)
{
if (mgr->fallback_ve == NULL) {
return;
}
/* Restore vertex elements. */
/* Note that saved_ve will be overwritten in bind_vertex_elements_state. */
mgr->pipe->bind_vertex_elements_state(mgr->pipe, mgr->saved_ve);
mgr->pipe->delete_vertex_elements_state(mgr->pipe, mgr->fallback_ve);
mgr->fallback_ve = NULL;
/* Delete the now-unused VBO. */
pipe_resource_reference(&mgr->b.real_vertex_buffer[mgr->translate_vb_slot],
NULL);
mgr->b.nr_real_vertex_buffers = mgr->b.nr_vertex_buffers;
}
#define FORMAT_REPLACE(what, withwhat) \
case PIPE_FORMAT_##what: format = PIPE_FORMAT_##withwhat; break
struct u_vbuf_mgr_elements *
u_vbuf_mgr_create_vertex_elements(struct u_vbuf_mgr *mgrb,
unsigned count,
const struct pipe_vertex_element *attribs,
struct pipe_vertex_element *native_attribs)
{
struct u_vbuf_mgr_priv *mgr = (struct u_vbuf_mgr_priv*)mgrb;
unsigned i;
struct u_vbuf_mgr_elements *ve = CALLOC_STRUCT(u_vbuf_mgr_elements);
ve->count = count;
if (!count) {
return ve;
}
memcpy(ve->ve, attribs, sizeof(struct pipe_vertex_element) * count);
memcpy(native_attribs, attribs, sizeof(struct pipe_vertex_element) * count);
/* Set the best native format in case the original format is not
* supported. */
for (i = 0; i < count; i++) {
enum pipe_format format = ve->ve[i].src_format;
/* Choose a native format.
* For now we don't care about the alignment, that's going to
* be sorted out later. */
if (!mgr->caps.format_fixed32) {
switch (format) {
FORMAT_REPLACE(R32_FIXED, R32_FLOAT);
FORMAT_REPLACE(R32G32_FIXED, R32G32_FLOAT);
FORMAT_REPLACE(R32G32B32_FIXED, R32G32B32_FLOAT);
FORMAT_REPLACE(R32G32B32A32_FIXED, R32G32B32A32_FLOAT);
default:;
}
}
if (!mgr->caps.format_float16) {
switch (format) {
FORMAT_REPLACE(R16_FLOAT, R32_FLOAT);
FORMAT_REPLACE(R16G16_FLOAT, R32G32_FLOAT);
FORMAT_REPLACE(R16G16B16_FLOAT, R32G32B32_FLOAT);
FORMAT_REPLACE(R16G16B16A16_FLOAT, R32G32B32A32_FLOAT);
default:;
}
}
if (!mgr->caps.format_float64) {
switch (format) {
FORMAT_REPLACE(R64_FLOAT, R32_FLOAT);
FORMAT_REPLACE(R64G64_FLOAT, R32G32_FLOAT);
FORMAT_REPLACE(R64G64B64_FLOAT, R32G32B32_FLOAT);
FORMAT_REPLACE(R64G64B64A64_FLOAT, R32G32B32A32_FLOAT);
default:;
}
}
if (!mgr->caps.format_norm32) {
switch (format) {
FORMAT_REPLACE(R32_UNORM, R32_FLOAT);
FORMAT_REPLACE(R32G32_UNORM, R32G32_FLOAT);
FORMAT_REPLACE(R32G32B32_UNORM, R32G32B32_FLOAT);
FORMAT_REPLACE(R32G32B32A32_UNORM, R32G32B32A32_FLOAT);
FORMAT_REPLACE(R32_SNORM, R32_FLOAT);
FORMAT_REPLACE(R32G32_SNORM, R32G32_FLOAT);
FORMAT_REPLACE(R32G32B32_SNORM, R32G32B32_FLOAT);
FORMAT_REPLACE(R32G32B32A32_SNORM, R32G32B32A32_FLOAT);
default:;
}
}
if (!mgr->caps.format_scaled32) {
switch (format) {
FORMAT_REPLACE(R32_USCALED, R32_FLOAT);
FORMAT_REPLACE(R32G32_USCALED, R32G32_FLOAT);
FORMAT_REPLACE(R32G32B32_USCALED, R32G32B32_FLOAT);
FORMAT_REPLACE(R32G32B32A32_USCALED,R32G32B32A32_FLOAT);
FORMAT_REPLACE(R32_SSCALED, R32_FLOAT);
FORMAT_REPLACE(R32G32_SSCALED, R32G32_FLOAT);
FORMAT_REPLACE(R32G32B32_SSCALED, R32G32B32_FLOAT);
FORMAT_REPLACE(R32G32B32A32_SSCALED,R32G32B32A32_FLOAT);
default:;
}
}
native_attribs[i].src_format = format;
ve->native_format[i] = format;
ve->native_format_size[i] =
util_format_get_blocksize(ve->native_format[i]);
ve->incompatible_layout =
ve->incompatible_layout ||
ve->ve[i].src_format != ve->native_format[i] ||
(!mgr->caps.fetch_dword_unaligned && ve->ve[i].src_offset % 4 != 0);
}
/* Align the formats to the size of DWORD if needed. */
if (!mgr->caps.fetch_dword_unaligned) {
for (i = 0; i < count; i++) {
ve->native_format_size[i] = align(ve->native_format_size[i], 4);
}
}
return ve;
}
void u_vbuf_mgr_bind_vertex_elements(struct u_vbuf_mgr *mgrb,
void *cso,
struct u_vbuf_mgr_elements *ve)
{
struct u_vbuf_mgr_priv *mgr = (struct u_vbuf_mgr_priv*)mgrb;
if (!cso) {
return;
}
if (!mgr->ve_binding_lock) {
mgr->saved_ve = cso;
mgr->ve = ve;
}
}
void u_vbuf_mgr_destroy_vertex_elements(struct u_vbuf_mgr *mgr,
struct u_vbuf_mgr_elements *ve)
{
FREE(ve);
}
void u_vbuf_mgr_set_vertex_buffers(struct u_vbuf_mgr *mgrb,
unsigned count,
const struct pipe_vertex_buffer *bufs)
{
struct u_vbuf_mgr_priv *mgr = (struct u_vbuf_mgr_priv*)mgrb;
unsigned i;
mgr->b.max_index = ~0;
mgr->any_user_vbs = FALSE;
mgr->incompatible_vb_layout = FALSE;
if (!mgr->caps.fetch_dword_unaligned) {
/* Check if the strides and offsets are aligned to the size of DWORD. */
for (i = 0; i < count; i++) {
if (bufs[i].buffer) {
if (bufs[i].stride % 4 != 0 ||
bufs[i].buffer_offset % 4 != 0) {
mgr->incompatible_vb_layout = TRUE;
break;
}
}
}
}
for (i = 0; i < count; i++) {
const struct pipe_vertex_buffer *vb = &bufs[i];
pipe_resource_reference(&mgr->b.vertex_buffer[i].buffer, vb->buffer);
pipe_resource_reference(&mgr->b.real_vertex_buffer[i], NULL);
if (u_vbuf_resource(vb->buffer)->user_ptr) {
mgr->any_user_vbs = TRUE;
continue;
}
pipe_resource_reference(&mgr->b.real_vertex_buffer[i], vb->buffer);
/* The stride of zero means we will be fetching only the first
* vertex, so don't care about max_index. */
if (!vb->stride) {
continue;
}
/* Update the maximum index. */
mgr->b.max_index =
MIN2(mgr->b.max_index,
(vb->buffer->width0 - vb->buffer_offset) / vb->stride);
}
for (; i < mgr->b.nr_real_vertex_buffers; i++) {
pipe_resource_reference(&mgr->b.vertex_buffer[i].buffer, NULL);
pipe_resource_reference(&mgr->b.real_vertex_buffer[i], NULL);
}
memcpy(mgr->b.vertex_buffer, bufs,
sizeof(struct pipe_vertex_buffer) * count);
mgr->b.nr_vertex_buffers = count;
mgr->b.nr_real_vertex_buffers = count;
}
static void u_vbuf_upload_buffers(struct u_vbuf_mgr_priv *mgr,
int min_index, int max_index,
boolean *upload_flushed)
{
int i, nr = mgr->ve->count;
unsigned count = max_index + 1 - min_index;
boolean uploaded[PIPE_MAX_ATTRIBS] = {0};
for (i = 0; i < nr; i++) {
unsigned index = mgr->ve->ve[i].vertex_buffer_index;
struct pipe_vertex_buffer *vb = &mgr->b.vertex_buffer[index];
if (vb->buffer &&
u_vbuf_resource(vb->buffer)->user_ptr &&
!uploaded[index]) {
unsigned first, size;
boolean flushed;
if (vb->stride) {
first = vb->stride * min_index;
size = vb->stride * count;
} else {
first = 0;
size = mgr->ve->native_format_size[i];
}
u_upload_data(mgr->uploader, first, size,
u_vbuf_resource(vb->buffer)->user_ptr + first,
&vb->buffer_offset,
&mgr->b.real_vertex_buffer[index],
&flushed);
vb->buffer_offset -= first;
uploaded[index] = TRUE;
*upload_flushed = *upload_flushed || flushed;
} else {
assert(mgr->b.real_vertex_buffer[index]);
}
}
}
void u_vbuf_mgr_draw_begin(struct u_vbuf_mgr *mgrb,
const struct pipe_draw_info *info,
boolean *buffers_updated,
boolean *uploader_flushed)
{
struct u_vbuf_mgr_priv *mgr = (struct u_vbuf_mgr_priv*)mgrb;
boolean bufs_updated, upload_flushed = FALSE;
int min_index, max_index;
min_index = info->min_index - info->index_bias;
max_index = MIN2(info->max_index, mgr->b.max_index) - info->index_bias;
/* Translate vertices with non-native layouts or formats. */
if (mgr->incompatible_vb_layout || mgr->ve->incompatible_layout) {
u_vbuf_translate_begin(mgr, min_index, max_index, &upload_flushed);
if (mgr->fallback_ve) {
bufs_updated = TRUE;
}
}
/* Upload user buffers. */
if (mgr->any_user_vbs) {
u_vbuf_upload_buffers(mgr, min_index, max_index, &upload_flushed);
bufs_updated = TRUE;
}
/* Set the return values. */
if (buffers_updated) {
*buffers_updated = bufs_updated;
}
if (uploader_flushed) {
*uploader_flushed = upload_flushed;
}
}
void u_vbuf_mgr_draw_end(struct u_vbuf_mgr *mgrb)
{
struct u_vbuf_mgr_priv *mgr = (struct u_vbuf_mgr_priv*)mgrb;
if (mgr->fallback_ve) {
u_vbuf_translate_end(mgr);
}
}
void u_vbuf_mgr_flush_uploader(struct u_vbuf_mgr *mgrb)
{
struct u_vbuf_mgr_priv *mgr = (struct u_vbuf_mgr_priv*)mgrb;
u_upload_flush(mgr->uploader);
}

View file

@ -0,0 +1,113 @@
/**************************************************************************
*
* Copyright 2011 Marek Olšák <maraeo@gmail.com>
* All Rights Reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the
* "Software"), to deal in the Software without restriction, including
* without limitation the rights to use, copy, modify, merge, publish,
* distribute, sub license, and/or sell copies of the Software, and to
* permit persons to whom the Software is furnished to do so, subject to
* the following conditions:
*
* The above copyright notice and this permission notice (including the
* next paragraph) shall be included in all copies or substantial portions
* of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
* OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
* IN NO EVENT SHALL AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*
**************************************************************************/
#ifndef U_VBUF_MGR_H
#define U_VBUF_MGR_H
/* This module builds upon u_upload_mgr and translate_cache and takes care of
* user buffer uploads and vertex format fallbacks. It's designed
* for the drivers which don't always use the Draw module. (e.g. for HWTCL)
*/
#include "pipe/p_context.h"
#include "pipe/p_state.h"
#include "util/u_transfer.h"
/* The manager.
* This structure should also be used to access vertex buffers
* from a driver. */
struct u_vbuf_mgr {
/* This is what was set in set_vertex_buffers.
* May contain user buffers. */
struct pipe_vertex_buffer vertex_buffer[PIPE_MAX_ATTRIBS];
unsigned nr_vertex_buffers;
/* Contains only real vertex buffers.
* Hardware drivers should use real_vertex_buffers[i]
* instead of vertex_buffers[i].buffer. */
struct pipe_resource *real_vertex_buffer[PIPE_MAX_ATTRIBS];
int nr_real_vertex_buffers;
/* Precomputed max_index for hardware vertex buffers. */
int max_index;
};
struct u_vbuf_resource {
struct u_resource b;
uint8_t *user_ptr;
};
/* Opaque type containing information about vertex elements for the manager. */
struct u_vbuf_mgr_elements;
enum u_fetch_alignment {
U_VERTEX_FETCH_BYTE_ALIGNED,
U_VERTEX_FETCH_DWORD_ALIGNED
};
struct u_vbuf_mgr *
u_vbuf_mgr_create(struct pipe_context *pipe,
unsigned upload_buffer_size,
unsigned upload_buffer_alignment,
enum u_fetch_alignment fetch_alignment);
void u_vbuf_mgr_destroy(struct u_vbuf_mgr *mgr);
struct u_vbuf_mgr_elements *
u_vbuf_mgr_create_vertex_elements(struct u_vbuf_mgr *mgr,
unsigned count,
const struct pipe_vertex_element *attrs,
struct pipe_vertex_element *native_attrs);
void u_vbuf_mgr_bind_vertex_elements(struct u_vbuf_mgr *mgr,
void *cso,
struct u_vbuf_mgr_elements *ve);
void u_vbuf_mgr_destroy_vertex_elements(struct u_vbuf_mgr *mgr,
struct u_vbuf_mgr_elements *ve);
void u_vbuf_mgr_set_vertex_buffers(struct u_vbuf_mgr *mgr,
unsigned count,
const struct pipe_vertex_buffer *bufs);
void u_vbuf_mgr_draw_begin(struct u_vbuf_mgr *mgr,
const struct pipe_draw_info *info,
boolean *buffers_updated,
boolean *uploader_flushed);
void u_vbuf_mgr_draw_end(struct u_vbuf_mgr *mgr);
void u_vbuf_mgr_flush_uploader(struct u_vbuf_mgr *mgr);
static INLINE struct u_vbuf_resource *u_vbuf_resource(struct pipe_resource *r)
{
return (struct u_vbuf_resource*)r;
}
#endif