panfrost: Prepare things for the common/JM cmdstream split

In order to be able to split the code into separate files, we need to
share some common definitions. Create a pan_cmdstream.h file containing
some structs and helpers.

This file is meant to be included from per-gen source files.

Signed-off-by: Boris Brezillon <boris.brezillon@collabora.com>
Reviewed-by: Erik Faye-Lund <erik.faye-lund@collabora.com>
Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/26249>
This commit is contained in:
Boris Brezillon 2023-10-05 18:56:53 +02:00 committed by Marge Bot
parent 635385a807
commit 5e5a766a81
2 changed files with 321 additions and 274 deletions

View file

@ -46,6 +46,7 @@
#include "pan_blend.h"
#include "pan_blitter.h"
#include "pan_bo.h"
#include "pan_cmdstream.h"
#include "pan_context.h"
#include "pan_indirect_dispatch.h"
#include "pan_job.h"
@ -54,43 +55,6 @@
#include "pan_texture.h"
#include "pan_util.h"
#define PAN_GPU_INDIRECTS (PAN_ARCH == 7)
struct panfrost_rasterizer {
struct pipe_rasterizer_state base;
#if PAN_ARCH <= 7
/* Partially packed RSD words */
struct mali_multisample_misc_packed multisample;
struct mali_stencil_mask_misc_packed stencil_misc;
#endif
};
struct panfrost_zsa_state {
struct pipe_depth_stencil_alpha_state base;
/* Is any depth, stencil, or alpha testing enabled? */
bool enabled;
/* Does the depth and stencil tests always pass? This ignores write
* masks, we are only interested in whether pixels may be killed.
*/
bool zs_always_passes;
/* Are depth or stencil writes possible? */
bool writes_zs;
#if PAN_ARCH <= 7
/* Prepacked words from the RSD */
struct mali_multisample_misc_packed rsd_depth;
struct mali_stencil_mask_misc_packed rsd_stencil;
struct mali_stencil_packed stencil_front, stencil_back;
#else
/* Depth/stencil descriptor template */
struct mali_depth_stencil_packed desc;
#endif
};
struct panfrost_sampler_state {
struct pipe_sampler_state base;
struct mali_sampler_packed hw;
@ -112,25 +76,6 @@ struct panfrost_sampler_view {
struct panfrost_pool *pool;
};
struct panfrost_vertex_state {
unsigned num_elements;
struct pipe_vertex_element pipe[PIPE_MAX_ATTRIBS];
uint16_t strides[PIPE_MAX_ATTRIBS];
#if PAN_ARCH >= 9
/* Packed attribute descriptors */
struct mali_attribute_packed attributes[PIPE_MAX_ATTRIBS];
#else
/* buffers corresponds to attribute buffer, element_buffers corresponds
* to an index in buffers for each vertex element */
struct pan_vertex_buffer buffers[PIPE_MAX_ATTRIBS];
unsigned element_buffer[PIPE_MAX_ATTRIBS];
unsigned nr_bufs;
unsigned formats[PIPE_MAX_ATTRIBS];
#endif
};
/* Statically assert that PIPE_* enums match the hardware enums.
* (As long as they match, we don't need to translate them.)
*/
@ -291,35 +236,6 @@ panfrost_create_sampler_state(struct pipe_context *pctx,
return so;
}
static bool
panfrost_fs_required(struct panfrost_compiled_shader *fs,
struct panfrost_blend_state *blend,
struct pipe_framebuffer_state *state,
const struct panfrost_zsa_state *zsa)
{
/* If we generally have side effects. This inclues use of discard,
* which can affect the results of an occlusion query. */
if (fs->info.fs.sidefx)
return true;
/* Using an empty FS requires early-z to be enabled, but alpha test
* needs it disabled. Alpha test is only native on Midgard, so only
* check there.
*/
if (PAN_ARCH <= 5 && zsa->base.alpha_func != PIPE_FUNC_ALWAYS)
return true;
/* If colour is written we need to execute */
for (unsigned i = 0; i < state->nr_cbufs; ++i) {
if (state->cbufs[i] && blend->info[i].enabled)
return true;
}
/* If depth is written and not implied we need to execute.
* TODO: Predicate on Z/S writes being enabled */
return (fs->info.fs.writes_depth || fs->info.fs.writes_stencil);
}
/* Get pointers to the blend shaders bound to each active render target. Used
* to emit the blend descriptors, as well as the fragment renderer state
* descriptor.
@ -358,34 +274,6 @@ pack_blend_constant(enum pipe_format format, float cons)
return unorm << (16 - chan_size);
}
/*
* Determine whether to set the respective overdraw alpha flag.
*
* The overdraw alpha=1 flag should be set when alpha=1 implies full overdraw,
* equivalently, all enabled render targets have alpha_one_store set. Likewise,
* overdraw alpha=0 should be set when alpha=0 implies no overdraw,
* equivalently, all enabled render targets have alpha_zero_nop set.
*/
#if PAN_ARCH >= 6
static bool
panfrost_overdraw_alpha(const struct panfrost_context *ctx, bool zero)
{
const struct panfrost_blend_state *so = ctx->blend;
for (unsigned i = 0; i < ctx->pipe_framebuffer.nr_cbufs; ++i) {
const struct pan_blend_info info = so->info[i];
bool enabled = ctx->pipe_framebuffer.cbufs[i] && !info.enabled;
bool flag = zero ? info.alpha_zero_nop : info.alpha_one_store;
if (enabled && !flag)
return false;
}
return true;
}
#endif
static void
panfrost_emit_blend(struct panfrost_batch *batch, void *rts,
mali_ptr *blend_shaders)
@ -499,23 +387,6 @@ panfrost_emit_blend(struct panfrost_batch *batch, void *rts,
}
#endif
static inline bool
pan_allow_forward_pixel_to_kill(struct panfrost_context *ctx,
struct panfrost_compiled_shader *fs)
{
/* Track if any colour buffer is reused across draws, either
* from reading it directly, or from failing to write it
*/
unsigned rt_mask = ctx->fb_rt_mask;
uint64_t rt_written = (fs->info.outputs_written >> FRAG_RESULT_DATA0) &
ctx->blend->enabled_mask;
bool blend_reads_dest = (ctx->blend->load_dest_mask & rt_mask);
bool alpha_to_coverage = ctx->blend->base.alpha_to_coverage;
return fs->info.fs.can_fpk && !(rt_mask & ~rt_written) &&
!alpha_to_coverage && !blend_reads_dest;
}
static mali_ptr
panfrost_emit_compute_shader_meta(struct panfrost_batch *batch,
enum pipe_shader_type stage)
@ -2688,34 +2559,6 @@ emit_fragment_job(struct panfrost_batch *batch, const struct pan_fb_info *pfb)
jm_emit_fragment_job(batch, pfb);
}
#define DEFINE_CASE(c) \
case MESA_PRIM_##c: \
return MALI_DRAW_MODE_##c;
static uint8_t
pan_draw_mode(enum mesa_prim mode)
{
switch (mode) {
DEFINE_CASE(POINTS);
DEFINE_CASE(LINES);
DEFINE_CASE(LINE_LOOP);
DEFINE_CASE(LINE_STRIP);
DEFINE_CASE(TRIANGLES);
DEFINE_CASE(TRIANGLE_STRIP);
DEFINE_CASE(TRIANGLE_FAN);
DEFINE_CASE(QUADS);
DEFINE_CASE(POLYGON);
#if PAN_ARCH <= 6
DEFINE_CASE(QUAD_STRIP);
#endif
default:
unreachable("Invalid draw mode");
}
}
#undef DEFINE_CASE
/* Count generated primitives (when there is no geom/tess shaders) for
* transform feedback */
@ -2751,16 +2594,6 @@ panfrost_update_streamout_offsets(struct panfrost_context *ctx)
}
}
static inline enum mali_index_type
panfrost_translate_index_size(unsigned size)
{
STATIC_ASSERT(MALI_INDEX_TYPE_NONE == 0);
STATIC_ASSERT(MALI_INDEX_TYPE_UINT8 == 1);
STATIC_ASSERT(MALI_INDEX_TYPE_UINT16 == 2);
return (size == 4) ? MALI_INDEX_TYPE_UINT32 : size;
}
#if PAN_ARCH <= 7
static inline void
jm_emit_draw_descs(struct panfrost_batch *batch, struct MALI_DRAW *d,
@ -2807,30 +2640,6 @@ jm_emit_vertex_job(struct panfrost_batch *batch,
}
#endif
static void
panfrost_emit_primitive_size(struct panfrost_context *ctx, bool points,
mali_ptr size_array, void *prim_size)
{
struct panfrost_rasterizer *rast = ctx->rasterizer;
pan_pack(prim_size, PRIMITIVE_SIZE, cfg) {
if (panfrost_writes_point_size(ctx)) {
cfg.size_array = size_array;
} else {
cfg.constant = points ? rast->base.point_size : rast->base.line_width;
}
}
}
static bool
panfrost_is_implicit_prim_restart(const struct pipe_draw_info *info)
{
/* As a reminder primitive_restart should always be checked before any
access to restart_index. */
return info->primitive_restart &&
info->restart_index == (unsigned)BITFIELD_MASK(info->index_size * 8);
}
/* On Bifrost and older, the Renderer State Descriptor aggregates many pieces of
* 3D state. In particular, it groups the fragment shader descriptor with
* depth/stencil, blend, polygon offset, and multisampling state. These pieces
@ -3002,14 +2811,6 @@ jm_emit_tiler_desc(struct panfrost_batch *batch)
}
#endif
static bool
allow_rotating_primitives(const struct panfrost_compiled_shader *fs,
const struct pipe_draw_info *info)
{
return u_reduced_prim(info->mode) != MESA_PRIM_LINES &&
!fs->info.bifrost.uses_flat_shading;
}
/* Packs a primitive descriptor, mostly common between Midgard/Bifrost tiler
* jobs and Valhall IDVS jobs
*/
@ -3086,47 +2887,6 @@ jm_emit_primitive(struct panfrost_batch *batch,
}
#if PAN_ARCH >= 9
static mali_ptr
panfrost_emit_resources(struct panfrost_batch *batch,
enum pipe_shader_type stage)
{
struct panfrost_context *ctx = batch->ctx;
struct panfrost_ptr T;
unsigned nr_tables = 12;
/* Although individual resources need only 16 byte alignment, the
* resource table as a whole must be 64-byte aligned.
*/
T = pan_pool_alloc_aligned(&batch->pool.base, nr_tables * pan_size(RESOURCE),
64);
memset(T.cpu, 0, nr_tables * pan_size(RESOURCE));
panfrost_make_resource_table(T, PAN_TABLE_UBO, batch->uniform_buffers[stage],
batch->nr_uniform_buffers[stage]);
panfrost_make_resource_table(T, PAN_TABLE_TEXTURE, batch->textures[stage],
ctx->sampler_view_count[stage]);
/* We always need at least 1 sampler for txf to work */
panfrost_make_resource_table(T, PAN_TABLE_SAMPLER, batch->samplers[stage],
MAX2(ctx->sampler_count[stage], 1));
panfrost_make_resource_table(T, PAN_TABLE_IMAGE, batch->images[stage],
util_last_bit(ctx->image_mask[stage]));
if (stage == PIPE_SHADER_VERTEX) {
panfrost_make_resource_table(T, PAN_TABLE_ATTRIBUTE,
batch->attribs[stage],
ctx->vertex->num_elements);
panfrost_make_resource_table(T, PAN_TABLE_ATTRIBUTE_BUFFER,
batch->attrib_bufs[stage],
util_last_bit(ctx->vb_mask));
}
return T.gpu | nr_tables;
}
static void
jm_emit_shader_env(struct panfrost_batch *batch,
struct MALI_SHADER_ENVIRONMENT *cfg,
@ -3297,39 +3057,6 @@ jm_emit_tiler_draw(void *out, struct panfrost_batch *batch, bool fs_required,
}
#if PAN_ARCH >= 9
static mali_ptr
panfrost_get_position_shader(struct panfrost_batch *batch,
const struct pipe_draw_info *info)
{
/* IDVS/points vertex shader */
mali_ptr vs_ptr = batch->rsd[PIPE_SHADER_VERTEX];
/* IDVS/triangle vertex shader */
if (vs_ptr && info->mode != MESA_PRIM_POINTS)
vs_ptr += pan_size(SHADER_PROGRAM);
return vs_ptr;
}
static mali_ptr
panfrost_get_varying_shader(struct panfrost_batch *batch)
{
return batch->rsd[PIPE_SHADER_VERTEX] + (2 * pan_size(SHADER_PROGRAM));
}
static unsigned
panfrost_vertex_attribute_stride(struct panfrost_compiled_shader *vs,
struct panfrost_compiled_shader *fs)
{
unsigned v = vs->info.varyings.output_count;
unsigned f = fs->info.varyings.input_count;
unsigned slots = MAX2(v, f);
slots += util_bitcount(fs->key.fs.fixed_varying_mask);
/* Assumes 16 byte slots. We could do better. */
return slots * 16;
}
static void
jm_emit_malloc_vertex_job(struct panfrost_batch *batch,
const struct pipe_draw_info *info,

View file

@ -0,0 +1,320 @@
/*
* Copyright (C) 2018 Alyssa Rosenzweig
* Copyright (C) 2020 Collabora Ltd.
* Copyright © 2017 Intel Corporation
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
#ifndef __PAN_CMDSTREAM_H__
#define __PAN_CMDSTREAM_H__
#ifndef PAN_ARCH
#error "PAN_ARCH undefined!"
#endif
#include "genxml/gen_macros.h"
#include "pan_context.h"
#include "pan_job.h"
#include "pipe/p_defines.h"
#include "pipe/p_state.h"
#include "util/u_prim.h"
#define PAN_GPU_INDIRECTS (PAN_ARCH == 7)
struct panfrost_rasterizer {
struct pipe_rasterizer_state base;
#if PAN_ARCH <= 7
/* Partially packed RSD words */
struct mali_multisample_misc_packed multisample;
struct mali_stencil_mask_misc_packed stencil_misc;
#endif
};
struct panfrost_zsa_state {
struct pipe_depth_stencil_alpha_state base;
/* Is any depth, stencil, or alpha testing enabled? */
bool enabled;
/* Does the depth and stencil tests always pass? This ignores write
* masks, we are only interested in whether pixels may be killed.
*/
bool zs_always_passes;
/* Are depth or stencil writes possible? */
bool writes_zs;
#if PAN_ARCH <= 7
/* Prepacked words from the RSD */
struct mali_multisample_misc_packed rsd_depth;
struct mali_stencil_mask_misc_packed rsd_stencil;
struct mali_stencil_packed stencil_front, stencil_back;
#else
/* Depth/stencil descriptor template */
struct mali_depth_stencil_packed desc;
#endif
};
struct panfrost_vertex_state {
unsigned num_elements;
struct pipe_vertex_element pipe[PIPE_MAX_ATTRIBS];
uint16_t strides[PIPE_MAX_ATTRIBS];
#if PAN_ARCH >= 9
/* Packed attribute descriptors */
struct mali_attribute_packed attributes[PIPE_MAX_ATTRIBS];
#else
/* buffers corresponds to attribute buffer, element_buffers corresponds
* to an index in buffers for each vertex element */
struct pan_vertex_buffer buffers[PIPE_MAX_ATTRIBS];
unsigned element_buffer[PIPE_MAX_ATTRIBS];
unsigned nr_bufs;
unsigned formats[PIPE_MAX_ATTRIBS];
#endif
};
static inline bool
panfrost_is_implicit_prim_restart(const struct pipe_draw_info *info)
{
/* As a reminder primitive_restart should always be checked before any
access to restart_index. */
return info->primitive_restart &&
info->restart_index == (unsigned)BITFIELD_MASK(info->index_size * 8);
}
static inline bool
pan_allow_forward_pixel_to_kill(struct panfrost_context *ctx,
struct panfrost_compiled_shader *fs)
{
/* Track if any colour buffer is reused across draws, either
* from reading it directly, or from failing to write it
*/
unsigned rt_mask = ctx->fb_rt_mask;
uint64_t rt_written = (fs->info.outputs_written >> FRAG_RESULT_DATA0) &
ctx->blend->enabled_mask;
bool blend_reads_dest = (ctx->blend->load_dest_mask & rt_mask);
bool alpha_to_coverage = ctx->blend->base.alpha_to_coverage;
return fs->info.fs.can_fpk && !(rt_mask & ~rt_written) &&
!alpha_to_coverage && !blend_reads_dest;
}
/*
* Determine whether to set the respective overdraw alpha flag.
*
* The overdraw alpha=1 flag should be set when alpha=1 implies full overdraw,
* equivalently, all enabled render targets have alpha_one_store set. Likewise,
* overdraw alpha=0 should be set when alpha=0 implies no overdraw,
* equivalently, all enabled render targets have alpha_zero_nop set.
*/
#if PAN_ARCH >= 6
static inline bool
panfrost_overdraw_alpha(const struct panfrost_context *ctx, bool zero)
{
const struct panfrost_blend_state *so = ctx->blend;
for (unsigned i = 0; i < ctx->pipe_framebuffer.nr_cbufs; ++i) {
const struct pan_blend_info info = so->info[i];
bool enabled = ctx->pipe_framebuffer.cbufs[i] && !info.enabled;
bool flag = zero ? info.alpha_zero_nop : info.alpha_one_store;
if (enabled && !flag)
return false;
}
return true;
}
#endif
static inline void
panfrost_emit_primitive_size(struct panfrost_context *ctx, bool points,
mali_ptr size_array, void *prim_size)
{
struct panfrost_rasterizer *rast = ctx->rasterizer;
pan_pack(prim_size, PRIMITIVE_SIZE, cfg) {
if (panfrost_writes_point_size(ctx)) {
cfg.size_array = size_array;
} else {
cfg.constant = points ? rast->base.point_size : rast->base.line_width;
}
}
}
static inline uint8_t
pan_draw_mode(enum mesa_prim mode)
{
switch (mode) {
#define DEFINE_CASE(c) \
case MESA_PRIM_##c: \
return MALI_DRAW_MODE_##c;
DEFINE_CASE(POINTS);
DEFINE_CASE(LINES);
DEFINE_CASE(LINE_LOOP);
DEFINE_CASE(LINE_STRIP);
DEFINE_CASE(TRIANGLES);
DEFINE_CASE(TRIANGLE_STRIP);
DEFINE_CASE(TRIANGLE_FAN);
DEFINE_CASE(QUADS);
DEFINE_CASE(POLYGON);
#if PAN_ARCH <= 6
DEFINE_CASE(QUAD_STRIP);
#endif
#undef DEFINE_CASE
default:
unreachable("Invalid draw mode");
}
}
static inline enum mali_index_type
panfrost_translate_index_size(unsigned size)
{
STATIC_ASSERT(MALI_INDEX_TYPE_NONE == 0);
STATIC_ASSERT(MALI_INDEX_TYPE_UINT8 == 1);
STATIC_ASSERT(MALI_INDEX_TYPE_UINT16 == 2);
return (size == 4) ? MALI_INDEX_TYPE_UINT32 : size;
}
static inline bool
panfrost_fs_required(struct panfrost_compiled_shader *fs,
struct panfrost_blend_state *blend,
struct pipe_framebuffer_state *state,
const struct panfrost_zsa_state *zsa)
{
/* If we generally have side effects. This inclues use of discard,
* which can affect the results of an occlusion query. */
if (fs->info.fs.sidefx)
return true;
/* Using an empty FS requires early-z to be enabled, but alpha test
* needs it disabled. Alpha test is only native on Midgard, so only
* check there.
*/
if (PAN_ARCH <= 5 && zsa->base.alpha_func != PIPE_FUNC_ALWAYS)
return true;
/* If colour is written we need to execute */
for (unsigned i = 0; i < state->nr_cbufs; ++i) {
if (state->cbufs[i] && blend->info[i].enabled)
return true;
}
/* If depth is written and not implied we need to execute.
* TODO: Predicate on Z/S writes being enabled */
return (fs->info.fs.writes_depth || fs->info.fs.writes_stencil);
}
#if PAN_ARCH >= 9
static inline mali_ptr
panfrost_get_position_shader(struct panfrost_batch *batch,
const struct pipe_draw_info *info)
{
/* IDVS/points vertex shader */
mali_ptr vs_ptr = batch->rsd[PIPE_SHADER_VERTEX];
/* IDVS/triangle vertex shader */
if (vs_ptr && info->mode != MESA_PRIM_POINTS)
vs_ptr += pan_size(SHADER_PROGRAM);
return vs_ptr;
}
static inline mali_ptr
panfrost_get_varying_shader(struct panfrost_batch *batch)
{
return batch->rsd[PIPE_SHADER_VERTEX] + (2 * pan_size(SHADER_PROGRAM));
}
static inline unsigned
panfrost_vertex_attribute_stride(struct panfrost_compiled_shader *vs,
struct panfrost_compiled_shader *fs)
{
unsigned v = vs->info.varyings.output_count;
unsigned f = fs->info.varyings.input_count;
unsigned slots = MAX2(v, f);
slots += util_bitcount(fs->key.fs.fixed_varying_mask);
/* Assumes 16 byte slots. We could do better. */
return slots * 16;
}
static inline mali_ptr
panfrost_emit_resources(struct panfrost_batch *batch,
enum pipe_shader_type stage)
{
struct panfrost_context *ctx = batch->ctx;
struct panfrost_ptr T;
unsigned nr_tables = 12;
/* Although individual resources need only 16 byte alignment, the
* resource table as a whole must be 64-byte aligned.
*/
T = pan_pool_alloc_aligned(&batch->pool.base, nr_tables * pan_size(RESOURCE),
64);
memset(T.cpu, 0, nr_tables * pan_size(RESOURCE));
panfrost_make_resource_table(T, PAN_TABLE_UBO, batch->uniform_buffers[stage],
batch->nr_uniform_buffers[stage]);
panfrost_make_resource_table(T, PAN_TABLE_TEXTURE, batch->textures[stage],
ctx->sampler_view_count[stage]);
/* We always need at least 1 sampler for txf to work */
panfrost_make_resource_table(T, PAN_TABLE_SAMPLER, batch->samplers[stage],
MAX2(ctx->sampler_count[stage], 1));
panfrost_make_resource_table(T, PAN_TABLE_IMAGE, batch->images[stage],
util_last_bit(ctx->image_mask[stage]));
if (stage == PIPE_SHADER_VERTEX) {
panfrost_make_resource_table(T, PAN_TABLE_ATTRIBUTE,
batch->attribs[stage],
ctx->vertex->num_elements);
panfrost_make_resource_table(T, PAN_TABLE_ATTRIBUTE_BUFFER,
batch->attrib_bufs[stage],
util_last_bit(ctx->vb_mask));
}
return T.gpu | nr_tables;
}
#endif /* PAN_ARCH >= 9 */
static bool
allow_rotating_primitives(const struct panfrost_compiled_shader *fs,
const struct pipe_draw_info *info)
{
return u_reduced_prim(info->mode) != MESA_PRIM_LINES &&
!fs->info.bifrost.uses_flat_shading;
}
#endif