/* * Copyright 2021 Collabora, Ltd. * SPDX-License-Identifier: MIT */ #include "agx_test.h" #include static void agx_optimize_and_dce(agx_context *ctx) { agx_optimizer(ctx); agx_dce(ctx, true); } #define CASE(instr, expected, size, returns) \ INSTRUCTION_CASE( \ { \ UNUSED agx_index out = agx_temp(b->shader, AGX_SIZE_##size); \ instr; \ if (returns) \ agx_unit_test(b, out); \ }, \ { \ UNUSED agx_index out = agx_temp(b->shader, AGX_SIZE_##size); \ expected; \ if (returns) \ agx_unit_test(b, out); \ }, \ agx_optimize_and_dce) #define NEGCASE(instr, size) CASE(instr, instr, size, true) #define CASE16(instr, expected) CASE(instr, expected, 16, true) #define CASE32(instr, expected) CASE(instr, expected, 32, true) #define CASE_NO_RETURN(instr, expected) \ CASE(instr, expected, 32 /* irrelevant */, false) #define NEGCASE16(instr) NEGCASE(instr, 16) #define NEGCASE32(instr) NEGCASE(instr, 32) static inline agx_index agx_fmov(agx_builder *b, agx_index s0) { agx_index tmp = agx_temp(b->shader, s0.size); agx_fmov_to(b, tmp, s0); return tmp; } class Optimizer : public testing::Test { protected: Optimizer() { mem_ctx = ralloc_context(NULL); wx = agx_register(0, AGX_SIZE_32); wy = agx_register(2, AGX_SIZE_32); wz = agx_register(4, AGX_SIZE_32); hx = agx_register(0, AGX_SIZE_16); hy = agx_register(1, AGX_SIZE_16); hz = agx_register(2, AGX_SIZE_16); } ~Optimizer() { ralloc_free(mem_ctx); } void *mem_ctx; agx_index wx, wy, wz, hx, hy, hz; }; TEST_F(Optimizer, FloatCopyprop) { CASE32(agx_fadd_to(b, out, agx_abs(agx_fmov(b, wx)), wy), agx_fadd_to(b, out, agx_abs(wx), wy)); CASE32(agx_fadd_to(b, out, agx_neg(agx_fmov(b, wx)), wy), agx_fadd_to(b, out, agx_neg(wx), wy)); } TEST_F(Optimizer, FloatConversion) { CASE32( { agx_index cvt = agx_temp(b->shader, AGX_SIZE_32); agx_fmov_to(b, cvt, hx); agx_fadd_to(b, out, cvt, wy); }, { agx_fadd_to(b, out, hx, wy); }); CASE16( { agx_index sum = agx_temp(b->shader, AGX_SIZE_32); agx_fadd_to(b, sum, wx, wy); agx_fmov_to(b, out, sum); }, { agx_fadd_to(b, out, wx, wy); }); } TEST_F(Optimizer, FusedFABSNEG) { CASE32(agx_fadd_to(b, out, agx_fmov(b, agx_abs(wx)), wy), agx_fadd_to(b, out, agx_abs(wx), wy)); CASE32(agx_fmul_to(b, out, wx, agx_fmov(b, agx_neg(agx_abs(wx)))), agx_fmul_to(b, out, wx, agx_neg(agx_abs(wx)))); } TEST_F(Optimizer, FusedFabsAbsorb) { CASE32(agx_fadd_to(b, out, agx_abs(agx_fmov(b, agx_abs(wx))), wy), agx_fadd_to(b, out, agx_abs(wx), wy)); } TEST_F(Optimizer, FusedFnegCancel) { CASE32(agx_fmul_to(b, out, wx, agx_neg(agx_fmov(b, agx_neg(wx)))), agx_fmul_to(b, out, wx, wx)); CASE32(agx_fmul_to(b, out, wx, agx_neg(agx_fmov(b, agx_neg(agx_abs(wx))))), agx_fmul_to(b, out, wx, agx_abs(wx))); } TEST_F(Optimizer, FmulFsatF2F16) { CASE16( { agx_index tmp = agx_temp(b->shader, AGX_SIZE_32); agx_fmov_to(b, tmp, agx_fmul(b, wx, wy))->saturate = true; agx_fmov_to(b, out, tmp); }, { agx_fmul_to(b, out, wx, wy)->saturate = true; }); } TEST_F(Optimizer, Copyprop) { CASE32(agx_fmul_to(b, out, wx, agx_mov(b, wy)), agx_fmul_to(b, out, wx, wy)); CASE32(agx_fmul_to(b, out, agx_mov(b, wx), agx_mov(b, wy)), agx_fmul_to(b, out, wx, wy)); } TEST_F(Optimizer, InlineHazards) { NEGCASE32({ agx_instr *I = agx_collect_to(b, out, 4); I->src[0] = agx_mov_imm(b, AGX_SIZE_32, 0); I->src[1] = wy; I->src[2] = wz; I->src[3] = wz; }); } TEST_F(Optimizer, CopypropRespectsAbsNeg) { CASE32(agx_fadd_to(b, out, agx_abs(agx_mov(b, wx)), wy), agx_fadd_to(b, out, agx_abs(wx), wy)); CASE32(agx_fadd_to(b, out, agx_neg(agx_mov(b, wx)), wy), agx_fadd_to(b, out, agx_neg(wx), wy)); CASE32(agx_fadd_to(b, out, agx_neg(agx_abs(agx_mov(b, wx))), wy), agx_fadd_to(b, out, agx_neg(agx_abs(wx)), wy)); } TEST_F(Optimizer, IntCopyprop) { CASE32(agx_xor_to(b, out, agx_mov(b, wx), wy), agx_xor_to(b, out, wx, wy)); } TEST_F(Optimizer, IntCopypropDoesntConvert) { NEGCASE32({ agx_index cvt = agx_temp(b->shader, AGX_SIZE_32); agx_mov_to(b, cvt, hx); agx_xor_to(b, out, cvt, wy); }); } TEST_F(Optimizer, SkipPreloads) { NEGCASE32({ agx_index preload = agx_preload(b, agx_register(0, AGX_SIZE_32)); agx_xor_to(b, out, preload, wy); }); } TEST_F(Optimizer, NoConversionsOn16BitALU) { NEGCASE16({ agx_index cvt = agx_temp(b->shader, AGX_SIZE_16); agx_fmov_to(b, cvt, wx); agx_fadd_to(b, out, cvt, hy); }); NEGCASE32(agx_fmov_to(b, out, agx_fadd(b, hx, hy))); } TEST_F(Optimizer, IfCondition) { CASE_NO_RETURN(agx_if_icmp(b, agx_icmp(b, wx, wy, AGX_ICOND_UEQ, true), agx_zero(), 1, AGX_ICOND_UEQ, true), agx_if_icmp(b, wx, wy, 1, AGX_ICOND_UEQ, true)); CASE_NO_RETURN(agx_if_icmp(b, agx_fcmp(b, wx, wy, AGX_FCOND_EQ, true), agx_zero(), 1, AGX_ICOND_UEQ, true), agx_if_fcmp(b, wx, wy, 1, AGX_FCOND_EQ, true)); CASE_NO_RETURN(agx_if_icmp(b, agx_fcmp(b, hx, hy, AGX_FCOND_LT, false), agx_zero(), 1, AGX_ICOND_UEQ, true), agx_if_fcmp(b, hx, hy, 1, AGX_FCOND_LT, false)); } TEST_F(Optimizer, SelectCondition) { CASE32(agx_icmpsel_to(b, out, agx_icmp(b, wx, wy, AGX_ICOND_UEQ, false), agx_zero(), wz, wx, AGX_ICOND_UEQ), agx_icmpsel_to(b, out, wx, wy, wx, wz, AGX_ICOND_UEQ)); CASE32(agx_icmpsel_to(b, out, agx_icmp(b, wx, wy, AGX_ICOND_UEQ, true), agx_zero(), wz, wx, AGX_ICOND_UEQ), agx_icmpsel_to(b, out, wx, wy, wz, wx, AGX_ICOND_UEQ)); CASE32(agx_icmpsel_to(b, out, agx_fcmp(b, wx, wy, AGX_FCOND_EQ, false), agx_zero(), wz, wx, AGX_ICOND_UEQ), agx_fcmpsel_to(b, out, wx, wy, wx, wz, AGX_FCOND_EQ)); CASE32(agx_icmpsel_to(b, out, agx_fcmp(b, wx, wy, AGX_FCOND_LT, true), agx_zero(), wz, wx, AGX_ICOND_UEQ), agx_fcmpsel_to(b, out, wx, wy, wz, wx, AGX_FCOND_LT)); }