mirror of
https://gitlab.freedesktop.org/mesa/mesa.git
synced 2025-12-24 21:50:12 +01:00
Signed-off-by: Rhys Perry <pendingchaos02@gmail.com> Reviewed-by: Georg Lehmann <dadschoorse@gmail.com> Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/25525>
2194 lines
85 KiB
C++
2194 lines
85 KiB
C++
/*
|
|
* Copyright © 2020 Valve Corporation
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
* IN THE SOFTWARE.
|
|
*
|
|
*/
|
|
#include "helpers.h"
|
|
|
|
using namespace aco;
|
|
|
|
BEGIN_TEST(optimize.neg)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, v1: %b, s1: %c, s1: %d = p_startpgm
|
|
if (!setup_cs("v1 v1 s1 s1", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
//! v1: %res0 = v_mul_f32 %a, -%b
|
|
//! p_unit_test 0, %res0
|
|
Temp neg_b = fneg(inputs[1]);
|
|
writeout(0, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], neg_b));
|
|
|
|
//~gfx9! v1: %neg_a = v_mul_f32 -1.0, %a
|
|
//~gfx9! v1: %res1 = v_mul_f32 0x123456, %neg_a
|
|
//~gfx10! v1: %res1 = v_mul_f32 0x123456, -%a
|
|
//! p_unit_test 1, %res1
|
|
Temp neg_a = fneg(inputs[0]);
|
|
writeout(1, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x123456u), neg_a));
|
|
|
|
//! v1: %res2 = v_mul_f32 %a, %b
|
|
//! p_unit_test 2, %res2
|
|
Temp neg_neg_a = fneg(neg_a);
|
|
writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), neg_neg_a, inputs[1]));
|
|
|
|
//! v1: %res3 = v_mul_f32 |%a|, %b
|
|
//! p_unit_test 3, %res3
|
|
Temp abs_neg_a = fabs(neg_a);
|
|
writeout(3, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), abs_neg_a, inputs[1]));
|
|
|
|
//! v1: %res4 = v_mul_f32 -|%a|, %b
|
|
//! p_unit_test 4, %res4
|
|
Temp abs_a = fabs(inputs[0]);
|
|
Temp neg_abs_a = fneg(abs_a);
|
|
writeout(4, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), neg_abs_a, inputs[1]));
|
|
|
|
//~gfx9! v1: %res5 = v_mul_f32 -%a, %b row_shl:1 bound_ctrl:1
|
|
//~gfx10! v1: %res5 = v_mul_f32 -%a, %b row_shl:1 bound_ctrl:1 fi
|
|
//! p_unit_test 5, %res5
|
|
writeout(5,
|
|
bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), neg_a, inputs[1], dpp_row_sl(1)));
|
|
|
|
//! v1: %res6 = v_subrev_f32 %a, %b
|
|
//! p_unit_test 6, %res6
|
|
writeout(6, bld.vop2(aco_opcode::v_add_f32, bld.def(v1), neg_a, inputs[1]));
|
|
|
|
//! v1: %res7 = v_sub_f32 %b, %a
|
|
//! p_unit_test 7, %res7
|
|
writeout(7, bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[1], neg_a));
|
|
|
|
//! v1: %res8 = v_mul_f32 %a, -%c
|
|
//! p_unit_test 8, %res8
|
|
Temp neg_c = fneg(bld.copy(bld.def(v1), inputs[2]));
|
|
writeout(8, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], neg_c));
|
|
|
|
// //! v1: %res9 = v_mul_f32 |%neg_a|, %b
|
|
// //! p_unit_test 9, %res9
|
|
Temp abs_neg_abs_a = fabs(neg_abs_a);
|
|
writeout(9, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), abs_neg_abs_a, inputs[1]));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.output_modifiers)
|
|
//>> v1: %a, v1: %b = p_startpgm
|
|
if (!setup_cs("v1 v1", GFX9))
|
|
return;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
/* 32-bit modifiers */
|
|
|
|
//! v1: %res0 = v_add_f32 %a, %b *0.5
|
|
//! p_unit_test 0, %res0
|
|
Temp tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
writeout(0, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x3f000000u), tmp));
|
|
|
|
//! v1: %res1 = v_add_f32 %a, %b *2
|
|
//! p_unit_test 1, %res1
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
writeout(1, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
|
|
|
|
//! v1: %res2 = v_add_f32 %a, %b *4
|
|
//! p_unit_test 2, %res2
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40800000u), tmp));
|
|
|
|
//! v1: %res3 = v_add_f32 %a, %b clamp
|
|
//! p_unit_test 3, %res3
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
writeout(3, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(),
|
|
Operand::c32(0x3f800000u), tmp));
|
|
|
|
//! v1: %res4 = v_add_f32 %a, %b *2 clamp
|
|
//! p_unit_test 4, %res4
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
tmp = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp);
|
|
writeout(4, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(),
|
|
Operand::c32(0x3f800000u), tmp));
|
|
|
|
/* 16-bit modifiers */
|
|
|
|
//! v2b: %res5 = v_add_f16 %a, %b *0.5
|
|
//! p_unit_test 5, %res5
|
|
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
|
|
writeout(5, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x3800u), tmp));
|
|
|
|
//! v2b: %res6 = v_add_f16 %a, %b *2
|
|
//! p_unit_test 6, %res6
|
|
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
|
|
writeout(6, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp));
|
|
|
|
//! v2b: %res7 = v_add_f16 %a, %b *4
|
|
//! p_unit_test 7, %res7
|
|
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
|
|
writeout(7, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4400u), tmp));
|
|
|
|
//! v2b: %res8 = v_add_f16 %a, %b clamp
|
|
//! p_unit_test 8, %res8
|
|
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
|
|
writeout(8, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u),
|
|
Operand::c16(0x3c00u), tmp));
|
|
|
|
//! v2b: %res9 = v_add_f16 %a, %b *2 clamp
|
|
//! p_unit_test 9, %res9
|
|
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
|
|
tmp = bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000), tmp);
|
|
writeout(9, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u),
|
|
Operand::c16(0x3c00u), tmp));
|
|
|
|
/* clamping is done after omod */
|
|
|
|
//! v1: %res10_tmp = v_add_f32 %a, %b clamp
|
|
//! v1: %res10 = v_mul_f32 2.0, %res10_tmp
|
|
//! p_unit_test 10, %res10
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
tmp = bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(), Operand::c32(0x3f800000u),
|
|
tmp);
|
|
writeout(10, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
|
|
|
|
/* unsupported instructions */
|
|
|
|
//! v1: %res11_tmp = v_xor_b32 %a, %b
|
|
//! v1: %res11 = v_mul_f32 2.0, %res11_tmp
|
|
//! p_unit_test 11, %res11
|
|
tmp = bld.vop2(aco_opcode::v_xor_b32, bld.def(v1), inputs[0], inputs[1]);
|
|
writeout(11, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
|
|
|
|
/* several users */
|
|
|
|
//! v1: %res12_tmp = v_add_f32 %a, %b
|
|
//! p_unit_test %res12_tmp
|
|
//! v1: %res12 = v_mul_f32 2.0, %res12_tmp
|
|
//! p_unit_test 12, %res12
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
bld.pseudo(aco_opcode::p_unit_test, tmp);
|
|
writeout(12, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
|
|
|
|
//! v1: %res13 = v_add_f32 %a, %b
|
|
//! p_unit_test 13, %res13
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp);
|
|
writeout(13, tmp);
|
|
|
|
/* omod has no effect if denormals are enabled but clamp is fine */
|
|
|
|
//>> BB1
|
|
//! /* logical preds: / linear preds: / kind: uniform, */
|
|
program->next_fp_mode.denorm32 = fp_denorm_keep;
|
|
program->next_fp_mode.denorm16_64 = fp_denorm_flush;
|
|
bld.reset(program->create_and_insert_block());
|
|
|
|
//! v1: %res14_tmp = v_add_f32 %a, %b
|
|
//! v1: %res14 = v_mul_f32 2.0, %res13_tmp
|
|
//! p_unit_test 14, %res14
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
writeout(14, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
|
|
|
|
//! v1: %res15 = v_add_f32 %a, %b clamp
|
|
//! p_unit_test 15, %res15
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
writeout(15, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(),
|
|
Operand::c32(0x3f800000u), tmp));
|
|
|
|
//>> BB2
|
|
//! /* logical preds: / linear preds: / kind: uniform, */
|
|
program->next_fp_mode.denorm32 = fp_denorm_flush;
|
|
program->next_fp_mode.denorm16_64 = fp_denorm_keep;
|
|
bld.reset(program->create_and_insert_block());
|
|
|
|
//! v2b: %res16_tmp = v_add_f16 %a, %b
|
|
//! v2b: %res16 = v_mul_f16 2.0, %res15_tmp
|
|
//! p_unit_test 16, %res16
|
|
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
|
|
writeout(16, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp));
|
|
|
|
//! v2b: %res17 = v_add_f16 %a, %b clamp
|
|
//! p_unit_test 17, %res17
|
|
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
|
|
writeout(17, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u),
|
|
Operand::c16(0x3c00u), tmp));
|
|
|
|
/* omod flushes -0.0 to +0.0 */
|
|
|
|
//>> BB3
|
|
//! /* logical preds: / linear preds: / kind: uniform, */
|
|
program->next_fp_mode.denorm32 = fp_denorm_keep;
|
|
program->next_fp_mode.denorm16_64 = fp_denorm_keep;
|
|
program->next_fp_mode.preserve_signed_zero_inf_nan32 = true;
|
|
program->next_fp_mode.preserve_signed_zero_inf_nan16_64 = false;
|
|
bld.reset(program->create_and_insert_block());
|
|
|
|
//! v1: %res18_tmp = v_add_f32 %a, %b
|
|
//! v1: %res18 = v_mul_f32 2.0, %res18_tmp
|
|
//! p_unit_test 18, %res18
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
writeout(18, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
|
|
//! v1: %res19 = v_add_f32 %a, %b clamp
|
|
//! p_unit_test 19, %res19
|
|
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
|
|
writeout(19, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(),
|
|
Operand::c32(0x3f800000u), tmp));
|
|
|
|
//>> BB4
|
|
//! /* logical preds: / linear preds: / kind: uniform, */
|
|
program->next_fp_mode.preserve_signed_zero_inf_nan32 = false;
|
|
program->next_fp_mode.preserve_signed_zero_inf_nan16_64 = true;
|
|
bld.reset(program->create_and_insert_block());
|
|
//! v2b: %res20_tmp = v_add_f16 %a, %b
|
|
//! v2b: %res20 = v_mul_f16 2.0, %res20_tmp
|
|
//! p_unit_test 20, %res20
|
|
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
|
|
writeout(20, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp));
|
|
//! v2b: %res21 = v_add_f16 %a, %b clamp
|
|
//! p_unit_test 21, %res21
|
|
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
|
|
writeout(21, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u),
|
|
Operand::c16(0x3c00u), tmp));
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
Temp
|
|
create_subbrev_co(Operand op0, Operand op1, Operand op2)
|
|
{
|
|
return bld.vop2_e64(aco_opcode::v_subbrev_co_u32, bld.def(v1), bld.def(bld.lm), op0, op1, op2);
|
|
}
|
|
|
|
BEGIN_TEST(optimize.cndmask)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, s1: %b, s2: %c = p_startpgm
|
|
if (!setup_cs("v1 s1 s2", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
Temp subbrev;
|
|
|
|
//! v1: %res0 = v_cndmask_b32 0, %a, %c
|
|
//! p_unit_test 0, %res0
|
|
subbrev = create_subbrev_co(Operand::zero(), Operand::zero(), Operand(inputs[2]));
|
|
writeout(0, bld.vop2(aco_opcode::v_and_b32, bld.def(v1), inputs[0], subbrev));
|
|
|
|
//! v1: %res1 = v_cndmask_b32 0, 42, %c
|
|
//! p_unit_test 1, %res1
|
|
subbrev = create_subbrev_co(Operand::zero(), Operand::zero(), Operand(inputs[2]));
|
|
writeout(1, bld.vop2(aco_opcode::v_and_b32, bld.def(v1), Operand::c32(42u), subbrev));
|
|
|
|
//~gfx9! v1: %subbrev, s2: %_ = v_subbrev_co_u32 0, 0, %c
|
|
//~gfx9! v1: %res2 = v_and_b32 %b, %subbrev
|
|
//~gfx10! v1: %res2 = v_cndmask_b32 0, %b, %c
|
|
//! p_unit_test 2, %res2
|
|
subbrev = create_subbrev_co(Operand::zero(), Operand::zero(), Operand(inputs[2]));
|
|
writeout(2, bld.vop2(aco_opcode::v_and_b32, bld.def(v1), inputs[1], subbrev));
|
|
|
|
//! v1: %subbrev1, s2: %_ = v_subbrev_co_u32 0, 0, %c
|
|
//! v1: %xor = v_xor_b32 %a, %subbrev1
|
|
//! v1: %res3 = v_cndmask_b32 0, %xor, %c
|
|
//! p_unit_test 3, %res3
|
|
subbrev = create_subbrev_co(Operand::zero(), Operand::zero(), Operand(inputs[2]));
|
|
Temp xor_a = bld.vop2(aco_opcode::v_xor_b32, bld.def(v1), inputs[0], subbrev);
|
|
writeout(3, bld.vop2(aco_opcode::v_and_b32, bld.def(v1), xor_a, subbrev));
|
|
|
|
//! v1: %res4 = v_cndmask_b32 0, %a, %c
|
|
//! p_unit_test 4, %res4
|
|
Temp cndmask = bld.vop2_e64(aco_opcode::v_cndmask_b32, bld.def(v1), Operand::zero(),
|
|
Operand::c32(1u), Operand(inputs[2]));
|
|
Temp sub = bld.vsub32(bld.def(v1), Operand::zero(), cndmask);
|
|
writeout(4, bld.vop2(aco_opcode::v_and_b32, bld.def(v1), Operand(inputs[0]), sub));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.add_lshl)
|
|
for (unsigned i = GFX8; i <= GFX10; i++) {
|
|
//>> s1: %a, v1: %b = p_startpgm
|
|
if (!setup_cs("s1 v1", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
Temp shift;
|
|
|
|
//~gfx8! s1: %lshl0, s1: %_:scc = s_lshl_b32 %a, 3
|
|
//~gfx8! s1: %res0, s1: %_:scc = s_add_u32 %lshl0, 4
|
|
//~gfx(9|10)! s1: %res0, s1: %_:scc = s_lshl3_add_u32 %a, 4
|
|
//! p_unit_test 0, %res0
|
|
shift = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), Operand(inputs[0]),
|
|
Operand::c32(3u));
|
|
writeout(0, bld.sop2(aco_opcode::s_add_u32, bld.def(s1), bld.def(s1, scc), shift,
|
|
Operand::c32(4u)));
|
|
|
|
//~gfx8! s1: %lshl1, s1: %_:scc = s_lshl_b32 %a, 3
|
|
//~gfx8! s1: %add1, s1: %_:scc = s_add_u32 %lshl1, 4
|
|
//~gfx8! v1: %add_co1, s2: %_ = v_add_co_u32 %lshl1, %b
|
|
//~gfx8! v1: %res1, s2: %_ = v_add_co_u32 %add1, %add_co1
|
|
//~gfx(9|10)! s1: %lshl1, s1: %_:scc = s_lshl3_add_u32 %a, 4
|
|
//~gfx(9|10)! v1: %lshl_add = v_lshl_add_u32 %a, 3, %b
|
|
//~gfx(9|10)! v1: %res1 = v_add_u32 %lshl1, %lshl_add
|
|
//! p_unit_test 1, %res1
|
|
shift = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), Operand(inputs[0]),
|
|
Operand::c32(3u));
|
|
Temp sadd =
|
|
bld.sop2(aco_opcode::s_add_u32, bld.def(s1), bld.def(s1, scc), shift, Operand::c32(4u));
|
|
Temp vadd = bld.vadd32(bld.def(v1), shift, Operand(inputs[1]));
|
|
writeout(1, bld.vadd32(bld.def(v1), sadd, vadd));
|
|
|
|
//~gfx8! s1: %lshl2 = s_lshl_b32 %a, 3
|
|
//~gfx8! v1: %res2, s2: %_ = v_add_co_u32 %lshl2, %b
|
|
//~gfx(9|10)! v1: %res2 = v_lshl_add_u32 %a, 3, %b
|
|
//! p_unit_test 2, %res2
|
|
Temp lshl =
|
|
bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), Operand(inputs[0]), Operand::c32(3u));
|
|
writeout(2, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
//~gfx8! s1: %lshl3 = s_lshl_b32 (is24bit)%a, 7
|
|
//~gfx8! v1: %res3, s2: %_ = v_add_co_u32 %lshl3, %b
|
|
//~gfx(9|10)! v1: %res3 = v_lshl_add_u32 (is24bit)%a, 7, %b
|
|
//! p_unit_test 3, %res3
|
|
Operand a_24bit = Operand(inputs[0]);
|
|
a_24bit.set24bit(true);
|
|
lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_24bit, Operand::c32(7u));
|
|
writeout(3, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
//! s1: %lshl4 = s_lshl_b32 (is24bit)%a, 3
|
|
//~gfx(8|9)! v1: %res4, s2: %carry = v_add_co_u32 %lshl4, %b
|
|
//~gfx10! v1: %res4, s2: %carry = v_add_co_u32_e64 %lshl4, %b
|
|
//! p_unit_test 4, %carry
|
|
lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_24bit, Operand::c32(3u));
|
|
Temp carry = bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]), true).def(1).getTemp();
|
|
writeout(4, carry);
|
|
|
|
//~gfx8! s1: %lshl5 = s_lshl_b32 (is24bit)%a, (is24bit)%a
|
|
//~gfx8! v1: %res5, s2: %_ = v_add_co_u32 %lshl5, %b
|
|
//~gfx(9|10)! v1: %res5 = v_lshl_add_u32 (is24bit)%a, (is24bit)%a, %b
|
|
//! p_unit_test 5, %res5
|
|
lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_24bit, a_24bit);
|
|
writeout(5, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
//~gfx8! v1: %res6 = v_mad_u32_u24 (is24bit)%a, 8, %b
|
|
//~gfx(9|10)! v1: %res6 = v_lshl_add_u32 (is24bit)%a, 3, %b
|
|
//! p_unit_test 6, %res6
|
|
lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_24bit, Operand::c32(3u));
|
|
writeout(6, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
//~gfx8! v1: %res7 = v_mad_u32_u24 (is16bit)%a, 16, %b
|
|
//~gfx(9|10)! v1: %res7 = v_lshl_add_u32 (is16bit)%a, 4, %b
|
|
//! p_unit_test 7, %res7
|
|
Operand a_16bit = Operand(inputs[0]);
|
|
a_16bit.set16bit(true);
|
|
lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_16bit, Operand::c32(4u));
|
|
writeout(7, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.bcnt)
|
|
for (unsigned i = GFX8; i <= GFX10; i++) {
|
|
//>> v1: %a, s1: %b = p_startpgm
|
|
if (!setup_cs("v1 s1", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
Temp bcnt;
|
|
|
|
//! v1: %res0 = v_bcnt_u32_b32 %a, %a
|
|
//! p_unit_test 0, %res0
|
|
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero());
|
|
writeout(0, bld.vadd32(bld.def(v1), bcnt, Operand(inputs[0])));
|
|
|
|
//! v1: %res1 = v_bcnt_u32_b32 %a, %b
|
|
//! p_unit_test 1, %res1
|
|
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero());
|
|
writeout(1, bld.vadd32(bld.def(v1), bcnt, Operand(inputs[1])));
|
|
|
|
//! v1: %res2 = v_bcnt_u32_b32 %a, 42
|
|
//! p_unit_test 2, %res2
|
|
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero());
|
|
writeout(2, bld.vadd32(bld.def(v1), bcnt, Operand::c32(42u)));
|
|
|
|
//! v1: %bnct3 = v_bcnt_u32_b32 %b, 0
|
|
//~gfx8! v1: %res3, s2: %_ = v_add_co_u32 %bcnt3, %a
|
|
//~gfx(9|10)! v1: %res3 = v_add_u32 %bcnt3, %a
|
|
//! p_unit_test 3, %res3
|
|
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[1]), Operand::zero());
|
|
writeout(3, bld.vadd32(bld.def(v1), bcnt, Operand(inputs[0])));
|
|
|
|
//! v1: %bnct4 = v_bcnt_u32_b32 %a, 0
|
|
//~gfx(8|9)! v1: %add4, s2: %carry = v_add_co_u32 %bcnt4, %a
|
|
//~gfx10! v1: %add4, s2: %carry = v_add_co_u32_e64 %bcnt4, %a
|
|
//! p_unit_test 4, %carry
|
|
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero());
|
|
Temp carry = bld.vadd32(bld.def(v1), bcnt, Operand(inputs[0]), true).def(1).getTemp();
|
|
writeout(4, carry);
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
struct clamp_config {
|
|
const char* name;
|
|
aco_opcode min, max, med3;
|
|
Operand lb, ub;
|
|
};
|
|
|
|
static const clamp_config clamp_configs[] = {
|
|
/* 0.0, 4.0 */
|
|
{"_0,4f32", aco_opcode::v_min_f32, aco_opcode::v_max_f32, aco_opcode::v_med3_f32,
|
|
Operand::zero(), Operand::c32(0x40800000u)},
|
|
{"_0,4f16", aco_opcode::v_min_f16, aco_opcode::v_max_f16, aco_opcode::v_med3_f16,
|
|
Operand::c16(0u), Operand::c16(0x4400)},
|
|
/* -1.0, 0.0 */
|
|
{"_-1,0f32", aco_opcode::v_min_f32, aco_opcode::v_max_f32, aco_opcode::v_med3_f32,
|
|
Operand::c32(0xbf800000u), Operand::zero()},
|
|
{"_-1,0f16", aco_opcode::v_min_f16, aco_opcode::v_max_f16, aco_opcode::v_med3_f16,
|
|
Operand::c16(0xBC00), Operand::c16(0u)},
|
|
/* 0, 3 */
|
|
{"_0,3u32", aco_opcode::v_min_u32, aco_opcode::v_max_u32, aco_opcode::v_med3_u32,
|
|
Operand::zero(), Operand::c32(3u)},
|
|
{"_0,3u16", aco_opcode::v_min_u16, aco_opcode::v_max_u16, aco_opcode::v_med3_u16,
|
|
Operand::c16(0u), Operand::c16(3u)},
|
|
{"_0,3i32", aco_opcode::v_min_i32, aco_opcode::v_max_i32, aco_opcode::v_med3_i32,
|
|
Operand::zero(), Operand::c32(3u)},
|
|
{"_0,3i16", aco_opcode::v_min_i16, aco_opcode::v_max_i16, aco_opcode::v_med3_i16,
|
|
Operand::c16(0u), Operand::c16(3u)},
|
|
/* -5, 0 */
|
|
{"_-5,0i32", aco_opcode::v_min_i32, aco_opcode::v_max_i32, aco_opcode::v_med3_i32,
|
|
Operand::c32(0xfffffffbu), Operand::zero()},
|
|
{"_-5,0i16", aco_opcode::v_min_i16, aco_opcode::v_max_i16, aco_opcode::v_med3_i16,
|
|
Operand::c16(0xfffbu), Operand::c16(0u)},
|
|
};
|
|
|
|
BEGIN_TEST(optimize.clamp)
|
|
for (clamp_config cfg : clamp_configs) {
|
|
if (!setup_cs("v1 v1 v1", GFX9, CHIP_UNKNOWN, cfg.name))
|
|
continue;
|
|
|
|
//! cfg: @match_func(min max med3 lb ub)
|
|
fprintf(output, "cfg: %s ", instr_info.name[(int)cfg.min]);
|
|
fprintf(output, "%s ", instr_info.name[(int)cfg.max]);
|
|
fprintf(output, "%s ", instr_info.name[(int)cfg.med3]);
|
|
aco_print_operand(&cfg.lb, output);
|
|
fprintf(output, " ");
|
|
aco_print_operand(&cfg.ub, output);
|
|
fprintf(output, "\n");
|
|
|
|
//>> v1: %a, v1: %b, v1: %c = p_startpgm
|
|
|
|
//! v1: %res0 = @med3 @ub, @lb, %a
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, bld.vop2(cfg.min, bld.def(v1), cfg.ub,
|
|
bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0])));
|
|
|
|
//! v1: %res1 = @med3 @lb, @ub, %a
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, bld.vop2(cfg.max, bld.def(v1), cfg.lb,
|
|
bld.vop2(cfg.min, bld.def(v1), cfg.ub, inputs[0])));
|
|
|
|
/* min constant must be greater than max constant */
|
|
//! v1: %res2_tmp = @min @lb, %a
|
|
//! v1: %res2 = @max @ub, %res2_tmp
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, bld.vop2(cfg.max, bld.def(v1), cfg.ub,
|
|
bld.vop2(cfg.min, bld.def(v1), cfg.lb, inputs[0])));
|
|
|
|
//! v1: %res3_tmp = @max @ub, %a
|
|
//! v1: %res3 = @min @lb, %res3_tmp
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, bld.vop2(cfg.min, bld.def(v1), cfg.lb,
|
|
bld.vop2(cfg.max, bld.def(v1), cfg.ub, inputs[0])));
|
|
|
|
/* needs two constants */
|
|
|
|
//! v1: %res4_tmp = @max @lb, %a
|
|
//! v1: %res4 = @min %b, %res4_tmp
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, bld.vop2(cfg.min, bld.def(v1), inputs[1],
|
|
bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0])));
|
|
|
|
//! v1: %res5_tmp = @max %b, %a
|
|
//! v1: %res5 = @min @ub, %res5_tmp
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, bld.vop2(cfg.min, bld.def(v1), cfg.ub,
|
|
bld.vop2(cfg.max, bld.def(v1), inputs[1], inputs[0])));
|
|
|
|
//! v1: %res6_tmp = @max %c, %a
|
|
//! v1: %res6 = @min %b, %res6_tmp
|
|
//! p_unit_test 6, %res6
|
|
writeout(6, bld.vop2(cfg.min, bld.def(v1), inputs[1],
|
|
bld.vop2(cfg.max, bld.def(v1), inputs[2], inputs[0])));
|
|
|
|
/* correct NaN behaviour with precise */
|
|
if (cfg.min == aco_opcode::v_min_f16 || cfg.min == aco_opcode::v_min_f32) {
|
|
//~f(16|32)! v1: %res7 = @med3 @ub, @lb, %a
|
|
//~f(16|32)! p_unit_test 7, %res7
|
|
Builder::Result max = bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0]);
|
|
max.def(0).setPrecise(true);
|
|
Builder::Result min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, max);
|
|
max.def(0).setPrecise(true);
|
|
writeout(7, min);
|
|
|
|
//~f(16|32)! v1: (precise)%res8_tmp = @min @ub, %a
|
|
//~f(16|32)! v1: %res8 = @max @lb, %res8_tmp
|
|
//~f(16|32)! p_unit_test 8, %res8
|
|
min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, inputs[0]);
|
|
min.def(0).setPrecise(true);
|
|
writeout(8, bld.vop2(cfg.max, bld.def(v1), cfg.lb, min));
|
|
}
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.const_comparison_ordering)
|
|
//>> v1: %a, v1: %b, v2: %c, v1: %d = p_startpgm
|
|
if (!setup_cs("v1 v1 v2 v1", GFX9))
|
|
return;
|
|
|
|
/* optimize to unordered comparison */
|
|
//! s2: %res0 = v_cmp_nge_f32 4.0, %a
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc),
|
|
bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), inputs[0], inputs[0]),
|
|
bld.vopc(aco_opcode::v_cmp_lt_f32, bld.def(bld.lm),
|
|
Operand::c32(0x40800000u), inputs[0])));
|
|
|
|
//! s2: %res1 = v_cmp_nge_f32 4.0, %a
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc),
|
|
bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), inputs[0], inputs[0]),
|
|
bld.vopc(aco_opcode::v_cmp_nge_f32, bld.def(bld.lm),
|
|
Operand::c32(0x40800000u), inputs[0])));
|
|
|
|
//! s2: %res2 = v_cmp_nge_f32 0x40a00000, %a
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc),
|
|
bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), inputs[0], inputs[0]),
|
|
bld.vopc(aco_opcode::v_cmp_lt_f32, bld.def(bld.lm),
|
|
bld.copy(bld.def(v1), Operand::c32(0x40a00000u)), inputs[0])));
|
|
|
|
/* optimize to ordered comparison */
|
|
//! s2: %res3 = v_cmp_lt_f32 4.0, %a
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, bld.sop2(aco_opcode::s_and_b64, bld.def(bld.lm), bld.def(s1, scc),
|
|
bld.vopc(aco_opcode::v_cmp_eq_f32, bld.def(bld.lm), inputs[0], inputs[0]),
|
|
bld.vopc(aco_opcode::v_cmp_nge_f32, bld.def(bld.lm),
|
|
Operand::c32(0x40800000u), inputs[0])));
|
|
|
|
//! s2: %res4 = v_cmp_lt_f32 4.0, %a
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, bld.sop2(aco_opcode::s_and_b64, bld.def(bld.lm), bld.def(s1, scc),
|
|
bld.vopc(aco_opcode::v_cmp_eq_f32, bld.def(bld.lm), inputs[0], inputs[0]),
|
|
bld.vopc(aco_opcode::v_cmp_lt_f32, bld.def(bld.lm),
|
|
Operand::c32(0x40800000u), inputs[0])));
|
|
|
|
//! s2: %res5 = v_cmp_lt_f32 0x40a00000, %a
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, bld.sop2(aco_opcode::s_and_b64, bld.def(bld.lm), bld.def(s1, scc),
|
|
bld.vopc(aco_opcode::v_cmp_eq_f32, bld.def(bld.lm), inputs[0], inputs[0]),
|
|
bld.vopc(aco_opcode::v_cmp_nge_f32, bld.def(bld.lm),
|
|
bld.copy(bld.def(v1), Operand::c32(0x40a00000u)), inputs[0])));
|
|
|
|
/* similar but unoptimizable expressions */
|
|
//! s2: %tmp6_0 = v_cmp_lt_f32 4.0, %a
|
|
//! s2: %tmp6_1 = v_cmp_neq_f32 %a, %a
|
|
//! s2: %res6, s1: %_:scc = s_and_b64 %tmp6_1, %tmp6_0
|
|
//! p_unit_test 6, %res6
|
|
Temp src1 =
|
|
bld.vopc(aco_opcode::v_cmp_lt_f32, bld.def(bld.lm), Operand::c32(0x40800000u), inputs[0]);
|
|
Temp src0 = bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), inputs[0], inputs[0]);
|
|
writeout(6, bld.sop2(aco_opcode::s_and_b64, bld.def(bld.lm), bld.def(s1, scc), src0, src1));
|
|
|
|
//! s2: %tmp7_0 = v_cmp_nge_f32 4.0, %a
|
|
//! s2: %tmp7_1 = v_cmp_eq_f32 %a, %a
|
|
//! s2: %res7, s1: %_:scc = s_or_b64 %tmp7_1, %tmp7_0
|
|
//! p_unit_test 7, %res7
|
|
src1 =
|
|
bld.vopc(aco_opcode::v_cmp_nge_f32, bld.def(bld.lm), Operand::c32(0x40800000u), inputs[0]);
|
|
src0 = bld.vopc(aco_opcode::v_cmp_eq_f32, bld.def(bld.lm), inputs[0], inputs[0]);
|
|
writeout(7, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc), src0, src1));
|
|
|
|
//! s2: %tmp8_0 = v_cmp_lt_f32 4.0, %d
|
|
//! s2: %tmp8_1 = v_cmp_neq_f32 %a, %a
|
|
//! s2: %res8, s1: %_:scc = s_or_b64 %tmp8_1, %tmp8_0
|
|
//! p_unit_test 8, %res8
|
|
src1 = bld.vopc(aco_opcode::v_cmp_lt_f32, bld.def(bld.lm), Operand::c32(0x40800000u), inputs[3]);
|
|
src0 = bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), inputs[0], inputs[0]);
|
|
writeout(8, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc), src0, src1));
|
|
|
|
//! s2: %tmp9_0 = v_cmp_lt_f32 4.0, %a
|
|
//! s2: %tmp9_1 = v_cmp_neq_f32 %a, %d
|
|
//! s2: %res9, s1: %_:scc = s_or_b64 %tmp9_1, %tmp9_0
|
|
//! p_unit_test 9, %res9
|
|
src1 = bld.vopc(aco_opcode::v_cmp_lt_f32, bld.def(bld.lm), Operand::c32(0x40800000u), inputs[0]);
|
|
src0 = bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), inputs[0], inputs[3]);
|
|
writeout(9, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc), src0, src1));
|
|
|
|
/* bit sizes */
|
|
//! s2: %res10 = v_cmp_nge_f16 4.0, %b
|
|
//! p_unit_test 10, %res10
|
|
Temp input1_16 =
|
|
bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), inputs[1], Operand::zero());
|
|
writeout(10, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc),
|
|
bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), input1_16, input1_16),
|
|
bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), Operand::c16(0x4400u),
|
|
input1_16)));
|
|
|
|
//! s2: %res11 = v_cmp_nge_f64 4.0, %c
|
|
//! p_unit_test 11, %res11
|
|
writeout(11, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc),
|
|
bld.vopc(aco_opcode::v_cmp_neq_f64, bld.def(bld.lm), inputs[2], inputs[2]),
|
|
bld.vopc(aco_opcode::v_cmp_lt_f64, bld.def(bld.lm),
|
|
Operand::c64(0x4010000000000000u), inputs[2])));
|
|
|
|
/* NaN */
|
|
uint16_t nan16 = 0x7e00;
|
|
uint32_t nan32 = 0x7fc00000;
|
|
uint64_t nan64 = 0xffffffffffffffffllu;
|
|
|
|
//! s2: %tmp12_0 = v_cmp_lt_f16 0x7e00, %a
|
|
//! s2: %tmp12_1 = v_cmp_neq_f16 %a, %a
|
|
//! s2: %res12, s1: %_:scc = s_or_b64 %tmp12_1, %tmp12_0
|
|
//! p_unit_test 12, %res12
|
|
src1 = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), Operand::c16(nan16), inputs[0]);
|
|
src0 = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), inputs[0], inputs[0]);
|
|
writeout(12, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc), src0, src1));
|
|
|
|
//! s2: %tmp13_0 = v_cmp_lt_f32 0x7fc00000, %a
|
|
//! s2: %tmp13_1 = v_cmp_neq_f32 %a, %a
|
|
//! s2: %res13, s1: %_:scc = s_or_b64 %tmp13_1, %tmp13_0
|
|
//! p_unit_test 13, %res13
|
|
src1 = bld.vopc(aco_opcode::v_cmp_lt_f32, bld.def(bld.lm), Operand::c32(nan32), inputs[0]);
|
|
src0 = bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), inputs[0], inputs[0]);
|
|
writeout(13, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc), src0, src1));
|
|
|
|
//! s2: %tmp14_0 = v_cmp_lt_f64 -1, %a
|
|
//! s2: %tmp14_1 = v_cmp_neq_f64 %a, %a
|
|
//! s2: %res14, s1: %_:scc = s_or_b64 %tmp14_1, %tmp14_0
|
|
//! p_unit_test 14, %res14
|
|
src1 = bld.vopc(aco_opcode::v_cmp_lt_f64, bld.def(bld.lm), Operand::c64(nan64), inputs[0]);
|
|
src0 = bld.vopc(aco_opcode::v_cmp_neq_f64, bld.def(bld.lm), inputs[0], inputs[0]);
|
|
writeout(14, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc), src0, src1));
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.add3)
|
|
//>> v1: %a, v1: %b, v1: %c = p_startpgm
|
|
if (!setup_cs("v1 v1 v1", GFX9))
|
|
return;
|
|
|
|
//! v1: %res0 = v_add3_u32 %a, %b, %c
|
|
//! p_unit_test 0, %res0
|
|
Builder::Result tmp = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]);
|
|
writeout(0, bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp));
|
|
|
|
//! v1: %tmp1 = v_add_u32 %b, %c clamp
|
|
//! v1: %res1 = v_add_u32 %a, %tmp1
|
|
//! p_unit_test 1, %res1
|
|
tmp = bld.vop2_e64(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]);
|
|
tmp->valu().clamp = true;
|
|
writeout(1, bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp));
|
|
|
|
//! v1: %tmp2 = v_add_u32 %b, %c
|
|
//! v1: %res2 = v_add_u32 %a, %tmp2 clamp
|
|
//! p_unit_test 2, %res2
|
|
tmp = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]);
|
|
tmp = bld.vop2_e64(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp);
|
|
tmp->valu().clamp = true;
|
|
writeout(2, tmp);
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.minmax)
|
|
for (unsigned i = GFX10_3; i <= GFX11; i++) {
|
|
//>> v1: %a, v1: %b, v1: %c = p_startpgm
|
|
if (!setup_cs("v1 v1 v1", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
Temp c = inputs[2];
|
|
|
|
//! v1: %res0 = v_min3_f32 %a, %b, %c
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fmin(c, fmin(a, b)));
|
|
|
|
//! v1: %res1 = v_max3_f32 %a, %b, %c
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fmax(c, fmax(a, b)));
|
|
|
|
//! v1: %res2 = v_min3_f32 -%a, -%b, %c
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, fmin(c, fneg(fmax(a, b))));
|
|
|
|
//! v1: %res3 = v_max3_f32 -%a, -%b, %c
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, fmax(c, fneg(fmin(a, b))));
|
|
|
|
//! v1: %res4 = v_max3_f32 -%a, %b, %c
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, fmax(c, fneg(fmin(a, fneg(b)))));
|
|
|
|
//~gfx10_3! v1: %res5_tmp = v_max_f32 %a, %b
|
|
//~gfx10_3! v1: %res5 = v_min_f32 %c, %res5_tmp
|
|
//~gfx11! v1: %res5 = v_maxmin_f32 %a, %b, %c
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, fmin(c, fmax(a, b)));
|
|
|
|
//~gfx10_3! v1: %res6_tmp = v_min_f32 %a, %b
|
|
//~gfx10_3! v1: %res6 = v_max_f32 %c, %res6_tmp
|
|
//~gfx11! v1: %res6 = v_minmax_f32 %a, %b, %c
|
|
//! p_unit_test 6, %res6
|
|
writeout(6, fmax(c, fmin(a, b)));
|
|
|
|
//~gfx10_3! v1: %res7_tmp = v_min_f32 %a, %b
|
|
//~gfx10_3! v1: %res7 = v_min_f32 %c, -%res7_tmp
|
|
//~gfx11! v1: %res7 = v_maxmin_f32 -%a, -%b, %c
|
|
//! p_unit_test 7, %res7
|
|
writeout(7, fmin(c, fneg(fmin(a, b))));
|
|
|
|
//~gfx10_3! v1: %res8_tmp = v_max_f32 %a, %b
|
|
//~gfx10_3! v1: %res8 = v_max_f32 %c, -%res8_tmp
|
|
//~gfx11! v1: %res8 = v_minmax_f32 -%a, -%b, %c
|
|
//! p_unit_test 8, %res8
|
|
writeout(8, fmax(c, fneg(fmax(a, b))));
|
|
|
|
//~gfx10_3! v1: %res9_tmp = v_max_f32 %a, -%b
|
|
//~gfx10_3! v1: %res9 = v_max_f32 %c, -%res9_tmp
|
|
//~gfx11! v1: %res9 = v_minmax_f32 -%a, %b, %c
|
|
//! p_unit_test 9, %res9
|
|
writeout(9, fmax(c, fneg(fmax(a, fneg(b)))));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_32_24)
|
|
for (unsigned i = GFX8; i <= GFX9; i++) {
|
|
//>> v1: %a, v1: %b, v1: %c = p_startpgm
|
|
if (!setup_cs("v1 v1 v1", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
//! v1: %res0 = v_mad_u32_u24 %b, %c, %a
|
|
//! p_unit_test 0, %res0
|
|
Temp mul = bld.vop2(aco_opcode::v_mul_u32_u24, bld.def(v1), inputs[1], inputs[2]);
|
|
writeout(0, bld.vadd32(bld.def(v1), inputs[0], mul));
|
|
|
|
//! v1: %res1_tmp = v_mul_u32_u24 %b, %c
|
|
//! v1: %_, s2: %res1 = v_add_co_u32 %a, %res1_tmp
|
|
//! p_unit_test 1, %res1
|
|
mul = bld.vop2(aco_opcode::v_mul_u32_u24, bld.def(v1), inputs[1], inputs[2]);
|
|
writeout(1, bld.vadd32(bld.def(v1), inputs[0], mul, true).def(1).getTemp());
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.add_lshlrev)
|
|
for (unsigned i = GFX8; i <= GFX10; i++) {
|
|
//>> v1: %a, v1: %b, s1: %c = p_startpgm
|
|
if (!setup_cs("v1 v1 s1", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
Temp lshl;
|
|
|
|
//~gfx8! v1: %lshl0 = v_lshlrev_b32 3, %a
|
|
//~gfx8! v1: %res0, s2: %_ = v_add_co_u32 %lshl0, %b
|
|
//~gfx(9|10)! v1: %res0 = v_lshl_add_u32 %a, 3, %b
|
|
//! p_unit_test 0, %res0
|
|
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(3u), Operand(inputs[0]));
|
|
writeout(0, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
//~gfx8! v1: %lshl1 = v_lshlrev_b32 7, (is24bit)%a
|
|
//~gfx8! v1: %res1, s2: %_ = v_add_co_u32 %lshl1, %b
|
|
//~gfx(9|10)! v1: %res1 = v_lshl_add_u32 (is24bit)%a, 7, %b
|
|
//! p_unit_test 1, %res1
|
|
Operand a_24bit = Operand(inputs[0]);
|
|
a_24bit.set24bit(true);
|
|
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(7u), a_24bit);
|
|
writeout(1, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
//~gfx8! v1: %lshl2 = v_lshlrev_b32 (is24bit)%a, (is24bit)%b
|
|
//~gfx8! v1: %res2, s2: %_ = v_add_co_u32 %lshl2, %b
|
|
//~gfx(9|10)! v1: %res2 = v_lshl_add_u32 (is24bit)%b, (is24bit)%a, %b
|
|
//! p_unit_test 2, %res2
|
|
Operand b_24bit = Operand(inputs[1]);
|
|
b_24bit.set24bit(true);
|
|
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), a_24bit, b_24bit);
|
|
writeout(2, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
//~gfx8! v1: %res3 = v_mad_u32_u24 (is24bit)%a, 8, %b
|
|
//~gfx(9|10)! v1: %res3 = v_lshl_add_u32 (is24bit)%a, 3, %b
|
|
//! p_unit_test 3, %res3
|
|
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(3u), a_24bit);
|
|
writeout(3, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
//~gfx8! v1: %res4 = v_mad_u32_u24 (is16bit)%a, 16, %b
|
|
//~gfx(9|10)! v1: %res4 = v_lshl_add_u32 (is16bit)%a, 4, %b
|
|
//! p_unit_test 4, %res4
|
|
Operand a_16bit = Operand(inputs[0]);
|
|
a_16bit.set16bit(true);
|
|
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(4u), a_16bit);
|
|
writeout(4, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
|
|
|
|
//~gfx8! v1: %res5 = v_mad_u32_u24 (is24bit)%c, 16, %c
|
|
//~gfx(9|10)! v1: %res5 = v_lshl_add_u32 (is24bit)%c, 4, %c
|
|
//! p_unit_test 5, %res5
|
|
Operand c_24bit = Operand(inputs[2]);
|
|
c_24bit.set24bit(true);
|
|
lshl = bld.vop2_e64(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(4u), c_24bit);
|
|
writeout(5, bld.vadd32(bld.def(v1), lshl, Operand(inputs[2])));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
enum denorm_op {
|
|
denorm_mul1 = 0,
|
|
denorm_fneg = 1,
|
|
denorm_fabs = 2,
|
|
denorm_fnegabs = 3,
|
|
};
|
|
|
|
static const char* denorm_op_names[] = {
|
|
"mul1",
|
|
"fneg",
|
|
"fabs",
|
|
"fnegabs",
|
|
};
|
|
|
|
struct denorm_config {
|
|
bool flush;
|
|
unsigned op;
|
|
aco_opcode src;
|
|
aco_opcode dest;
|
|
};
|
|
|
|
static const char*
|
|
srcdest_op_name(aco_opcode op)
|
|
{
|
|
switch (op) {
|
|
case aco_opcode::v_cndmask_b32: return "cndmask";
|
|
case aco_opcode::v_min_f32: return "min";
|
|
case aco_opcode::v_rcp_f32: return "rcp";
|
|
default: return "none";
|
|
}
|
|
}
|
|
|
|
static Temp
|
|
emit_denorm_srcdest(aco_opcode op, Temp val)
|
|
{
|
|
switch (op) {
|
|
case aco_opcode::v_cndmask_b32:
|
|
return bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), Operand::zero(), val, inputs[1]);
|
|
case aco_opcode::v_min_f32:
|
|
return bld.vop2(aco_opcode::v_min_f32, bld.def(v1), Operand::zero(), val);
|
|
case aco_opcode::v_rcp_f32: return bld.vop1(aco_opcode::v_rcp_f32, bld.def(v1), val);
|
|
default: return val;
|
|
}
|
|
}
|
|
|
|
BEGIN_TEST(optimize.denorm_propagation)
|
|
for (unsigned i = GFX8; i <= GFX9; i++) {
|
|
std::vector<denorm_config> configs;
|
|
for (bool flush : {false, true}) {
|
|
for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs})
|
|
configs.push_back({flush, op, aco_opcode::num_opcodes, aco_opcode::num_opcodes});
|
|
|
|
for (aco_opcode dest : {aco_opcode::v_min_f32, aco_opcode::v_rcp_f32}) {
|
|
for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs})
|
|
configs.push_back({flush, op, aco_opcode::num_opcodes, dest});
|
|
}
|
|
|
|
for (aco_opcode src :
|
|
{aco_opcode::v_cndmask_b32, aco_opcode::v_min_f32, aco_opcode::v_rcp_f32}) {
|
|
for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs})
|
|
configs.push_back({flush, op, src, aco_opcode::num_opcodes});
|
|
}
|
|
}
|
|
|
|
for (denorm_config cfg : configs) {
|
|
char subvariant[128];
|
|
sprintf(subvariant, "_%s_%s_%s_%s", cfg.flush ? "flush" : "keep", srcdest_op_name(cfg.src),
|
|
denorm_op_names[(int)cfg.op], srcdest_op_name(cfg.dest));
|
|
if (!setup_cs("v1 s2", (amd_gfx_level)i, CHIP_UNKNOWN, subvariant))
|
|
continue;
|
|
|
|
bool can_propagate = cfg.src == aco_opcode::v_rcp_f32 ||
|
|
(i >= GFX9 && cfg.src == aco_opcode::v_min_f32) ||
|
|
cfg.dest == aco_opcode::v_rcp_f32 ||
|
|
(i >= GFX9 && cfg.dest == aco_opcode::v_min_f32) || !cfg.flush;
|
|
|
|
fprintf(output, "src, dest, op: %s %s %s\n", srcdest_op_name(cfg.src),
|
|
srcdest_op_name(cfg.dest), denorm_op_names[(int)cfg.op]);
|
|
fprintf(output, "can_propagate: %u\n", can_propagate);
|
|
//! src, dest, op: $src $dest $op
|
|
//! can_propagate: #can_propagate
|
|
//>> v1: %a, s2: %b = p_startpgm
|
|
|
|
//; patterns = {'cndmask': 'v1: %{} = v_cndmask_b32 0, {}, %b',
|
|
//; 'min': 'v1: %{} = v_min_f32 0, {}',
|
|
//; 'rcp': 'v1: %{} = v_rcp_f32 {}'}
|
|
//; ops = {'mul1': 'v1: %{} = v_mul_f32 1.0, %{}',
|
|
//; 'fneg': 'v1: %{} = v_mul_f32 -1.0, %{}',
|
|
//; 'fabs': 'v1: %{} = v_mul_f32 1.0, |%{}|',
|
|
//; 'fnegabs': 'v1: %{} = v_mul_f32 -1.0, |%{}|'}
|
|
//; inline_ops = {'mul1': '%{}', 'fneg': '-%{}', 'fabs': '|%{}|', 'fnegabs': '-|%{}|'}
|
|
|
|
//; name = 'a'
|
|
//; if src != 'none':
|
|
//; insert_pattern(patterns[src].format('src_res', '%'+name))
|
|
//; name = 'src_res'
|
|
|
|
//; if can_propagate:
|
|
//; name = inline_ops[op].format(name)
|
|
//; else:
|
|
//; insert_pattern(ops[op].format('op_res', name))
|
|
//; name = '%op_res'
|
|
|
|
//; if dest != 'none':
|
|
//; insert_pattern(patterns[dest].format('dest_res', name))
|
|
//; name = '%dest_res'
|
|
|
|
//; insert_pattern('v1: %res = v_cndmask_b32 0, {}, %b'.format(name))
|
|
//! p_unit_test 0, %res
|
|
|
|
program->blocks[0].fp_mode.denorm32 = cfg.flush ? fp_denorm_flush : fp_denorm_keep;
|
|
|
|
Temp val = emit_denorm_srcdest(cfg.src, inputs[0]);
|
|
switch (cfg.op) {
|
|
case denorm_mul1:
|
|
val = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x3f800000u), val);
|
|
break;
|
|
case denorm_fneg: val = fneg(val); break;
|
|
case denorm_fabs: val = fabs(val); break;
|
|
case denorm_fnegabs: val = fneg(fabs(val)); break;
|
|
}
|
|
val = emit_denorm_srcdest(cfg.dest, val);
|
|
writeout(
|
|
0, bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), Operand::zero(), val, inputs[1]));
|
|
|
|
finish_opt_test();
|
|
}
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimizer.dpp)
|
|
//>> v1: %a, v1: %b, s2: %c, s1: %d = p_startpgm
|
|
if (!setup_cs("v1 v1 s2 s1", GFX10_3))
|
|
return;
|
|
|
|
Operand a(inputs[0]);
|
|
Operand b(inputs[1]);
|
|
Operand c(inputs[2]);
|
|
Operand d(inputs[3]);
|
|
|
|
/* basic optimization */
|
|
//! v1: %res0 = v_add_f32 %a, %b row_mirror bound_ctrl:1 fi
|
|
//! p_unit_test 0, %res0
|
|
Temp tmp0 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
Temp res0 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), tmp0, b);
|
|
writeout(0, res0);
|
|
|
|
/* operand swapping */
|
|
//! v1: %res1 = v_subrev_f32 %a, %b row_mirror bound_ctrl:1 fi
|
|
//! p_unit_test 1, %res1
|
|
Temp tmp1 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
Temp res1 = bld.vop2(aco_opcode::v_sub_f32, bld.def(v1), b, tmp1);
|
|
writeout(1, res1);
|
|
|
|
//! v1: %tmp2 = v_mov_b32 %a row_mirror bound_ctrl:1 fi
|
|
//! v1: %res2 = v_sub_f32 %b, %tmp2 row_half_mirror bound_ctrl:1 fi
|
|
//! p_unit_test 2, %res2
|
|
Temp tmp2 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
Temp res2 = bld.vop2_dpp(aco_opcode::v_sub_f32, bld.def(v1), b, tmp2, dpp_row_half_mirror);
|
|
writeout(2, res2);
|
|
|
|
/* modifiers */
|
|
//! v1: %res3 = v_add_f32 -%a, %b row_mirror bound_ctrl:1 fi
|
|
//! p_unit_test 3, %res3
|
|
auto tmp3 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
tmp3->dpp16().neg[0] = true;
|
|
Temp res3 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), tmp3, b);
|
|
writeout(3, res3);
|
|
|
|
//! v1: %res4 = v_add_f32 -%a, %b row_mirror bound_ctrl:1 fi
|
|
//! p_unit_test 4, %res4
|
|
Temp tmp4 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
auto res4 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp4, b);
|
|
res4->valu().neg[0] = true;
|
|
writeout(4, res4);
|
|
|
|
//! v1: %tmp5 = v_mov_b32 %a row_mirror bound_ctrl:1 fi
|
|
//! v1: %res5 = v_add_f32 %tmp5, %b clamp
|
|
//! p_unit_test 5, %res5
|
|
Temp tmp5 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
auto res5 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp5, b);
|
|
res5->valu().clamp = true;
|
|
writeout(5, res5);
|
|
|
|
//! v1: %res6 = v_add_f32 |%a|, %b row_mirror bound_ctrl:1 fi
|
|
//! p_unit_test 6, %res6
|
|
auto tmp6 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
tmp6->dpp16().neg[0] = true;
|
|
auto res6 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp6, b);
|
|
res6->valu().abs[0] = true;
|
|
writeout(6, res6);
|
|
|
|
//! v1: %res7 = v_subrev_f32 %a, |%b| row_mirror bound_ctrl:1 fi
|
|
//! p_unit_test 7, %res7
|
|
Temp tmp7 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
auto res7 = bld.vop2_e64(aco_opcode::v_sub_f32, bld.def(v1), b, tmp7);
|
|
res7->valu().abs[0] = true;
|
|
writeout(7, res7);
|
|
|
|
//! v1: %tmp11 = v_mov_b32 -%a row_mirror bound_ctrl:1 fi
|
|
//! v1: %res11 = v_add_u32 %tmp11, %b
|
|
//! p_unit_test 11, %res11
|
|
auto tmp11 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
tmp11->dpp16().neg[0] = true;
|
|
Temp res11 = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), tmp11, b);
|
|
writeout(11, res11);
|
|
|
|
//! v1: %tmp12 = v_mov_b32 -%a row_mirror bound_ctrl:1 fi
|
|
//! v1: %res12 = v_add_f16 %tmp12, %b
|
|
//! p_unit_test 12, %res12
|
|
auto tmp12 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
tmp12->dpp16().neg[0] = true;
|
|
Temp res12 = bld.vop2(aco_opcode::v_add_f16, bld.def(v1), tmp12, b);
|
|
writeout(12, res12);
|
|
|
|
/* vcc */
|
|
//! v1: %res8 = v_cndmask_b32 %a, %b, %c:vcc row_mirror bound_ctrl:1 fi
|
|
//! p_unit_test 8, %res8
|
|
Temp tmp8 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
Temp res8 = bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), tmp8, b, c);
|
|
writeout(8, res8);
|
|
|
|
/* sgprs */
|
|
//! v1: %tmp9 = v_mov_b32 %a row_mirror bound_ctrl:1 fi
|
|
//! v1: %res9 = v_add_f32 %tmp9, %d
|
|
//! p_unit_test 9, %res9
|
|
Temp tmp9 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
Temp res9 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp9, d);
|
|
writeout(9, res9);
|
|
|
|
//! v1: %tmp10 = v_mov_b32 %a row_mirror bound_ctrl:1 fi
|
|
//! v1: %res10 = v_add_f32 %d, %tmp10
|
|
//! p_unit_test 10, %res10
|
|
Temp tmp10 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
Temp res10 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), d, tmp10);
|
|
writeout(10, res10);
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.dpp_prop)
|
|
//>> v1: %a, s1: %b = p_startpgm
|
|
if (!setup_cs("v1 s1", GFX10))
|
|
return;
|
|
|
|
//! v1: %one = p_parallelcopy 1
|
|
//! v1: %res0 = v_mul_f32 1, %a
|
|
//! p_unit_test 0, %res0
|
|
Temp one = bld.copy(bld.def(v1), Operand::c32(1));
|
|
writeout(0, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), one, inputs[0], dpp_row_sl(1)));
|
|
|
|
//! v1: %res1 = v_mul_f32 %a, %one row_shl:1 bound_ctrl:1 fi
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], one, dpp_row_sl(1)));
|
|
|
|
//! v1: %res2 = v_mul_f32 0x12345678, %a
|
|
//! p_unit_test 2, %res2
|
|
Temp literal1 = bld.copy(bld.def(v1), Operand::c32(0x12345678u));
|
|
writeout(2,
|
|
bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), literal1, inputs[0], dpp_row_sl(1)));
|
|
|
|
//! v1: %literal2 = p_parallelcopy 0x12345679
|
|
//! v1: %res3 = v_mul_f32 %a, %literal row_shl:1 bound_ctrl:1 fi
|
|
//! p_unit_test 3, %res3
|
|
Temp literal2 = bld.copy(bld.def(v1), Operand::c32(0x12345679u));
|
|
writeout(3,
|
|
bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], literal2, dpp_row_sl(1)));
|
|
|
|
//! v1: %b_v = p_parallelcopy %b
|
|
//! v1: %res4 = v_mul_f32 %b, %a
|
|
//! p_unit_test 4, %res4
|
|
Temp b_v = bld.copy(bld.def(v1), inputs[1]);
|
|
writeout(4, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), b_v, inputs[0], dpp_row_sl(1)));
|
|
|
|
//! v1: %res5 = v_mul_f32 %a, %b_v row_shl:1 bound_ctrl:1 fi
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], b_v, dpp_row_sl(1)));
|
|
|
|
//! v1: %res6 = v_rcp_f32 %b
|
|
//! p_unit_test 6, %res6
|
|
writeout(6, bld.vop1_dpp(aco_opcode::v_rcp_f32, bld.def(v1), b_v, dpp_row_sl(1)));
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.casts)
|
|
//>> v1: %a, v2b: %a16 = p_startpgm
|
|
if (!setup_cs("v1 v2b", GFX10_3))
|
|
return;
|
|
|
|
Temp a = inputs[0];
|
|
Temp a16 = inputs[1];
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
//! v1: %res0_tmp = v_mul_f32 -1.0, %a
|
|
//! v2b: %res0 = v_mul_f16 %res0_tmp, %a16
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fmul(u2u16(fneg(a)), a16));
|
|
|
|
//! v2b: %res1_tmp = v_mul_f16 -1.0, %a16
|
|
//! v1: %res1 = v_mul_f32 %res1_tmp, %a
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fmul(bld.as_uniform(fneg(a16)), a));
|
|
|
|
//! v1: %res2_tmp = v_mul_f32 -1.0, %a16
|
|
//! v2b: %res2 = v_mul_f16 %res2_tmp, %a16
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, fmul(u2u16(bld.vop2_e64(aco_opcode::v_mul_f32, bld.def(v1),
|
|
Operand::c32(0xbf800000u), bld.as_uniform(a16))),
|
|
a16));
|
|
|
|
//! v1: %res3_tmp = v_mul_f32 %a, %a
|
|
//! v2b: %res3 = v_add_f16 %res3_tmp, 0 clamp
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, fsat(u2u16(fmul(a, a))));
|
|
|
|
//! v2b: %res4_tmp = v_mul_f16 %a16, %a16
|
|
//! v1: %res4 = v_add_f32 %res4_tmp, 0 clamp
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, fsat(bld.as_uniform(fmul(a16, a16))));
|
|
|
|
//! v1: %res5_tmp = v_mul_f32 %a, %a
|
|
//! v2b: %res5 = v_mul_f16 2.0, %res5_tmp
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, fmul(u2u16(fmul(a, a)), bld.copy(bld.def(v2b), Operand::c16(0x4000))));
|
|
|
|
//! v2b: %res6_tmp = v_mul_f16 %a16, %a16
|
|
//! v1: %res6 = v_mul_f32 2.0, %res6_tmp
|
|
//! p_unit_test 6, %res6
|
|
writeout(6,
|
|
fmul(bld.as_uniform(fmul(a16, a16)), bld.copy(bld.def(v1), Operand::c32(0x40000000))));
|
|
|
|
//! v1: %res7_tmp = v_mul_f32 %a, %a
|
|
//! v2b: %res7 = v_add_f16 %res7_tmp, %a16
|
|
//! p_unit_test 7, %res7
|
|
writeout(7, fadd(u2u16(fmul(a, a)), a16));
|
|
|
|
//! v2b: %res8_tmp = v_mul_f16 %a16, %a16
|
|
//! v1: %res8 = v_add_f32 %res8_tmp, %a
|
|
//! p_unit_test 8, %res8
|
|
writeout(8, fadd(bld.as_uniform(fmul(a16, a16)), a));
|
|
|
|
//! v1: %res9_tmp = v_mul_f32 %a, %a
|
|
//! v2b: %res9 = v_mul_f16 -1.0, %res9_tmp
|
|
//! p_unit_test 9, %res9
|
|
writeout(9, fneg(u2u16(fmul(a, a))));
|
|
|
|
//! v2b: %res10_tmp = v_mul_f16 %a16, %a16
|
|
//! v1: %res10 = v_mul_f32 -1.0, %res10_tmp
|
|
//! p_unit_test 10, %res10
|
|
writeout(10, bld.vop2_e64(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0xbf800000u),
|
|
bld.as_uniform(fmul(a16, a16))));
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.input_conv.basic)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, v2b: %a16 = p_startpgm
|
|
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a = inputs[0];
|
|
Temp a16 = inputs[1];
|
|
|
|
//! v1: %res0 = v_fma_mix_f32 %a, lo(%a16), -0
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fmul(a, f2f32(a16)));
|
|
|
|
//! v1: %res1 = v_fma_mix_f32 1.0, %a, lo(%a16)
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fadd(a, f2f32(a16)));
|
|
|
|
//! v1: %res2 = v_fma_mix_f32 1.0, lo(%a16), %a
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, fadd(f2f32(a16), a));
|
|
|
|
//! v1: %res3 = v_fma_mix_f32 %a, %a, lo(%a16)
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, fma(a, a, f2f32(a16)));
|
|
|
|
//! v1: %res4 = v_fma_mix_f32 %a, %a, lo(%a16)
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, fma(a, a, f2f32(a16)));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.input_conv.precision)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, v2b: %a16 = p_startpgm
|
|
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a = inputs[0];
|
|
Temp a16 = inputs[1];
|
|
|
|
/* precise arithmetic */
|
|
//~gfx9! v1: %res0_cvt = v_cvt_f32_f16 %a16
|
|
//~gfx9! v1: (precise)%res0 = v_fma_f32 %a, %a, %res0_cvt
|
|
//~gfx10! v1: (precise)%res0 = v_fma_mix_f32 %a, %a, lo(%a16)
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fma(a, a, f2f32(a16), bld.precise()));
|
|
|
|
//! v2b: %res1_cvt = v_cvt_f16_f32 %a
|
|
//! v2b: (precise)%res1 = v_mul_f16 %a16, %res1_cvt
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fmul(a16, f2f16(a), bld.precise()));
|
|
|
|
//! v2b: %res2_cvt = v_cvt_f16_f32 %a
|
|
//! v2b: (precise)%res2 = v_add_f16 %a16, %res2_cvt
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, fadd(a16, f2f16(a), bld.precise()));
|
|
|
|
//! v2b: %res3_cvt = v_cvt_f16_f32 %a
|
|
//! v2b: (precise)%res3 = v_fma_f16 %a16, %a16, %res3_cvt
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, fma(a16, a16, f2f16(a), bld.precise()));
|
|
|
|
/* precise conversions */
|
|
//! v2b: (precise)%res4_cvt = v_cvt_f16_f32 %a
|
|
//! v2b: %res4 = v_mul_f16 %a16, %res4_cvt
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, fmul(a16, f2f16(a, bld.precise())));
|
|
|
|
//! v2b: (precise)%res5_cvt = v_cvt_f16_f32 %a
|
|
//! v2b: %res5 = v_add_f16 %a16, %res5_cvt
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, fadd(a16, f2f16(a, bld.precise())));
|
|
|
|
//! v2b: (precise)%res6_cvt = v_cvt_f16_f32 %a
|
|
//! v2b: %res6 = v_fma_f16 %a16, %a16, %res6_cvt
|
|
//! p_unit_test 6, %res6
|
|
writeout(6, fma(a16, a16, f2f16(a, bld.precise())));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.input_conv.modifiers)
|
|
for (unsigned i = GFX9; i <= GFX11; i++) {
|
|
if (i == GFX10_3)
|
|
continue;
|
|
//>> v1: %a, v2b: %a16 = p_startpgm
|
|
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a = inputs[0];
|
|
Temp a16 = inputs[1];
|
|
|
|
/* check whether modifiers are preserved when converting to VOP3P */
|
|
//! v1: %res0 = v_fma_mix_f32 -%a, lo(%a16), -0
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fmul(fneg(a), f2f32(a16)));
|
|
|
|
//! v1: %res1 = v_fma_mix_f32 |%a|, lo(%a16), -0
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fmul(fabs(a), f2f32(a16)));
|
|
|
|
/* fneg modifiers */
|
|
//! v1: %res2 = v_fma_mix_f32 %a, -lo(%a16), -0
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, fmul(a, fneg(f2f32(a16))));
|
|
|
|
//! v1: %res3 = v_fma_mix_f32 %a, -lo(%a16), -0
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, fmul(a, f2f32(fneg(a16))));
|
|
|
|
/* fabs modifiers */
|
|
//! v1: %res4 = v_fma_mix_f32 %a, |lo(%a16)|, -0
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, fmul(a, fabs(f2f32(a16))));
|
|
|
|
//! v1: %res5 = v_fma_mix_f32 %a, |lo(%a16)|, -0
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, fmul(a, f2f32(fabs(a16))));
|
|
|
|
/* both fabs and fneg modifiers */
|
|
//! v1: %res6 = v_fma_mix_f32 %a, -|lo(%a16)|, -0
|
|
//! p_unit_test 6, %res6
|
|
writeout(6, fmul(a, fneg(f2f32(fabs(a16)))));
|
|
|
|
//! v1: %res7 = v_fma_mix_f32 %a, |lo(%a16)|, -0
|
|
//! p_unit_test 7, %res7
|
|
writeout(7, fmul(a, fabs(f2f32(fabs(a16)))));
|
|
|
|
//! v1: %res8 = v_fma_mix_f32 %a, -|lo(%a16)|, -0
|
|
//! p_unit_test 8, %res8
|
|
writeout(8, fmul(a, fneg(fabs(f2f32(fabs(a16))))));
|
|
|
|
//! v1: %res9 = v_fma_mix_f32 %a, -|lo(%a16)|, -0
|
|
//! p_unit_test 9, %res9
|
|
writeout(9, fmul(a, f2f32(fneg(fabs(a16)))));
|
|
|
|
//! v1: %res10 = v_fma_mix_f32 %a, |lo(%a16)|, -0
|
|
//! p_unit_test 10, %res10
|
|
writeout(10, fmul(a, fneg(f2f32(fneg(fabs(a16))))));
|
|
|
|
//! v1: %res11 = v_fma_mix_f32 %a, |lo(%a16)|, -0
|
|
//! p_unit_test 11, %res11
|
|
writeout(11, fmul(a, fabs(f2f32(fneg(fabs(a16))))));
|
|
|
|
//! v1: %res12 = v_fma_mix_f32 %a, -|lo(%a16)|, -0
|
|
//! p_unit_test 12, %res12
|
|
writeout(12, fmul(a, fneg(fabs(f2f32(fneg(fabs(a16)))))));
|
|
|
|
/* sdwa */
|
|
//! v1: %res13 = v_fma_mix_f32 lo(%a), %a, -0
|
|
//! p_unit_test 13, %res13
|
|
writeout(13, fmul(f2f32(ext_ushort(a, 0)), a));
|
|
|
|
//! v1: %res14 = v_fma_mix_f32 hi(%a), %a, -0
|
|
//! p_unit_test 14, %res14
|
|
writeout(14, fmul(f2f32(ext_ushort(a, 1)), a));
|
|
|
|
//~gfx(9|10)! v1: %res15_cvt = v_cvt_f32_f16 %a dst_sel:uword0 src0_sel:dword
|
|
//~gfx11! v1: %res16_cvt1 = v_fma_mix_f32 lo(%a), 1.0, -0
|
|
//~gfx11! v1: %res15_cvt = p_extract %res16_cvt1, 0, 16, 0
|
|
//! v1: %res15 = v_mul_f32 %res15_cvt, %a
|
|
//! p_unit_test 15, %res15
|
|
writeout(15, fmul(ext_ushort(f2f32(a), 0), a));
|
|
|
|
//~gfx(9|10)! v1: %res16_cvt = v_cvt_f32_f16 %a
|
|
//~gfx(9|10)! v1: %res16 = v_mul_f32 %res16_cvt, %a dst_sel:dword src0_sel:uword1 src1_sel:dword
|
|
//~gfx11! v1: %res16_cvt = v_fma_mix_f32 lo(%a), 1.0, -0
|
|
//~gfx11! v1: %res16_ext = p_extract %res16_cvt, 1, 16, 0
|
|
//~gfx11! v1: %res16 = v_mul_f32 %res16_ext, %a
|
|
//! p_unit_test 16, %res16
|
|
writeout(16, fmul(ext_ushort(f2f32(a), 1), a));
|
|
|
|
//~gfx(9|10)! v1: %res17_cvt = v_cvt_f32_f16 %a dst_sel:dword src0_sel:ubyte2
|
|
//~gfx(9|10)! v1: %res17 = v_mul_f32 %res17_cvt, %a
|
|
//~gfx11! v1: %res17_ext = p_extract %a, 2, 8, 0
|
|
//~gfx11! v1: %res17 = v_fma_mix_f32 lo(%res17_ext), %a, -0
|
|
//! p_unit_test 17, %res17
|
|
writeout(17, fmul(f2f32(ext_ubyte(a, 2)), a));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.output_conv.basic)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %b16 = p_startpgm
|
|
if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
Temp c = inputs[2];
|
|
Temp a16 = inputs[3];
|
|
Temp b16 = inputs[4];
|
|
|
|
//! v2b: %res0 = v_fma_mixlo_f16 %a, %b, -0
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, f2f16(fmul(a, b)));
|
|
|
|
//! v2b: %res1 = v_fma_mixlo_f16 1.0, %a, %b
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, f2f16(fadd(a, b)));
|
|
|
|
//! v2b: %res2 = v_fma_mixlo_f16 %a, %b, %c
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, f2f16(fma(a, b, c)));
|
|
|
|
//! v2b: %res3 = v_fma_mixlo_f16 lo(%a16), %b, -0
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, f2f16(fmul(f2f32(a16), b)));
|
|
|
|
//! v2b: %res4 = v_fma_mixlo_f16 1.0, %a, lo(%b16)
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, f2f16(fadd(a, f2f32(b16))));
|
|
|
|
//! v2b: %res5 = v_fma_mixlo_f16 %a, lo(%b16), %c
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, f2f16(fma(a, f2f32(b16), c)));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.output_conv.precision)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v2b: %a16 = p_startpgm
|
|
if (!setup_cs("v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a16 = inputs[0];
|
|
|
|
//! v2b: %res0_tmp = v_mul_f16 %a16, %a16
|
|
//! v1: (precise)%res0 = v_cvt_f32_f16 %res0_tmp
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, f2f32(fmul(a16, a16), bld.precise()));
|
|
|
|
//! v2b: (precise)%res1_tmp = v_mul_f16 %a16, %a16
|
|
//! v1: %res1 = v_cvt_f32_f16 %res1_tmp
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, f2f32(fmul(a16, a16, bld.precise())));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.output_conv.modifiers)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, v1: %b, v2b: %a16, v2b: %b16 = p_startpgm
|
|
if (!setup_cs("v1 v1 v2b v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
Temp a16 = inputs[2];
|
|
Temp b16 = inputs[3];
|
|
|
|
/* fneg/fabs */
|
|
//! v1: %res0_add = v_add_f32 %1, %2
|
|
//! v2b: %res0 = v_cvt_f16_f32 |%res0_add|
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, f2f16(fabs(fadd(a, b))));
|
|
|
|
//! v1: %res1_add = v_add_f32 %1, %2
|
|
//! v2b: %res1 = v_cvt_f16_f32 -%res1_add
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, f2f16(fneg(fadd(a, b))));
|
|
|
|
//! v2b: %res2_add = v_add_f16 %3, %4
|
|
//! v1: %res2 = v_cvt_f32_f16 |%res2_add|
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, f2f32(fabs(fadd(a16, b16))));
|
|
|
|
//! v2b: %res3_add = v_add_f16 %3, %4
|
|
//! v1: %res3 = v_cvt_f32_f16 -%res3_add
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, f2f32(fneg(fadd(a16, b16))));
|
|
|
|
/* sdwa */
|
|
//! v2b: %res4_add = v_fma_mixlo_f16 1.0, %a, %b
|
|
//! v2b: %res4 = p_extract %res4_add, 0, 8, 0
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, ext_ubyte(f2f16(fadd(a, b)), 0));
|
|
|
|
//! v1: %res5_mul = v_add_f32 %a, %b dst_sel:uword0 src0_sel:dword src1_sel:dword
|
|
//! v2b: %res5 = v_cvt_f16_f32 %res5_mul
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, f2f16(ext_ushort(fadd(a, b), 0)));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.fma.basic)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %c16 = p_startpgm
|
|
if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
Temp c = inputs[2];
|
|
Temp a16 = inputs[3];
|
|
Temp c16 = inputs[4];
|
|
|
|
//! v1: %res0 = v_fma_mix_f32 lo(%a16), %b, %c
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fadd(fmul(f2f32(a16), b), c));
|
|
|
|
//! v1: %res1 = v_fma_mix_f32 %a, %b, lo(%c16)
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fadd(fmul(a, b), f2f32(c16)));
|
|
|
|
/* omod/clamp check */
|
|
//! v1: %res2_mul = v_fma_mix_f32 lo(%a16), %b, -0
|
|
//! v1: %res2 = v_add_f32 %res2_mul, %c *2
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000),
|
|
fadd(fmul(f2f32(a16), b), c)));
|
|
|
|
/* neg/abs modifiers */
|
|
//! v1: %res3 = v_fma_mix_f32 -lo(%a16), %b, |lo(%c16)|
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, fadd(fmul(fneg(f2f32(a16)), b), fabs(f2f32(c16))));
|
|
|
|
//! v1: %res4 = v_fma_mix_f32 |%a|, |%b|, lo(%c16)
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, fadd(fabs(fmul(fneg(a), fneg(b))), f2f32(c16)));
|
|
|
|
//! v1: %res5 = v_fma_mix_f32 %a, -%b, lo(%c16)
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, fadd(fneg(fmul(a, b)), f2f32(c16)));
|
|
|
|
//! v1: %res6 = v_fma_mix_f32 |%a|, -|%b|, lo(%c16)
|
|
//! p_unit_test 6, %res6
|
|
writeout(6, fadd(fneg(fabs(fmul(fneg(a), fneg(b)))), f2f32(c16)));
|
|
|
|
/* output conversions */
|
|
//! v2b: %res7 = v_fma_mixlo_f16 %a, %b, %c
|
|
//! p_unit_test 7, %res7
|
|
writeout(7, f2f16(fadd(fmul(a, b), c)));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.fma.precision)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %b16 = p_startpgm
|
|
if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
Temp c = inputs[2];
|
|
Temp a16 = inputs[3];
|
|
Temp b16 = inputs[4];
|
|
|
|
/* the optimization is precise for 32-bit on GFX9 */
|
|
//~gfx9! v1: (precise)%res0 = v_fma_mix_f32 lo(%a16), %b, %c
|
|
//~gfx10! v1: (precise)%res0_tmp = v_fma_mix_f32 lo(%a16), %b, -0
|
|
//~gfx10! v1: %res0 = v_add_f32 %res0_tmp, %c
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fadd(fmul(f2f32(a16), b, bld.precise()), c));
|
|
|
|
//~gfx9! v1: (precise)%res1 = v_fma_mix_f32 lo(%a16), %b, %c
|
|
//~gfx10! v1: %res1_tmp = v_fma_mix_f32 lo(%a16), %b, -0
|
|
//~gfx10! v1: (precise)%res1 = v_add_f32 %res1_tmp, %c
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fadd(fmul(f2f32(a16), b), c, bld.precise()));
|
|
|
|
/* never promote 16-bit arithmetic to 32-bit */
|
|
//! v2b: %res2_tmp = v_cvt_f16_f32 %a
|
|
//! v2b: %res2 = v_add_f16 %res2_tmp, %b16
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, fadd(f2f16(a), b16));
|
|
|
|
//! v2b: %res3_tmp = v_cvt_f16_f32 %a
|
|
//! v2b: %res3 = v_mul_f16 %res3_tmp, %b16
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, fmul(f2f16(a), b16));
|
|
|
|
//! v2b: %res4_tmp = v_mul_f16 %a16, %b16
|
|
//! v1: %res4 = v_cvt_f32_f16 %res4_tmp
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, f2f32(fmul(a16, b16)));
|
|
|
|
//! v2b: %res5_tmp = v_add_f16 %a16, %b16
|
|
//! v1: %res5 = v_cvt_f32_f16 %res5_tmp
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, f2f32(fadd(a16, b16)));
|
|
|
|
//! v2b: %res6_tmp = v_fma_mixlo_f16 %a, %b, -0
|
|
//! v2b: %res6 = v_add_f16 %res6_tmp, %a16
|
|
//! p_unit_test 6, %res6
|
|
writeout(6, fadd(f2f16(fmul(a, b)), a16));
|
|
|
|
//! v2b: %res7_tmp = v_mul_f16 %a16, %b16
|
|
//! v1: %res7 = v_fma_mix_f32 1.0, lo(%res7_tmp), %c
|
|
//! p_unit_test 7, %res7
|
|
writeout(7, fadd(f2f32(fmul(a16, b16)), c));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.clamp)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, v2b: %a16 = p_startpgm
|
|
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a = inputs[0];
|
|
Temp a16 = inputs[1];
|
|
|
|
//! v1: %res0 = v_fma_mix_f32 lo(%a16), %a, -0 clamp
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fsat(fmul(f2f32(a16), a)));
|
|
|
|
//! v2b: %res1 = v_fma_mixlo_f16 %a, %a, -0 clamp
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, f2f16(fsat(fmul(a, a))));
|
|
|
|
//! v2b: %res2 = v_fma_mixlo_f16 %a, %a, -0 clamp
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, fsat(f2f16(fmul(a, a))));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.mad_mix.cast)
|
|
for (unsigned i = GFX9; i <= GFX10; i++) {
|
|
//>> v1: %a, v2b: %a16 = p_startpgm
|
|
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
|
|
|
|
Temp a = inputs[0];
|
|
Temp a16 = inputs[1];
|
|
|
|
/* The optimizer copy-propagates v2b=p_extract_vector(v1, 0) and p_as_uniform, so the
|
|
* optimizer has to check compatibility.
|
|
*/
|
|
|
|
//! v1: %res0_cvt = v_cvt_f32_f16 %a16
|
|
//! v2b: %res0 = v_mul_f16 %res0_cvt, %a16
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fmul(u2u16(f2f32(a16)), a16));
|
|
|
|
//! v2b: %res1_cvt = v_cvt_f16_f32 %a
|
|
//! v1: %res1 = v_mul_f32 %res1_cvt, %a
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fmul(bld.as_uniform(f2f16(a)), a));
|
|
|
|
//! v2b: %res2_mul = v_mul_f16 %a16, %a16
|
|
//! v2b: %res2 = v_cvt_f16_f32 %res2_mul
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, f2f16(bld.as_uniform(fmul(a16, a16))));
|
|
|
|
//! v1: %res3_mul = v_mul_f32 %a, %a
|
|
//! v1: %res3 = v_cvt_f32_f16 %res3_mul
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, f2f32(u2u16(fmul(a, a))));
|
|
|
|
//! v1: %res4_mul = v_fma_mix_f32 lo(%a16), %a, -0
|
|
//! v2b: %res4 = v_add_f16 %res4_mul, 0 clamp
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, fsat(u2u16(fmul(f2f32(a16), a))));
|
|
|
|
//! v2b: %res5_mul = v_fma_mixlo_f16 %a, %a, -0
|
|
//! v1: %res5 = v_add_f32 %res5_mul, 0 clamp
|
|
//! p_unit_test 5, %res5
|
|
writeout(5, fsat(bld.as_uniform(f2f16(fmul(a, a)))));
|
|
|
|
//! v1: %res6_mul = v_mul_f32 %a, %a
|
|
//! v1: %res6 = v_fma_mix_f32 1.0, lo(%res6_mul), %a
|
|
//! p_unit_test 6, %res6
|
|
writeout(6, fadd(f2f32(u2u16(fmul(a, a))), a));
|
|
|
|
//! v2b: %res7_mul = v_mul_f16 %a16, %a16
|
|
//! v1: %res7 = v_fma_mix_f32 1.0, %res7_mul, lo(%a16)
|
|
//! p_unit_test 7, %res7
|
|
writeout(7, fadd(bld.as_uniform(fmul(a16, a16)), f2f32(a16)));
|
|
|
|
/* opsel_hi should be obtained from the original opcode, not the operand regclass */
|
|
//! v1: %res8 = v_fma_mix_f32 lo(%a16), %a16, -0
|
|
//! p_unit_test 8, %res8
|
|
writeout(8, fmul(f2f32(a16), a16));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
static void
|
|
vop3p_constant(unsigned* idx, aco_opcode op, const char* swizzle, uint32_t val)
|
|
{
|
|
uint32_t halves[2] = {val & 0xffff, val >> 16};
|
|
uint32_t expected = halves[swizzle[0] - 'x'] | (halves[swizzle[1] - 'x'] << 16);
|
|
fprintf(output, "Expected for %u: 0x%.8x / %u\n", *idx, expected, expected);
|
|
|
|
unsigned opsel_lo = swizzle[0] == 'x' ? 0x0 : 0x1;
|
|
unsigned opsel_hi = swizzle[1] == 'x' ? 0x2 : 0x3;
|
|
writeout((*idx)++, bld.vop3p(op, bld.def(v1), bld.copy(bld.def(v1), Operand::c32(val)),
|
|
inputs[0], opsel_lo, opsel_hi));
|
|
}
|
|
|
|
BEGIN_TEST(optimize.vop3p_constants)
|
|
for (aco_opcode op : {aco_opcode::v_pk_add_f16, aco_opcode::v_pk_add_u16}) {
|
|
for (const char* swizzle : {"xx", "yy", "xy", "yx"}) {
|
|
char variant[16];
|
|
strcpy(variant, op == aco_opcode::v_pk_add_f16 ? "_f16" : "_u16");
|
|
strcat(variant, "_");
|
|
strcat(variant, swizzle);
|
|
|
|
//; for i in range(36):
|
|
//; insert_pattern('Expected for %u: $_ / #expected%u' % (i, i))
|
|
|
|
//>> v1: %a = p_startpgm
|
|
if (!setup_cs("v1", GFX10_3, CHIP_UNKNOWN, variant))
|
|
continue;
|
|
|
|
//; opcode = 'v_pk_add_u16' if 'u16' in variant else 'v_pk_add_f16'
|
|
//; for i in range(36):
|
|
//; insert_pattern('v1: %%res%u = %s $got%u %%a' % (i, opcode, i))
|
|
//; insert_pattern('p_unit_test %u, %%res%u' % (i, i))
|
|
//! s_endpgm
|
|
|
|
//; def parse_op(op):
|
|
//; is_int = opcode == 'v_pk_add_u16'
|
|
//; op = op.rstrip(',')
|
|
//;
|
|
//; mods = lambda v: v
|
|
//; if op.endswith('*[1,-1]'):
|
|
//; mods = lambda v: v ^ 0x80000000
|
|
//; assert(not is_int)
|
|
//; elif op.endswith('*[-1,1]'):
|
|
//; mods = lambda v: v ^ 0x00008000
|
|
//; assert(not is_int)
|
|
//; op = op.split('*')[0]
|
|
//;
|
|
//; swizzle = lambda v: v
|
|
//; if op.endswith('.xx'):
|
|
//; swizzle = lambda v: ((v & 0xffff) | (v << 16)) & 0xffffffff;
|
|
//; elif op.endswith('.yy'):
|
|
//; swizzle = lambda v: (v >> 16) | (v & 0xffff0000);
|
|
//; elif op.endswith('.yx'):
|
|
//; swizzle = lambda v: ((v >> 16) | (v << 16)) & 0xffffffff;
|
|
//; op = op.rstrip('xy.')
|
|
//;
|
|
//; val = None
|
|
//; if op.startswith('0x'):
|
|
//; val = int(op[2:], 16)
|
|
//; elif op == '-1.0':
|
|
//; val = 0xbf800000 if is_int else 0xbC00
|
|
//; elif op == '1.0':
|
|
//; val = 0x3f800000 if is_int else 0x3c00
|
|
//; else:
|
|
//; val = int(op) & 0xffffffff
|
|
//;
|
|
//; return mods(swizzle(val))
|
|
|
|
//; # Check correctness
|
|
//; for i in range(36):
|
|
//; expected = globals()['expected%u' % i]
|
|
//; got = globals()['got%u' % i]
|
|
//; got_parsed = parse_op(got)
|
|
//; if got_parsed != expected:
|
|
//; raise Exception('Check %u failed: expected 0x%.8x, got 0x%.8x ("%s")' % (i, expected, got_parsed, got))
|
|
|
|
//; # Check that all literals are ones that cannot be encoded as inline constants
|
|
//; allowed_literals = [0x00004242, 0x0000fffe, 0x00308030, 0x0030ffff, 0x3c00ffff,
|
|
//; 0x42420000, 0x42424242, 0x4242c242, 0x4242ffff, 0x7ffefffe,
|
|
//; 0x80300030, 0xbeefdead, 0xc2424242, 0xdeadbeef, 0xfffe0000,
|
|
//; 0xfffe7ffe, 0xffff0030, 0xffff3c00, 0xffff4242]
|
|
//; if opcode == 'v_pk_add_u16':
|
|
//; allowed_literals.extend([0x00003c00, 0x3c000000, 0x3c003c00, 0x3c00bc00, 0xbc003c00])
|
|
//; else:
|
|
//; allowed_literals.extend([0x00003f80, 0x3f800000])
|
|
//;
|
|
//; for i in range(36):
|
|
//; got = globals()['got%u' % i]
|
|
//; if not got.startswith('0x'):
|
|
//; continue;
|
|
//; got = int(got[2:].rstrip(',').split('*')[0].split('.')[0], 16)
|
|
//; if got not in allowed_literals:
|
|
//; raise Exception('Literal check %u failed: 0x%.8x not in allowed literals' % (i, got))
|
|
|
|
unsigned idx = 0;
|
|
for (uint32_t constant : {0x3C00, 0x0030, 0xfffe, 0x4242}) {
|
|
vop3p_constant(&idx, op, swizzle, constant);
|
|
vop3p_constant(&idx, op, swizzle, constant | 0xffff0000);
|
|
vop3p_constant(&idx, op, swizzle, constant | (constant << 16));
|
|
vop3p_constant(&idx, op, swizzle, constant << 16);
|
|
vop3p_constant(&idx, op, swizzle, (constant << 16) | 0x0000ffff);
|
|
vop3p_constant(&idx, op, swizzle, constant | ((constant ^ 0x8000) << 16));
|
|
vop3p_constant(&idx, op, swizzle, (constant ^ 0x8000) | (constant << 16));
|
|
}
|
|
|
|
for (uint32_t constant : {0x3f800000u, 0xfffffffeu, 0x00000030u, 0xdeadbeefu}) {
|
|
uint32_t lo = constant & 0xffff;
|
|
uint32_t hi = constant >> 16;
|
|
vop3p_constant(&idx, op, swizzle, constant);
|
|
vop3p_constant(&idx, op, swizzle, hi | (lo << 16));
|
|
}
|
|
|
|
finish_opt_test();
|
|
}
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.fmamix_two_literals)
|
|
/* This test has to recreate literals sometimes because we don't combine them at all if there's
|
|
* at least one uncombined use.
|
|
*/
|
|
for (unsigned i = GFX10; i <= GFX10_3; i++) {
|
|
//>> v1: %a, v1: %b = p_startpgm
|
|
if (!setup_cs("v1 v1", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
|
|
Temp c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f)));
|
|
Temp c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f)));
|
|
Temp c_denorm = bld.copy(bld.def(v1), Operand::c32(0x387fc000));
|
|
|
|
//! v1: %res0 = v_fma_mix_f32 %a, lo(0x42003e00), hi(0x42003e00)
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fma(a, c15, c30));
|
|
|
|
/* No need to use v_fma_mix_f32. */
|
|
//! v1: %res1 = v_fmaak_f32 %a, %b, 0x40400000
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fma(a, b, c30));
|
|
|
|
/* Separate mul/add can become v_fma_mix_f32 if it's not precise. */
|
|
//! v1: %res2 = v_fma_mix_f32 %a, lo(0x42003e00), hi(0x42003e00)
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, fadd(fmul(a, c15), c30));
|
|
|
|
//~gfx10! v1: %c15 = p_parallelcopy 0x3fc00000
|
|
c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f)));
|
|
c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f)));
|
|
|
|
/* v_fma_mix_f32 is a fused mul/add, so it can't be used for precise separate mul/add. */
|
|
//~gfx10! v1: (precise)%res3 = v_madak_f32 %a, %c15, 0x40400000
|
|
//~gfx10_3! v1: (precise)%res3_tmp = v_mul_f32 %a, 0x3fc00000
|
|
//~gfx10_3! v1: %res3 = v_add_f32 %res3_tmp, 0x40400000
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, fadd(bld.precise().vop2(aco_opcode::v_mul_f32, bld.def(v1), a, c15), c30));
|
|
|
|
//~gfx10! v1: (precise)%res4 = v_madak_f32 %1, %c16, 0x40400000
|
|
//~gfx10_3! v1: %res4_tmp = v_mul_f32 %a, 0x3fc00000
|
|
//~gfx10_3! v1: (precise)%res4 = v_add_f32 %res4_tmp, 0x40400000
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, bld.precise().vop2(aco_opcode::v_add_f32, bld.def(v1), fmul(a, c15), c30));
|
|
|
|
/* Can't convert to fp16 if it will be flushed as a denormal. */
|
|
//! v1: %res5 = v_fma_mix_f32 %1, lo(0x3ff3e00), hi(0x3ff3e00)
|
|
//! p_unit_test 5, %res5
|
|
c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f)));
|
|
writeout(5, fma(a, c15, c_denorm));
|
|
|
|
//>> BB1
|
|
//! /* logical preds: / linear preds: / kind: uniform, */
|
|
program->next_fp_mode.denorm16_64 = fp_denorm_flush;
|
|
bld.reset(program->create_and_insert_block());
|
|
|
|
//~gfx10; del c15
|
|
//! v1: %c15 = p_parallelcopy 0x3fc00000
|
|
//! v1: %res6 = v_fmaak_f32 %a, %c15, 0x387fc000
|
|
//! p_unit_test 6, %res6
|
|
c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f)));
|
|
c_denorm = bld.copy(bld.def(v1), Operand::c32(0x387fc000));
|
|
writeout(6, fma(a, c15, c_denorm));
|
|
|
|
/* Can't accept more than 3 unique fp16 literals. */
|
|
//! v1: %c45 = p_parallelcopy 0x40900000
|
|
//! v1: %res7 = v_fma_mix_f32 lo(0x42003e00), hi(0x42003e00), %c45
|
|
//! p_unit_test 7, %res7
|
|
Temp c45 = bld.copy(bld.def(v1), Operand::c32(fui(4.5f)));
|
|
writeout(7, fma(c15, c30, c45));
|
|
|
|
/* Modifiers must be preserved. */
|
|
//! v1: %res8 = v_fma_mix_f32 -%a, lo(0x44804200), hi(0x44804200)
|
|
//! p_unit_test 8, %res8
|
|
writeout(8, fma(fneg(a), c30, c45));
|
|
|
|
//! v1: %res9 = v_fma_mix_f32 lo(0x44804200), |%a|, hi(0x44804200)
|
|
//! p_unit_test 9, %res9
|
|
writeout(9, fma(c30, fabs(a), c45));
|
|
|
|
//! v1: %res10 = v_fma_mix_f32 %a, lo(0x44804200), hi(0x44804200) clamp
|
|
//! p_unit_test 10, %res10
|
|
writeout(10, fsat(fma(a, c30, c45)));
|
|
|
|
/* Output modifiers are not supported by v_fma_mix_f32. */
|
|
c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f)));
|
|
//; del c45
|
|
//! v1: %c45 = p_parallelcopy 0x40900000
|
|
//! v1: %res11 = v_fma_f32 %a, 0x40400000, %c45 *0.5
|
|
//! p_unit_test 11, %res11
|
|
c45 = bld.copy(bld.def(v1), Operand::c32(fui(4.5f)));
|
|
writeout(11, fmul(fma(a, c30, c45), bld.copy(bld.def(v1), Operand::c32(0x3f000000))));
|
|
|
|
/* Has a literal which can't be represented as fp16. */
|
|
//! v1: %c03 = p_parallelcopy 0x3e99999a
|
|
//! v1: %res12 = v_fmaak_f32 %a, %c03, 0x40400000
|
|
//! p_unit_test 12, %res12
|
|
Temp c03 = bld.copy(bld.def(v1), Operand::c32(fui(0.3f)));
|
|
writeout(12, fma(a, c03, c30));
|
|
|
|
/* We should still use fmaak/fmamk if the two literals are identical. */
|
|
//! v1: %res13 = v_fmaak_f32 0x40400000, %a, 0x40400000
|
|
//! p_unit_test 13, %res13
|
|
writeout(13, fma(a, c30, c30));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.fma_opsel)
|
|
/* TODO make these work before GFX11 using SDWA. */
|
|
for (unsigned i = GFX11; i <= GFX11; i++) {
|
|
//>> v2b: %a, v2b: %b, v1: %c, v1: %d, v1: %e = p_startpgm
|
|
if (!setup_cs("v2b v2b v1 v1 v1", (amd_gfx_level)i))
|
|
continue;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
Temp c = inputs[2];
|
|
Temp d = inputs[3];
|
|
Temp e = inputs[4];
|
|
Temp c_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), c, Operand::c32(1));
|
|
Temp d_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), d, Operand::c32(1));
|
|
Temp e_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), e, Operand::c32(1));
|
|
|
|
//! v2b: %res0 = v_fma_f16 %b, hi(%c), %a
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fadd(fmul(b, c_hi), a));
|
|
|
|
//! v2b: %res1 = v_fma_f16 %a, %b, hi(%d)
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fadd(fmul(a, b), d_hi));
|
|
|
|
//! v2b: %res2 = v_fma_f16 %a, %b, hi(%e)
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, fma(a, b, e_hi));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.dpp_opsel)
|
|
//>> v1: %a, v1: %b = p_startpgm
|
|
if (!setup_cs("v1 v1", GFX11))
|
|
return;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
|
|
Temp dpp16 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
|
|
Temp dpp16_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), dpp16, Operand::c32(1));
|
|
Temp dpp8 = bld.vop1_dpp8(aco_opcode::v_mov_b32, bld.def(v1), a);
|
|
Temp dpp8_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), dpp8, Operand::c32(1));
|
|
|
|
Temp b_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(1));
|
|
Temp b_lo = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(0));
|
|
|
|
//! v2b: %res0 = v_add_f16 hi(%a), hi(%b) row_mirror bound_ctrl:1 fi
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fadd(dpp16_hi, b_hi));
|
|
|
|
//! v2b: %res1 = v_add_f16 hi(%a), %b dpp8:[0,0,0,0,0,0,0,0] fi
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fadd(b_lo, dpp8_hi));
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.apply_sgpr_swap_opsel)
|
|
//>> v1: %a, s1: %b = p_startpgm
|
|
if (!setup_cs("v1 s1", GFX11))
|
|
return;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
|
|
Temp b_vgpr = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), bld.copy(bld.def(v1), b),
|
|
Operand::c32(0));
|
|
|
|
Temp res0 = bld.tmp(v2b);
|
|
VALU_instruction& valu = bld.vop2(aco_opcode::v_sub_f16, Definition(res0), a, b_vgpr)->valu();
|
|
valu.opsel[0] = true;
|
|
|
|
//! v2b: %res0 = v_subrev_f16 %b, hi(%a)
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, res0);
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.combine_comparison_ordering)
|
|
//>> v1: %a, v1: %b, v1: %c = p_startpgm
|
|
if (!setup_cs("v1 v1 v1", GFX11))
|
|
return;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
Temp c = inputs[2];
|
|
|
|
Temp a_unordered = bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), a, a);
|
|
Temp b_unordered = bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), b, b);
|
|
Temp unordered =
|
|
bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), a_unordered, b_unordered);
|
|
|
|
Temp a_lt_a = bld.vopc(aco_opcode::v_cmp_lt_f32, bld.def(bld.lm), a, a);
|
|
Temp unordered_cmp =
|
|
bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), unordered, a_lt_a);
|
|
|
|
//! s2: %res0_unordered = v_cmp_u_f32 %a, %b
|
|
//! s2: %res0_cmp = v_cmp_lt_f32 %a, %a
|
|
//! s2: %res0, s2: %_:scc = s_or_b64 %res0_unordered, %res0_cmp
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, unordered_cmp);
|
|
|
|
Temp c_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), c, Operand::c32(1));
|
|
|
|
Temp c_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), c, c);
|
|
Temp c_hi_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), c_hi, c_hi);
|
|
unordered =
|
|
bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), c_unordered, c_hi_unordered);
|
|
|
|
Temp c_lt_c_hi = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), c, c_hi);
|
|
unordered_cmp =
|
|
bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), unordered, c_lt_c_hi);
|
|
|
|
//! s2: %res1 = v_cmp_nge_f16 %c, hi(%c)
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, unordered_cmp);
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.combine_comparison_ordering_opsel)
|
|
//>> v1: %a, v2b: %b = p_startpgm
|
|
if (!setup_cs("v1 v2b", GFX11))
|
|
return;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
|
|
Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1));
|
|
|
|
Temp ahi_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a_hi, a_hi);
|
|
Temp b_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), b, b);
|
|
Temp unordered =
|
|
bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), ahi_unordered, b_unordered);
|
|
|
|
Temp ahi_lt_b = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), a_hi, b);
|
|
Temp unordered_cmp =
|
|
bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), unordered, ahi_lt_b);
|
|
|
|
//! s2: %res0 = v_cmp_nge_f16 hi(%a), %b
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, unordered_cmp);
|
|
|
|
Temp ahi_cmp_const = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), a_hi,
|
|
bld.copy(bld.def(v2b), Operand::c16(0x4400)));
|
|
Temp ahi_ucmp_const =
|
|
bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), ahi_unordered, ahi_cmp_const);
|
|
//! s2: %res1 = v_cmp_nle_f16 4.0, hi(%a)
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, ahi_ucmp_const);
|
|
|
|
a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1));
|
|
ahi_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a_hi, a_hi);
|
|
b_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), b, b);
|
|
unordered =
|
|
bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), ahi_unordered, b_unordered);
|
|
Temp alo_lt_b = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), a, b);
|
|
Temp noopt = bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), unordered, alo_lt_b);
|
|
//! s2: %u2 = v_cmp_u_f16 hi(%a), %b
|
|
//! s2: %cmp2 = v_cmp_lt_f16 %a, %b
|
|
//! s2: %res2, s2: %scc2:scc = s_or_b64 %u2, %cmp2
|
|
//! p_unit_test 2, %res2
|
|
writeout(2, noopt);
|
|
|
|
Temp hi_neq_lo = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a, a_hi);
|
|
Temp a_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a, a);
|
|
noopt = bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), hi_neq_lo, a_unordered);
|
|
//! s2: %nan31 = v_cmp_neq_f16 %a, hi(%a)
|
|
//! s2: %nan32 = v_cmp_neq_f16 %a, %a
|
|
//! s2: %res3, s2: %scc3:scc = s_or_b64 %nan31, %nan32
|
|
//! p_unit_test 3, %res3
|
|
writeout(3, noopt);
|
|
|
|
ahi_cmp_const = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), a_hi,
|
|
bld.copy(bld.def(v2b), Operand::c16(0x4400)));
|
|
a_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a, a);
|
|
noopt =
|
|
bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), a_unordered, ahi_cmp_const);
|
|
//! s2: %cmp4 = v_cmp_gt_f16 4.0, hi(%a)
|
|
//! s2: %nan4 = v_cmp_neq_f16 %a, %a
|
|
//! s2: %res4, s2: %scc4:scc = s_or_b64 %nan4, %cmp4
|
|
//! p_unit_test 4, %res4
|
|
writeout(4, noopt);
|
|
|
|
finish_opt_test();
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.max3_opsel)
|
|
/* TODO make these work before GFX11 using SDWA. */
|
|
for (unsigned i = GFX11; i <= GFX11; i++) {
|
|
//>> v1: %a, v1: %b, v2b: %c = p_startpgm
|
|
if (!setup_cs("v1 v1 v2b", GFX11))
|
|
continue;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
Temp c = inputs[2];
|
|
|
|
Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1));
|
|
Temp b_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(1));
|
|
|
|
//! v2b: %res0 = v_max3_f16 hi(%a), %c, hi(%b)
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, bld.vop2(aco_opcode::v_max_f16, bld.def(v2b),
|
|
bld.vop2(aco_opcode::v_max_f16, bld.def(v2b), a_hi, c), b_hi));
|
|
|
|
finish_opt_test();
|
|
}
|
|
END_TEST
|
|
|
|
BEGIN_TEST(optimize.neg_mul_opsel)
|
|
//>> v1: %a, v2b: %b = p_startpgm
|
|
if (!setup_cs("v1 v2b", GFX11))
|
|
return;
|
|
|
|
Temp a = inputs[0];
|
|
Temp b = inputs[1];
|
|
|
|
Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1));
|
|
|
|
//! v2b: %res0 = v_mul_f16 -hi(%a), %b
|
|
//! p_unit_test 0, %res0
|
|
writeout(0, fneg(fmul(a_hi, b)));
|
|
|
|
//! v1: %res1 = v_fma_mix_f32 -hi(%a), lo(%b), -0
|
|
//! p_unit_test 1, %res1
|
|
writeout(1, fneg(fmul(f2f32(a_hi), f2f32(b))));
|
|
|
|
finish_opt_test();
|
|
END_TEST
|