diff options
Diffstat (limited to 'src/amd/compiler/tests/test_optimizer.cpp')
-rw-r--r-- | src/amd/compiler/tests/test_optimizer.cpp | 1405 |
1 files changed, 1275 insertions, 130 deletions
diff --git a/src/amd/compiler/tests/test_optimizer.cpp b/src/amd/compiler/tests/test_optimizer.cpp index 31a229f99e9..c9a28599cc0 100644 --- a/src/amd/compiler/tests/test_optimizer.cpp +++ b/src/amd/compiler/tests/test_optimizer.cpp @@ -1,25 +1,7 @@ /* * Copyright © 2020 Valve Corporation * - * Permission is hereby granted, free of charge, to any person obtaining a - * copy of this software and associated documentation files (the "Software"), - * to deal in the Software without restriction, including without limitation - * the rights to use, copy, modify, merge, publish, distribute, sublicense, - * and/or sell copies of the Software, and to permit persons to whom the - * Software is furnished to do so, subject to the following conditions: - * - * The above copyright notice and this permission notice (including the next - * paragraph) shall be included in all copies or substantial portions of the - * Software. - * - * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR - * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL - * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER - * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING - * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS - * IN THE SOFTWARE. - * + * SPDX-License-Identifier: MIT */ #include "helpers.h" @@ -28,7 +10,7 @@ using namespace aco; BEGIN_TEST(optimize.neg) for (unsigned i = GFX9; i <= GFX10; i++) { //>> v1: %a, v1: %b, s1: %c, s1: %d = p_startpgm - if (!setup_cs("v1 v1 s1 s1", (chip_class)i)) + if (!setup_cs("v1 v1 s1 s1", (amd_gfx_level)i)) continue; //! v1: %res0 = v_mul_f32 %a, -%b @@ -59,9 +41,11 @@ BEGIN_TEST(optimize.neg) Temp neg_abs_a = fneg(abs_a); writeout(4, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), neg_abs_a, inputs[1])); - //! v1: %res5 = v_mul_f32 -%a, %b row_shl:1 bound_ctrl:1 + //~gfx9! v1: %res5 = v_mul_f32 -%a, %b row_shl:1 bound_ctrl:1 + //~gfx10! v1: %res5 = v_mul_f32 -%a, %b row_shl:1 bound_ctrl:1 fi //! p_unit_test 5, %res5 - writeout(5, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), neg_a, inputs[1], dpp_row_sl(1))); + writeout(5, + bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), neg_a, inputs[1], dpp_row_sl(1))); //! v1: %res6 = v_subrev_f32 %a, %b //! p_unit_test 6, %res6 @@ -264,15 +248,16 @@ BEGIN_TEST(optimize.output_modifiers) finish_opt_test(); END_TEST -Temp create_subbrev_co(Operand op0, Operand op1, Operand op2) +Temp +create_subbrev_co(Operand op0, Operand op1, Operand op2) { - return bld.vop2_e64(aco_opcode::v_subbrev_co_u32, bld.def(v1), bld.hint_vcc(bld.def(bld.lm)), op0, op1, op2); + return bld.vop2_e64(aco_opcode::v_subbrev_co_u32, bld.def(v1), bld.def(bld.lm), op0, op1, op2); } BEGIN_TEST(optimize.cndmask) for (unsigned i = GFX9; i <= GFX10; i++) { //>> v1: %a, s1: %b, s2: %c = p_startpgm - if (!setup_cs("v1 s1 s2", (chip_class)i)) + if (!setup_cs("v1 s1 s2", (amd_gfx_level)i)) continue; Temp subbrev; @@ -316,7 +301,7 @@ END_TEST BEGIN_TEST(optimize.add_lshl) for (unsigned i = GFX8; i <= GFX10; i++) { //>> s1: %a, v1: %b = p_startpgm - if (!setup_cs("s1 v1", (chip_class)i)) + if (!setup_cs("s1 v1", (amd_gfx_level)i)) continue; Temp shift; @@ -345,42 +330,45 @@ BEGIN_TEST(optimize.add_lshl) Temp vadd = bld.vadd32(bld.def(v1), shift, Operand(inputs[1])); writeout(1, bld.vadd32(bld.def(v1), sadd, vadd)); - //~gfx8! s1: %lshl2 = s_lshl_b32 %a, 3 + //~gfx8! s1: %lshl2, s1: %_:scc = s_lshl_b32 %a, 3 //~gfx8! v1: %res2, s2: %_ = v_add_co_u32 %lshl2, %b //~gfx(9|10)! v1: %res2 = v_lshl_add_u32 %a, 3, %b //! p_unit_test 2, %res2 - Temp lshl = - bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), Operand(inputs[0]), Operand::c32(3u)); + Temp lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), + Operand(inputs[0]), Operand::c32(3u)); writeout(2, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); - //~gfx8! s1: %lshl3 = s_lshl_b32 (is24bit)%a, 7 + //~gfx8! s1: %lshl3, s1: %_:scc = s_lshl_b32 (is24bit)%a, 7 //~gfx8! v1: %res3, s2: %_ = v_add_co_u32 %lshl3, %b //~gfx(9|10)! v1: %res3 = v_lshl_add_u32 (is24bit)%a, 7, %b //! p_unit_test 3, %res3 Operand a_24bit = Operand(inputs[0]); a_24bit.set24bit(true); - lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_24bit, Operand::c32(7u)); + lshl = + bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, Operand::c32(7u)); writeout(3, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); - //! s1: %lshl4 = s_lshl_b32 (is24bit)%a, 3 + //! s1: %lshl4, s1: %_:scc = s_lshl_b32 (is24bit)%a, 3 //~gfx(8|9)! v1: %res4, s2: %carry = v_add_co_u32 %lshl4, %b //~gfx10! v1: %res4, s2: %carry = v_add_co_u32_e64 %lshl4, %b //! p_unit_test 4, %carry - lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_24bit, Operand::c32(3u)); + lshl = + bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, Operand::c32(3u)); Temp carry = bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]), true).def(1).getTemp(); writeout(4, carry); - //~gfx8! s1: %lshl5 = s_lshl_b32 (is24bit)%a, (is24bit)%a + //~gfx8! s1: %lshl5, s1: %_:scc = s_lshl_b32 (is24bit)%a, (is24bit)%a //~gfx8! v1: %res5, s2: %_ = v_add_co_u32 %lshl5, %b //~gfx(9|10)! v1: %res5 = v_lshl_add_u32 (is24bit)%a, (is24bit)%a, %b //! p_unit_test 5, %res5 - lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_24bit, a_24bit); + lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, a_24bit); writeout(5, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); //~gfx8! v1: %res6 = v_mad_u32_u24 (is24bit)%a, 8, %b //~gfx(9|10)! v1: %res6 = v_lshl_add_u32 (is24bit)%a, 3, %b //! p_unit_test 6, %res6 - lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_24bit, Operand::c32(3u)); + lshl = + bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, Operand::c32(3u)); writeout(6, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); //~gfx8! v1: %res7 = v_mad_u32_u24 (is16bit)%a, 16, %b @@ -388,7 +376,8 @@ BEGIN_TEST(optimize.add_lshl) //! p_unit_test 7, %res7 Operand a_16bit = Operand(inputs[0]); a_16bit.set16bit(true); - lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), a_16bit, Operand::c32(4u)); + lshl = + bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_16bit, Operand::c32(4u)); writeout(7, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); finish_opt_test(); @@ -398,7 +387,7 @@ END_TEST BEGIN_TEST(optimize.bcnt) for (unsigned i = GFX8; i <= GFX10; i++) { //>> v1: %a, s1: %b = p_startpgm - if (!setup_cs("v1 s1", (chip_class)i)) + if (!setup_cs("v1 s1", (amd_gfx_level)i)) continue; Temp bcnt; @@ -438,7 +427,7 @@ BEGIN_TEST(optimize.bcnt) END_TEST struct clamp_config { - const char *name; + const char* name; aco_opcode min, max, med3; Operand lb, ub; }; @@ -530,21 +519,22 @@ BEGIN_TEST(optimize.clamp) bld.vop2(cfg.max, bld.def(v1), inputs[2], inputs[0]))); /* correct NaN behaviour with precise */ - - //! v1: %res7 = @med3 @ub, @lb, %a - //! p_unit_test 7, %res7 - Builder::Result max = bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0]); - max.def(0).setPrecise(true); - Builder::Result min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, max); - max.def(0).setPrecise(true); - writeout(7, min); - - //! v1: (precise)%res8_tmp = @min @ub, %a - //! v1: %res8 = @max @lb, %res8_tmp - //! p_unit_test 8, %res8 - min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, inputs[0]); - min.def(0).setPrecise(true); - writeout(8, bld.vop2(cfg.max, bld.def(v1), cfg.lb, min)); + if (cfg.min == aco_opcode::v_min_f16 || cfg.min == aco_opcode::v_min_f32) { + //~f(16|32)! v1: %res7 = @med3 @ub, @lb, %a + //~f(16|32)! p_unit_test 7, %res7 + Builder::Result max = bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0]); + max.def(0).setPrecise(true); + Builder::Result min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, max); + max.def(0).setPrecise(true); + writeout(7, min); + + //~f(16|32)! v1: (precise)%res8_tmp = @min @ub, %a + //~f(16|32)! v1: %res8 = @max @lb, %res8_tmp + //~f(16|32)! p_unit_test 8, %res8 + min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, inputs[0]); + min.def(0).setPrecise(true); + writeout(8, bld.vop2(cfg.max, bld.def(v1), cfg.lb, min)); + } finish_opt_test(); } @@ -672,12 +662,12 @@ BEGIN_TEST(optimize.const_comparison_ordering) src0 = bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), inputs[0], inputs[0]); writeout(13, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc), src0, src1)); - //! s2: %tmp14_0 = v_cmp_lt_f64 -1, %a - //! s2: %tmp14_1 = v_cmp_neq_f64 %a, %a + //! s2: %tmp14_0 = v_cmp_lt_f64 -1, %c + //! s2: %tmp14_1 = v_cmp_neq_f64 %c, %c //! s2: %res14, s1: %_:scc = s_or_b64 %tmp14_1, %tmp14_0 //! p_unit_test 14, %res14 - src1 = bld.vopc(aco_opcode::v_cmp_lt_f64, bld.def(bld.lm), Operand::c64(nan64), inputs[0]); - src0 = bld.vopc(aco_opcode::v_cmp_neq_f64, bld.def(bld.lm), inputs[0], inputs[0]); + src1 = bld.vopc(aco_opcode::v_cmp_lt_f64, bld.def(bld.lm), Operand::c64(nan64), inputs[2]); + src0 = bld.vopc(aco_opcode::v_cmp_neq_f64, bld.def(bld.lm), inputs[2], inputs[2]); writeout(14, bld.sop2(aco_opcode::s_or_b64, bld.def(bld.lm), bld.def(s1, scc), src0, src1)); finish_opt_test(); @@ -697,7 +687,7 @@ BEGIN_TEST(optimize.add3) //! v1: %res1 = v_add_u32 %a, %tmp1 //! p_unit_test 1, %res1 tmp = bld.vop2_e64(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]); - tmp.instr->vop3().clamp = true; + tmp->valu().clamp = true; writeout(1, bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp)); //! v1: %tmp2 = v_add_u32 %b, %c @@ -705,30 +695,71 @@ BEGIN_TEST(optimize.add3) //! p_unit_test 2, %res2 tmp = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]); tmp = bld.vop2_e64(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp); - tmp.instr->vop3().clamp = true; + tmp->valu().clamp = true; writeout(2, tmp); finish_opt_test(); END_TEST BEGIN_TEST(optimize.minmax) - for (unsigned i = GFX9; i <= GFX10; i++) { - //>> v1: %a = p_startpgm - if (!setup_cs("v1", (chip_class)i)) + for (unsigned i = GFX10_3; i <= GFX11; i++) { + //>> v1: %a, v1: %b, v1: %c = p_startpgm + if (!setup_cs("v1 v1 v1", (amd_gfx_level)i)) continue; - //! v1: %res0 = v_max3_f32 0, -0, %a + Temp a = inputs[0]; + Temp b = inputs[1]; + Temp c = inputs[2]; + + //! v1: %res0 = v_min3_f32 %a, %b, %c //! p_unit_test 0, %res0 - Temp xor0 = fneg(inputs[0]); - Temp min = bld.vop2(aco_opcode::v_min_f32, bld.def(v1), Operand::zero(), xor0); - Temp xor1 = fneg(min); - writeout(0, bld.vop2(aco_opcode::v_max_f32, bld.def(v1), Operand::zero(), xor1)); + writeout(0, fmin(c, fmin(a, b))); - //! v1: %res1 = v_max3_f32 0, -0, -%a + //! v1: %res1 = v_max3_f32 %a, %b, %c //! p_unit_test 1, %res1 - min = bld.vop2(aco_opcode::v_min_f32, bld.def(v1), Operand::zero(), Operand(inputs[0])); - xor1 = fneg(min); - writeout(1, bld.vop2(aco_opcode::v_max_f32, bld.def(v1), Operand::zero(), xor1)); + writeout(1, fmax(c, fmax(a, b))); + + //! v1: %res2 = v_min3_f32 -%a, -%b, %c + //! p_unit_test 2, %res2 + writeout(2, fmin(c, fneg(fmax(a, b)))); + + //! v1: %res3 = v_max3_f32 -%a, -%b, %c + //! p_unit_test 3, %res3 + writeout(3, fmax(c, fneg(fmin(a, b)))); + + //! v1: %res4 = v_max3_f32 -%a, %b, %c + //! p_unit_test 4, %res4 + writeout(4, fmax(c, fneg(fmin(a, fneg(b))))); + + //~gfx10_3! v1: %res5_tmp = v_max_f32 %a, %b + //~gfx10_3! v1: %res5 = v_min_f32 %c, %res5_tmp + //~gfx11! v1: %res5 = v_maxmin_f32 %a, %b, %c + //! p_unit_test 5, %res5 + writeout(5, fmin(c, fmax(a, b))); + + //~gfx10_3! v1: %res6_tmp = v_min_f32 %a, %b + //~gfx10_3! v1: %res6 = v_max_f32 %c, %res6_tmp + //~gfx11! v1: %res6 = v_minmax_f32 %a, %b, %c + //! p_unit_test 6, %res6 + writeout(6, fmax(c, fmin(a, b))); + + //~gfx10_3! v1: %res7_tmp = v_min_f32 %a, %b + //~gfx10_3! v1: %res7 = v_min_f32 %c, -%res7_tmp + //~gfx11! v1: %res7 = v_maxmin_f32 -%a, -%b, %c + //! p_unit_test 7, %res7 + writeout(7, fmin(c, fneg(fmin(a, b)))); + + //~gfx10_3! v1: %res8_tmp = v_max_f32 %a, %b + //~gfx10_3! v1: %res8 = v_max_f32 %c, -%res8_tmp + //~gfx11! v1: %res8 = v_minmax_f32 -%a, -%b, %c + //! p_unit_test 8, %res8 + writeout(8, fmax(c, fneg(fmax(a, b)))); + + //~gfx10_3! v1: %res9_tmp = v_max_f32 %a, -%b + //~gfx10_3! v1: %res9 = v_max_f32 %c, -%res9_tmp + //~gfx11! v1: %res9 = v_minmax_f32 -%a, %b, %c + //! p_unit_test 9, %res9 + writeout(9, fmax(c, fneg(fmax(a, fneg(b))))); finish_opt_test(); } @@ -737,7 +768,7 @@ END_TEST BEGIN_TEST(optimize.mad_32_24) for (unsigned i = GFX8; i <= GFX9; i++) { //>> v1: %a, v1: %b, v1: %c = p_startpgm - if (!setup_cs("v1 v1 v1", (chip_class)i)) + if (!setup_cs("v1 v1 v1", (amd_gfx_level)i)) continue; //! v1: %res0 = v_mad_u32_u24 %b, %c, %a @@ -758,7 +789,7 @@ END_TEST BEGIN_TEST(optimize.add_lshlrev) for (unsigned i = GFX8; i <= GFX10; i++) { //>> v1: %a, v1: %b, s1: %c = p_startpgm - if (!setup_cs("v1 v1 s1", (chip_class)i)) + if (!setup_cs("v1 v1 s1", (amd_gfx_level)i)) continue; Temp lshl; @@ -821,7 +852,7 @@ enum denorm_op { denorm_fnegabs = 3, }; -static const char *denorm_op_names[] = { +static const char* denorm_op_names[] = { "mul1", "fneg", "fabs", @@ -835,31 +866,27 @@ struct denorm_config { aco_opcode dest; }; -static const char *srcdest_op_name(aco_opcode op) +static const char* +srcdest_op_name(aco_opcode op) { switch (op) { - case aco_opcode::v_cndmask_b32: - return "cndmask"; - case aco_opcode::v_min_f32: - return "min"; - case aco_opcode::v_rcp_f32: - return "rcp"; - default: - return "none"; + case aco_opcode::v_cndmask_b32: return "cndmask"; + case aco_opcode::v_min_f32: return "min"; + case aco_opcode::v_rcp_f32: return "rcp"; + default: return "none"; } } -static Temp emit_denorm_srcdest(aco_opcode op, Temp val) +static Temp +emit_denorm_srcdest(aco_opcode op, Temp val) { switch (op) { case aco_opcode::v_cndmask_b32: return bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), Operand::zero(), val, inputs[1]); case aco_opcode::v_min_f32: return bld.vop2(aco_opcode::v_min_f32, bld.def(v1), Operand::zero(), val); - case aco_opcode::v_rcp_f32: - return bld.vop1(aco_opcode::v_rcp_f32, bld.def(v1), val); - default: - return val; + case aco_opcode::v_rcp_f32: return bld.vop1(aco_opcode::v_rcp_f32, bld.def(v1), val); + default: return val; } } @@ -875,7 +902,8 @@ BEGIN_TEST(optimize.denorm_propagation) configs.push_back({flush, op, aco_opcode::num_opcodes, dest}); } - for (aco_opcode src : {aco_opcode::v_cndmask_b32, aco_opcode::v_min_f32, aco_opcode::v_rcp_f32}) { + for (aco_opcode src : + {aco_opcode::v_cndmask_b32, aco_opcode::v_min_f32, aco_opcode::v_rcp_f32}) { for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs}) configs.push_back({flush, op, src, aco_opcode::num_opcodes}); } @@ -883,18 +911,18 @@ BEGIN_TEST(optimize.denorm_propagation) for (denorm_config cfg : configs) { char subvariant[128]; - sprintf(subvariant, "_%s_%s_%s_%s", - cfg.flush ? "flush" : "keep", srcdest_op_name(cfg.src), + sprintf(subvariant, "_%s_%s_%s_%s", cfg.flush ? "flush" : "keep", srcdest_op_name(cfg.src), denorm_op_names[(int)cfg.op], srcdest_op_name(cfg.dest)); - if (!setup_cs("v1 s2", (chip_class)i, CHIP_UNKNOWN, subvariant)) + if (!setup_cs("v1 s2", (amd_gfx_level)i, CHIP_UNKNOWN, subvariant)) continue; - bool can_propagate = cfg.src == aco_opcode::v_rcp_f32 || (i >= GFX9 && cfg.src == aco_opcode::v_min_f32) || - cfg.dest == aco_opcode::v_rcp_f32 || (i >= GFX9 && cfg.dest == aco_opcode::v_min_f32) || - !cfg.flush; + bool can_propagate = cfg.src == aco_opcode::v_rcp_f32 || + (i >= GFX9 && cfg.src == aco_opcode::v_min_f32) || + cfg.dest == aco_opcode::v_rcp_f32 || + (i >= GFX9 && cfg.dest == aco_opcode::v_min_f32) || !cfg.flush; - fprintf(output, "src, dest, op: %s %s %s\n", - srcdest_op_name(cfg.src), srcdest_op_name(cfg.dest), denorm_op_names[(int)cfg.op]); + fprintf(output, "src, dest, op: %s %s %s\n", srcdest_op_name(cfg.src), + srcdest_op_name(cfg.dest), denorm_op_names[(int)cfg.op]); fprintf(output, "can_propagate: %u\n", can_propagate); //! src, dest, op: $src $dest $op //! can_propagate: #can_propagate @@ -934,15 +962,9 @@ BEGIN_TEST(optimize.denorm_propagation) case denorm_mul1: val = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x3f800000u), val); break; - case denorm_fneg: - val = fneg(val); - break; - case denorm_fabs: - val = fabs(val); - break; - case denorm_fnegabs: - val = fneg(fabs(val)); - break; + case denorm_fneg: val = fneg(val); break; + case denorm_fabs: val = fabs(val); break; + case denorm_fnegabs: val = fneg(fabs(val)); break; } val = emit_denorm_srcdest(cfg.dest, val); writeout( @@ -954,80 +976,112 @@ BEGIN_TEST(optimize.denorm_propagation) END_TEST BEGIN_TEST(optimizer.dpp) - //>> v1: %a, v1: %b, s2: %c = p_startpgm - if (!setup_cs("v1 v1 s2", GFX10_3)) + //>> v1: %a, v1: %b, s2: %c, s1: %d = p_startpgm + if (!setup_cs("v1 v1 s2 s1", GFX10_3)) return; Operand a(inputs[0]); Operand b(inputs[1]); Operand c(inputs[2]); + Operand d(inputs[3]); /* basic optimization */ - //! v1: %res0 = v_add_f32 %a, %b row_mirror bound_ctrl:1 + //! v1: %res0 = v_add_f32 %a, %b row_mirror bound_ctrl:1 fi //! p_unit_test 0, %res0 Temp tmp0 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); Temp res0 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), tmp0, b); writeout(0, res0); /* operand swapping */ - //! v1: %res1 = v_subrev_f32 %a, %b row_mirror bound_ctrl:1 + //! v1: %res1 = v_subrev_f32 %a, %b row_mirror bound_ctrl:1 fi //! p_unit_test 1, %res1 Temp tmp1 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); Temp res1 = bld.vop2(aco_opcode::v_sub_f32, bld.def(v1), b, tmp1); writeout(1, res1); - //! v1: %tmp2 = v_mov_b32 %a row_mirror bound_ctrl:1 - //! v1: %res2 = v_sub_f32 %b, %tmp2 row_half_mirror bound_ctrl:1 + //! v1: %tmp2 = v_mov_b32 %a row_mirror bound_ctrl:1 fi + //! v1: %res2 = v_sub_f32 %b, %tmp2 row_half_mirror bound_ctrl:1 fi //! p_unit_test 2, %res2 Temp tmp2 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); Temp res2 = bld.vop2_dpp(aco_opcode::v_sub_f32, bld.def(v1), b, tmp2, dpp_row_half_mirror); writeout(2, res2); /* modifiers */ - //! v1: %res3 = v_add_f32 -%a, %b row_mirror bound_ctrl:1 + //! v1: %res3 = v_add_f32 -%a, %b row_mirror bound_ctrl:1 fi //! p_unit_test 3, %res3 auto tmp3 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); - tmp3.instr->dpp().neg[0] = true; + tmp3->dpp16().neg[0] = true; Temp res3 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), tmp3, b); writeout(3, res3); - //! v1: %res4 = v_add_f32 -%a, %b row_mirror bound_ctrl:1 + //! v1: %res4 = v_add_f32 -%a, %b row_mirror bound_ctrl:1 fi //! p_unit_test 4, %res4 Temp tmp4 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); auto res4 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp4, b); - res4.instr->vop3().neg[0] = true; + res4->valu().neg[0] = true; writeout(4, res4); - //! v1: %tmp5 = v_mov_b32 %a row_mirror bound_ctrl:1 + //! v1: %tmp5 = v_mov_b32 %a row_mirror bound_ctrl:1 fi //! v1: %res5 = v_add_f32 %tmp5, %b clamp //! p_unit_test 5, %res5 Temp tmp5 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); auto res5 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp5, b); - res5.instr->vop3().clamp = true; + res5->valu().clamp = true; writeout(5, res5); - //! v1: %res6 = v_add_f32 |%a|, %b row_mirror bound_ctrl:1 + //! v1: %res6 = v_add_f32 |%a|, %b row_mirror bound_ctrl:1 fi //! p_unit_test 6, %res6 auto tmp6 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); - tmp6.instr->dpp().neg[0] = true; + tmp6->dpp16().neg[0] = true; auto res6 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp6, b); - res6.instr->vop3().abs[0] = true; + res6->valu().abs[0] = true; writeout(6, res6); - //! v1: %res7 = v_subrev_f32 %a, |%b| row_mirror bound_ctrl:1 + //! v1: %res7 = v_subrev_f32 %a, |%b| row_mirror bound_ctrl:1 fi //! p_unit_test 7, %res7 Temp tmp7 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); auto res7 = bld.vop2_e64(aco_opcode::v_sub_f32, bld.def(v1), b, tmp7); - res7.instr->vop3().abs[0] = true; + res7->valu().abs[0] = true; writeout(7, res7); + //! v1: %tmp11 = v_mov_b32 -%a row_mirror bound_ctrl:1 fi + //! v1: %res11 = v_add_u32 %tmp11, %b + //! p_unit_test 11, %res11 + auto tmp11 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); + tmp11->dpp16().neg[0] = true; + Temp res11 = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), tmp11, b); + writeout(11, res11); + + //! v1: %tmp12 = v_mov_b32 -%a row_mirror bound_ctrl:1 fi + //! v1: %res12 = v_add_f16 %tmp12, %b + //! p_unit_test 12, %res12 + auto tmp12 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); + tmp12->dpp16().neg[0] = true; + Temp res12 = bld.vop2(aco_opcode::v_add_f16, bld.def(v1), tmp12, b); + writeout(12, res12); + /* vcc */ - //! v1: %res8 = v_cndmask_b32 %a, %b, %c:vcc row_mirror bound_ctrl:1 + //! v1: %res8 = v_cndmask_b32 %a, %b, %c:vcc row_mirror bound_ctrl:1 fi //! p_unit_test 8, %res8 Temp tmp8 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); Temp res8 = bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), tmp8, b, c); writeout(8, res8); + /* sgprs */ + //! v1: %tmp9 = v_mov_b32 %a row_mirror bound_ctrl:1 fi + //! v1: %res9 = v_add_f32 %tmp9, %d + //! p_unit_test 9, %res9 + Temp tmp9 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); + Temp res9 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp9, d); + writeout(9, res9); + + //! v1: %tmp10 = v_mov_b32 %a row_mirror bound_ctrl:1 fi + //! v1: %res10 = v_add_f32 %d, %tmp10 + //! p_unit_test 10, %res10 + Temp tmp10 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); + Temp res10 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), d, tmp10); + writeout(10, res10); + finish_opt_test(); END_TEST @@ -1042,20 +1096,22 @@ BEGIN_TEST(optimize.dpp_prop) Temp one = bld.copy(bld.def(v1), Operand::c32(1)); writeout(0, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), one, inputs[0], dpp_row_sl(1))); - //! v1: %res1 = v_mul_f32 %a, %one row_shl:1 bound_ctrl:1 + //! v1: %res1 = v_mul_f32 %a, %one row_shl:1 bound_ctrl:1 fi //! p_unit_test 1, %res1 writeout(1, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], one, dpp_row_sl(1))); //! v1: %res2 = v_mul_f32 0x12345678, %a //! p_unit_test 2, %res2 Temp literal1 = bld.copy(bld.def(v1), Operand::c32(0x12345678u)); - writeout(2, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), literal1, inputs[0], dpp_row_sl(1))); + writeout(2, + bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), literal1, inputs[0], dpp_row_sl(1))); //! v1: %literal2 = p_parallelcopy 0x12345679 - //! v1: %res3 = v_mul_f32 %a, %literal row_shl:1 bound_ctrl:1 + //! v1: %res3 = v_mul_f32 %a, %literal row_shl:1 bound_ctrl:1 fi //! p_unit_test 3, %res3 Temp literal2 = bld.copy(bld.def(v1), Operand::c32(0x12345679u)); - writeout(3, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], literal2, dpp_row_sl(1))); + writeout(3, + bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], literal2, dpp_row_sl(1))); //! v1: %b_v = p_parallelcopy %b //! v1: %res4 = v_mul_f32 %b, %a @@ -1063,7 +1119,7 @@ BEGIN_TEST(optimize.dpp_prop) Temp b_v = bld.copy(bld.def(v1), inputs[1]); writeout(4, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), b_v, inputs[0], dpp_row_sl(1))); - //! v1: %res5 = v_mul_f32 %a, %b_v row_shl:1 bound_ctrl:1 + //! v1: %res5 = v_mul_f32 %a, %b_v row_shl:1 bound_ctrl:1 fi //! p_unit_test 5, %res5 writeout(5, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], b_v, dpp_row_sl(1))); @@ -1074,3 +1130,1092 @@ BEGIN_TEST(optimize.dpp_prop) finish_opt_test(); END_TEST +BEGIN_TEST(optimize.casts) + //>> v1: %a, v2b: %a16 = p_startpgm + if (!setup_cs("v1 v2b", GFX10_3)) + return; + + Temp a = inputs[0]; + Temp a16 = inputs[1]; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + //! v1: %res0_tmp = v_mul_f32 -1.0, %a + //! v2b: %res0 = v_mul_f16 %res0_tmp, %a16 + //! p_unit_test 0, %res0 + writeout(0, fmul(u2u16(fneg(a)), a16)); + + //! v2b: %res1_tmp = v_mul_f16 -1.0, %a16 + //! v1: %res1 = v_mul_f32 %res1_tmp, %a + //! p_unit_test 1, %res1 + writeout(1, fmul(bld.as_uniform(fneg(a16)), a)); + + //! v1: %res2_tmp = v_mul_f32 -1.0, %a16 + //! v2b: %res2 = v_mul_f16 %res2_tmp, %a16 + //! p_unit_test 2, %res2 + writeout(2, fmul(u2u16(bld.vop2_e64(aco_opcode::v_mul_f32, bld.def(v1), + Operand::c32(0xbf800000u), bld.as_uniform(a16))), + a16)); + + //! v1: %res3_tmp = v_mul_f32 %a, %a + //! v2b: %res3 = v_add_f16 %res3_tmp, 0 clamp + //! p_unit_test 3, %res3 + writeout(3, fsat(u2u16(fmul(a, a)))); + + //! v2b: %res4_tmp = v_mul_f16 %a16, %a16 + //! v1: %res4 = v_add_f32 %res4_tmp, 0 clamp + //! p_unit_test 4, %res4 + writeout(4, fsat(bld.as_uniform(fmul(a16, a16)))); + + //! v1: %res5_tmp = v_mul_f32 %a, %a + //! v2b: %res5 = v_mul_f16 2.0, %res5_tmp + //! p_unit_test 5, %res5 + writeout(5, fmul(u2u16(fmul(a, a)), bld.copy(bld.def(v2b), Operand::c16(0x4000)))); + + //! v2b: %res6_tmp = v_mul_f16 %a16, %a16 + //! v1: %res6 = v_mul_f32 2.0, %res6_tmp + //! p_unit_test 6, %res6 + writeout(6, + fmul(bld.as_uniform(fmul(a16, a16)), bld.copy(bld.def(v1), Operand::c32(0x40000000)))); + + //! v1: %res7_tmp = v_mul_f32 %a, %a + //! v2b: %res7 = v_add_f16 %res7_tmp, %a16 + //! p_unit_test 7, %res7 + writeout(7, fadd(u2u16(fmul(a, a)), a16)); + + //! v2b: %res8_tmp = v_mul_f16 %a16, %a16 + //! v1: %res8 = v_add_f32 %res8_tmp, %a + //! p_unit_test 8, %res8 + writeout(8, fadd(bld.as_uniform(fmul(a16, a16)), a)); + + //! v1: %res9_tmp = v_mul_f32 %a, %a + //! v2b: %res9 = v_mul_f16 -1.0, %res9_tmp + //! p_unit_test 9, %res9 + writeout(9, fneg(u2u16(fmul(a, a)))); + + //! v2b: %res10_tmp = v_mul_f16 %a16, %a16 + //! v1: %res10 = v_mul_f32 -1.0, %res10_tmp + //! p_unit_test 10, %res10 + writeout(10, bld.vop2_e64(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0xbf800000u), + bld.as_uniform(fmul(a16, a16)))); + + finish_opt_test(); +END_TEST + +BEGIN_TEST(optimize.mad_mix.input_conv.basic) + for (unsigned i = GFX9; i <= GFX10; i++) { + //>> v1: %a, v2b: %a16 = p_startpgm + if (!setup_cs("v1 v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a = inputs[0]; + Temp a16 = inputs[1]; + + //! v1: %res0 = v_fma_mix_f32 %a, lo(%a16), -0 + //! p_unit_test 0, %res0 + writeout(0, fmul(a, f2f32(a16))); + + //! v1: %res1 = v_fma_mix_f32 1.0, %a, lo(%a16) + //! p_unit_test 1, %res1 + writeout(1, fadd(a, f2f32(a16))); + + //! v1: %res2 = v_fma_mix_f32 1.0, lo(%a16), %a + //! p_unit_test 2, %res2 + writeout(2, fadd(f2f32(a16), a)); + + //! v1: %res3 = v_fma_mix_f32 %a, %a, lo(%a16) + //! p_unit_test 3, %res3 + writeout(3, fma(a, a, f2f32(a16))); + + //! v1: %res4 = v_fma_mix_f32 %a, %a, lo(%a16) + //! p_unit_test 4, %res4 + writeout(4, fma(a, a, f2f32(a16))); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.mad_mix.input_conv.precision) + for (unsigned i = GFX9; i <= GFX10; i++) { + //>> v1: %a, v2b: %a16 = p_startpgm + if (!setup_cs("v1 v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a = inputs[0]; + Temp a16 = inputs[1]; + + /* precise arithmetic */ + //~gfx9! v1: %res0_cvt = v_cvt_f32_f16 %a16 + //~gfx9! v1: (precise)%res0 = v_fma_f32 %a, %a, %res0_cvt + //~gfx10! v1: (precise)%res0 = v_fma_mix_f32 %a, %a, lo(%a16) + //! p_unit_test 0, %res0 + writeout(0, fma(a, a, f2f32(a16), bld.precise())); + + //! v2b: %res1_cvt = v_cvt_f16_f32 %a + //! v2b: (precise)%res1 = v_mul_f16 %a16, %res1_cvt + //! p_unit_test 1, %res1 + writeout(1, fmul(a16, f2f16(a), bld.precise())); + + //! v2b: %res2_cvt = v_cvt_f16_f32 %a + //! v2b: (precise)%res2 = v_add_f16 %a16, %res2_cvt + //! p_unit_test 2, %res2 + writeout(2, fadd(a16, f2f16(a), bld.precise())); + + //! v2b: %res3_cvt = v_cvt_f16_f32 %a + //! v2b: (precise)%res3 = v_fma_f16 %a16, %a16, %res3_cvt + //! p_unit_test 3, %res3 + writeout(3, fma(a16, a16, f2f16(a), bld.precise())); + + /* precise conversions */ + //! v2b: (precise)%res4_cvt = v_cvt_f16_f32 %a + //! v2b: %res4 = v_mul_f16 %a16, %res4_cvt + //! p_unit_test 4, %res4 + writeout(4, fmul(a16, f2f16(a, bld.precise()))); + + //! v2b: (precise)%res5_cvt = v_cvt_f16_f32 %a + //! v2b: %res5 = v_add_f16 %a16, %res5_cvt + //! p_unit_test 5, %res5 + writeout(5, fadd(a16, f2f16(a, bld.precise()))); + + //! v2b: (precise)%res6_cvt = v_cvt_f16_f32 %a + //! v2b: %res6 = v_fma_f16 %a16, %a16, %res6_cvt + //! p_unit_test 6, %res6 + writeout(6, fma(a16, a16, f2f16(a, bld.precise()))); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.mad_mix.input_conv.modifiers) + for (unsigned i = GFX9; i <= GFX11; i++) { + if (i == GFX10_3) + continue; + //>> v1: %a, v2b: %a16 = p_startpgm + if (!setup_cs("v1 v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a = inputs[0]; + Temp a16 = inputs[1]; + + /* check whether modifiers are preserved when converting to VOP3P */ + //! v1: %res0 = v_fma_mix_f32 -%a, lo(%a16), -0 + //! p_unit_test 0, %res0 + writeout(0, fmul(fneg(a), f2f32(a16))); + + //! v1: %res1 = v_fma_mix_f32 |%a|, lo(%a16), -0 + //! p_unit_test 1, %res1 + writeout(1, fmul(fabs(a), f2f32(a16))); + + /* fneg modifiers */ + //! v1: %res2 = v_fma_mix_f32 %a, -lo(%a16), -0 + //! p_unit_test 2, %res2 + writeout(2, fmul(a, fneg(f2f32(a16)))); + + //! v1: %res3 = v_fma_mix_f32 %a, -lo(%a16), -0 + //! p_unit_test 3, %res3 + writeout(3, fmul(a, f2f32(fneg(a16)))); + + /* fabs modifiers */ + //! v1: %res4 = v_fma_mix_f32 %a, |lo(%a16)|, -0 + //! p_unit_test 4, %res4 + writeout(4, fmul(a, fabs(f2f32(a16)))); + + //! v1: %res5 = v_fma_mix_f32 %a, |lo(%a16)|, -0 + //! p_unit_test 5, %res5 + writeout(5, fmul(a, f2f32(fabs(a16)))); + + /* both fabs and fneg modifiers */ + //! v1: %res6 = v_fma_mix_f32 %a, -|lo(%a16)|, -0 + //! p_unit_test 6, %res6 + writeout(6, fmul(a, fneg(f2f32(fabs(a16))))); + + //! v1: %res7 = v_fma_mix_f32 %a, |lo(%a16)|, -0 + //! p_unit_test 7, %res7 + writeout(7, fmul(a, fabs(f2f32(fabs(a16))))); + + //! v1: %res8 = v_fma_mix_f32 %a, -|lo(%a16)|, -0 + //! p_unit_test 8, %res8 + writeout(8, fmul(a, fneg(fabs(f2f32(fabs(a16)))))); + + //! v1: %res9 = v_fma_mix_f32 %a, -|lo(%a16)|, -0 + //! p_unit_test 9, %res9 + writeout(9, fmul(a, f2f32(fneg(fabs(a16))))); + + //! v1: %res10 = v_fma_mix_f32 %a, |lo(%a16)|, -0 + //! p_unit_test 10, %res10 + writeout(10, fmul(a, fneg(f2f32(fneg(fabs(a16)))))); + + //! v1: %res11 = v_fma_mix_f32 %a, |lo(%a16)|, -0 + //! p_unit_test 11, %res11 + writeout(11, fmul(a, fabs(f2f32(fneg(fabs(a16)))))); + + //! v1: %res12 = v_fma_mix_f32 %a, -|lo(%a16)|, -0 + //! p_unit_test 12, %res12 + writeout(12, fmul(a, fneg(fabs(f2f32(fneg(fabs(a16))))))); + + /* sdwa */ + //! v1: %res13 = v_fma_mix_f32 lo(%a), %a, -0 + //! p_unit_test 13, %res13 + writeout(13, fmul(f2f32(ext_ushort(a, 0)), a)); + + //! v1: %res14 = v_fma_mix_f32 hi(%a), %a, -0 + //! p_unit_test 14, %res14 + writeout(14, fmul(f2f32(ext_ushort(a, 1)), a)); + + //~gfx(9|10)! v1: %res15_cvt = v_cvt_f32_f16 %a dst_sel:uword0 src0_sel:dword + //~gfx11! v1: %res16_cvt1 = v_fma_mix_f32 lo(%a), 1.0, -0 + //~gfx11! v1: %res15_cvt = p_extract %res16_cvt1, 0, 16, 0 + //! v1: %res15 = v_mul_f32 %res15_cvt, %a + //! p_unit_test 15, %res15 + writeout(15, fmul(ext_ushort(f2f32(a), 0), a)); + + //~gfx(9|10)! v1: %res16_cvt = v_cvt_f32_f16 %a + //~gfx(9|10)! v1: %res16 = v_mul_f32 %res16_cvt, %a dst_sel:dword src0_sel:uword1 src1_sel:dword + //~gfx11! v1: %res16_cvt = v_fma_mix_f32 lo(%a), 1.0, -0 + //~gfx11! v1: %res16_ext = p_extract %res16_cvt, 1, 16, 0 + //~gfx11! v1: %res16 = v_mul_f32 %res16_ext, %a + //! p_unit_test 16, %res16 + writeout(16, fmul(ext_ushort(f2f32(a), 1), a)); + + //~gfx(9|10)! v1: %res17_cvt = v_cvt_f32_f16 %a dst_sel:dword src0_sel:ubyte2 + //~gfx(9|10)! v1: %res17 = v_mul_f32 %res17_cvt, %a + //~gfx11! v1: %res17_ext = p_extract %a, 2, 8, 0 + //~gfx11! v1: %res17 = v_fma_mix_f32 lo(%res17_ext), %a, -0 + //! p_unit_test 17, %res17 + writeout(17, fmul(f2f32(ext_ubyte(a, 2)), a)); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.mad_mix.output_conv.basic) + for (unsigned i = GFX9; i <= GFX10; i++) { + //>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %b16 = p_startpgm + if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a = inputs[0]; + Temp b = inputs[1]; + Temp c = inputs[2]; + Temp a16 = inputs[3]; + Temp b16 = inputs[4]; + + //! v2b: %res0 = v_fma_mixlo_f16 %a, %b, -0 + //! p_unit_test 0, %res0 + writeout(0, f2f16(fmul(a, b))); + + //! v2b: %res1 = v_fma_mixlo_f16 1.0, %a, %b + //! p_unit_test 1, %res1 + writeout(1, f2f16(fadd(a, b))); + + //! v2b: %res2 = v_fma_mixlo_f16 %a, %b, %c + //! p_unit_test 2, %res2 + writeout(2, f2f16(fma(a, b, c))); + + //! v2b: %res3 = v_fma_mixlo_f16 lo(%a16), %b, -0 + //! p_unit_test 3, %res3 + writeout(3, f2f16(fmul(f2f32(a16), b))); + + //! v2b: %res4 = v_fma_mixlo_f16 1.0, %a, lo(%b16) + //! p_unit_test 4, %res4 + writeout(4, f2f16(fadd(a, f2f32(b16)))); + + //! v2b: %res5 = v_fma_mixlo_f16 %a, lo(%b16), %c + //! p_unit_test 5, %res5 + writeout(5, f2f16(fma(a, f2f32(b16), c))); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.mad_mix.output_conv.precision) + for (unsigned i = GFX9; i <= GFX10; i++) { + //>> v2b: %a16 = p_startpgm + if (!setup_cs("v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a16 = inputs[0]; + + //! v2b: %res0_tmp = v_mul_f16 %a16, %a16 + //! v1: (precise)%res0 = v_cvt_f32_f16 %res0_tmp + //! p_unit_test 0, %res0 + writeout(0, f2f32(fmul(a16, a16), bld.precise())); + + //! v2b: (precise)%res1_tmp = v_mul_f16 %a16, %a16 + //! v1: %res1 = v_cvt_f32_f16 %res1_tmp + //! p_unit_test 1, %res1 + writeout(1, f2f32(fmul(a16, a16, bld.precise()))); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.mad_mix.output_conv.modifiers) + for (unsigned i = GFX9; i <= GFX10; i++) { + //>> v1: %a, v1: %b, v2b: %a16, v2b: %b16 = p_startpgm + if (!setup_cs("v1 v1 v2b v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a = inputs[0]; + Temp b = inputs[1]; + Temp a16 = inputs[2]; + Temp b16 = inputs[3]; + + /* fneg/fabs */ + //! v1: %res0_add = v_add_f32 %1, %2 + //! v2b: %res0 = v_cvt_f16_f32 |%res0_add| + //! p_unit_test 0, %res0 + writeout(0, f2f16(fabs(fadd(a, b)))); + + //! v1: %res1_add = v_add_f32 %1, %2 + //! v2b: %res1 = v_cvt_f16_f32 -%res1_add + //! p_unit_test 1, %res1 + writeout(1, f2f16(fneg(fadd(a, b)))); + + //! v2b: %res2_add = v_add_f16 %3, %4 + //! v1: %res2 = v_cvt_f32_f16 |%res2_add| + //! p_unit_test 2, %res2 + writeout(2, f2f32(fabs(fadd(a16, b16)))); + + //! v2b: %res3_add = v_add_f16 %3, %4 + //! v1: %res3 = v_cvt_f32_f16 -%res3_add + //! p_unit_test 3, %res3 + writeout(3, f2f32(fneg(fadd(a16, b16)))); + + /* sdwa */ + //! v2b: %res4_add = v_fma_mixlo_f16 1.0, %a, %b + //! v2b: %res4 = p_extract %res4_add, 0, 8, 0 + //! p_unit_test 4, %res4 + writeout(4, ext_ubyte(f2f16(fadd(a, b)), 0)); + + //! v1: %res5_mul = v_add_f32 %a, %b dst_sel:uword0 src0_sel:dword src1_sel:dword + //! v2b: %res5 = v_cvt_f16_f32 %res5_mul + //! p_unit_test 5, %res5 + writeout(5, f2f16(ext_ushort(fadd(a, b), 0))); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.mad_mix.fma.basic) + for (unsigned i = GFX9; i <= GFX10; i++) { + //>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %c16 = p_startpgm + if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a = inputs[0]; + Temp b = inputs[1]; + Temp c = inputs[2]; + Temp a16 = inputs[3]; + Temp c16 = inputs[4]; + + //! v1: %res0 = v_fma_mix_f32 lo(%a16), %b, %c + //! p_unit_test 0, %res0 + writeout(0, fadd(fmul(f2f32(a16), b), c)); + + //! v1: %res1 = v_fma_mix_f32 %a, %b, lo(%c16) + //! p_unit_test 1, %res1 + writeout(1, fadd(fmul(a, b), f2f32(c16))); + + /* omod/clamp check */ + //! v1: %res2_mul = v_fma_mix_f32 lo(%a16), %b, -0 + //! v1: %res2 = v_add_f32 %res2_mul, %c *2 + //! p_unit_test 2, %res2 + writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000), + fadd(fmul(f2f32(a16), b), c))); + + /* neg/abs modifiers */ + //! v1: %res3 = v_fma_mix_f32 -lo(%a16), %b, |lo(%c16)| + //! p_unit_test 3, %res3 + writeout(3, fadd(fmul(fneg(f2f32(a16)), b), fabs(f2f32(c16)))); + + //! v1: %res4 = v_fma_mix_f32 |%a|, |%b|, lo(%c16) + //! p_unit_test 4, %res4 + writeout(4, fadd(fabs(fmul(fneg(a), fneg(b))), f2f32(c16))); + + //! v1: %res5 = v_fma_mix_f32 %a, -%b, lo(%c16) + //! p_unit_test 5, %res5 + writeout(5, fadd(fneg(fmul(a, b)), f2f32(c16))); + + //! v1: %res6 = v_fma_mix_f32 |%a|, -|%b|, lo(%c16) + //! p_unit_test 6, %res6 + writeout(6, fadd(fneg(fabs(fmul(fneg(a), fneg(b)))), f2f32(c16))); + + /* output conversions */ + //! v2b: %res7 = v_fma_mixlo_f16 %a, %b, %c + //! p_unit_test 7, %res7 + writeout(7, f2f16(fadd(fmul(a, b), c))); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.mad_mix.fma.precision) + for (unsigned i = GFX9; i <= GFX10; i++) { + //>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %b16 = p_startpgm + if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a = inputs[0]; + Temp b = inputs[1]; + Temp c = inputs[2]; + Temp a16 = inputs[3]; + Temp b16 = inputs[4]; + + /* the optimization is precise for 32-bit on GFX9 */ + //~gfx9! v1: (precise)%res0 = v_fma_mix_f32 lo(%a16), %b, %c + //~gfx10! v1: (precise)%res0_tmp = v_fma_mix_f32 lo(%a16), %b, -0 + //~gfx10! v1: %res0 = v_add_f32 %res0_tmp, %c + //! p_unit_test 0, %res0 + writeout(0, fadd(fmul(f2f32(a16), b, bld.precise()), c)); + + //~gfx9! v1: (precise)%res1 = v_fma_mix_f32 lo(%a16), %b, %c + //~gfx10! v1: %res1_tmp = v_fma_mix_f32 lo(%a16), %b, -0 + //~gfx10! v1: (precise)%res1 = v_add_f32 %res1_tmp, %c + //! p_unit_test 1, %res1 + writeout(1, fadd(fmul(f2f32(a16), b), c, bld.precise())); + + /* never promote 16-bit arithmetic to 32-bit */ + //! v2b: %res2_tmp = v_cvt_f16_f32 %a + //! v2b: %res2 = v_add_f16 %res2_tmp, %b16 + //! p_unit_test 2, %res2 + writeout(2, fadd(f2f16(a), b16)); + + //! v2b: %res3_tmp = v_cvt_f16_f32 %a + //! v2b: %res3 = v_mul_f16 %res3_tmp, %b16 + //! p_unit_test 3, %res3 + writeout(3, fmul(f2f16(a), b16)); + + //! v2b: %res4_tmp = v_mul_f16 %a16, %b16 + //! v1: %res4 = v_cvt_f32_f16 %res4_tmp + //! p_unit_test 4, %res4 + writeout(4, f2f32(fmul(a16, b16))); + + //! v2b: %res5_tmp = v_add_f16 %a16, %b16 + //! v1: %res5 = v_cvt_f32_f16 %res5_tmp + //! p_unit_test 5, %res5 + writeout(5, f2f32(fadd(a16, b16))); + + //! v2b: %res6_tmp = v_fma_mixlo_f16 %a, %b, -0 + //! v2b: %res6 = v_add_f16 %res6_tmp, %a16 + //! p_unit_test 6, %res6 + writeout(6, fadd(f2f16(fmul(a, b)), a16)); + + //! v2b: %res7_tmp = v_mul_f16 %a16, %b16 + //! v1: %res7 = v_fma_mix_f32 1.0, lo(%res7_tmp), %c + //! p_unit_test 7, %res7 + writeout(7, fadd(f2f32(fmul(a16, b16)), c)); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.mad_mix.clamp) + for (unsigned i = GFX9; i <= GFX10; i++) { + //>> v1: %a, v2b: %a16 = p_startpgm + if (!setup_cs("v1 v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a = inputs[0]; + Temp a16 = inputs[1]; + + //! v1: %res0 = v_fma_mix_f32 lo(%a16), %a, -0 clamp + //! p_unit_test 0, %res0 + writeout(0, fsat(fmul(f2f32(a16), a))); + + //! v2b: %res1 = v_fma_mixlo_f16 %a, %a, -0 clamp + //! p_unit_test 1, %res1 + writeout(1, f2f16(fsat(fmul(a, a)))); + + //! v2b: %res2 = v_fma_mixlo_f16 %a, %a, -0 clamp + //! p_unit_test 2, %res2 + writeout(2, fsat(f2f16(fmul(a, a)))); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.mad_mix.cast) + for (unsigned i = GFX9; i <= GFX10; i++) { + //>> v1: %a, v2b: %a16 = p_startpgm + if (!setup_cs("v1 v2b", (amd_gfx_level)i)) + continue; + + program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; + + Temp a = inputs[0]; + Temp a16 = inputs[1]; + + /* The optimizer copy-propagates v2b=p_extract_vector(v1, 0) and p_as_uniform, so the + * optimizer has to check compatibility. + */ + + //! v1: %res0_cvt = v_cvt_f32_f16 %a16 + //! v2b: %res0 = v_mul_f16 %res0_cvt, %a16 + //! p_unit_test 0, %res0 + writeout(0, fmul(u2u16(f2f32(a16)), a16)); + + //! v2b: %res1_cvt = v_cvt_f16_f32 %a + //! v1: %res1 = v_mul_f32 %res1_cvt, %a + //! p_unit_test 1, %res1 + writeout(1, fmul(bld.as_uniform(f2f16(a)), a)); + + //! v2b: %res2_mul = v_mul_f16 %a16, %a16 + //! v2b: %res2 = v_cvt_f16_f32 %res2_mul + //! p_unit_test 2, %res2 + writeout(2, f2f16(bld.as_uniform(fmul(a16, a16)))); + + //! v1: %res3_mul = v_mul_f32 %a, %a + //! v1: %res3 = v_cvt_f32_f16 %res3_mul + //! p_unit_test 3, %res3 + writeout(3, f2f32(u2u16(fmul(a, a)))); + + //! v1: %res4_mul = v_fma_mix_f32 lo(%a16), %a, -0 + //! v2b: %res4 = v_add_f16 %res4_mul, 0 clamp + //! p_unit_test 4, %res4 + writeout(4, fsat(u2u16(fmul(f2f32(a16), a)))); + + //! v2b: %res5_mul = v_fma_mixlo_f16 %a, %a, -0 + //! v1: %res5 = v_add_f32 %res5_mul, 0 clamp + //! p_unit_test 5, %res5 + writeout(5, fsat(bld.as_uniform(f2f16(fmul(a, a))))); + + //! v1: %res6_mul = v_mul_f32 %a, %a + //! v1: %res6 = v_fma_mix_f32 1.0, lo(%res6_mul), %a + //! p_unit_test 6, %res6 + writeout(6, fadd(f2f32(u2u16(fmul(a, a))), a)); + + //! v2b: %res7_mul = v_mul_f16 %a16, %a16 + //! v1: %res7 = v_fma_mix_f32 1.0, %res7_mul, lo(%a16) + //! p_unit_test 7, %res7 + writeout(7, fadd(bld.as_uniform(fmul(a16, a16)), f2f32(a16))); + + /* opsel_hi should be obtained from the original opcode, not the operand regclass */ + //! v1: %res8 = v_fma_mix_f32 lo(%a16), %a16, -0 + //! p_unit_test 8, %res8 + writeout(8, fmul(f2f32(a16), a16)); + + finish_opt_test(); + } +END_TEST + +static void +vop3p_constant(unsigned* idx, aco_opcode op, const char* swizzle, uint32_t val) +{ + uint32_t halves[2] = {val & 0xffff, val >> 16}; + uint32_t expected = halves[swizzle[0] - 'x'] | (halves[swizzle[1] - 'x'] << 16); + fprintf(output, "Expected for %u: 0x%.8x / %u\n", *idx, expected, expected); + + unsigned opsel_lo = swizzle[0] == 'x' ? 0x0 : 0x1; + unsigned opsel_hi = swizzle[1] == 'x' ? 0x2 : 0x3; + writeout((*idx)++, bld.vop3p(op, bld.def(v1), bld.copy(bld.def(v1), Operand::c32(val)), + inputs[0], opsel_lo, opsel_hi)); +} + +BEGIN_TEST(optimize.vop3p_constants) + for (aco_opcode op : {aco_opcode::v_pk_add_f16, aco_opcode::v_pk_add_u16}) { + for (const char* swizzle : {"xx", "yy", "xy", "yx"}) { + char variant[16]; + strcpy(variant, op == aco_opcode::v_pk_add_f16 ? "_f16" : "_u16"); + strcat(variant, "_"); + strcat(variant, swizzle); + + //; for i in range(36): + //; insert_pattern('Expected for %u: $_ / #expected%u' % (i, i)) + + //>> v1: %a = p_startpgm + if (!setup_cs("v1", GFX10_3, CHIP_UNKNOWN, variant)) + continue; + + //; opcode = 'v_pk_add_u16' if 'u16' in variant else 'v_pk_add_f16' + //; for i in range(36): + //; insert_pattern('v1: %%res%u = %s $got%u %%a' % (i, opcode, i)) + //; insert_pattern('p_unit_test %u, %%res%u' % (i, i)) + //! s_endpgm + + //; def parse_op(op): + //; is_int = opcode == 'v_pk_add_u16' + //; op = op.rstrip(',') + //; + //; mods = lambda v: v + //; if op.endswith('*[1,-1]'): + //; mods = lambda v: v ^ 0x80000000 + //; assert(not is_int) + //; elif op.endswith('*[-1,1]'): + //; mods = lambda v: v ^ 0x00008000 + //; assert(not is_int) + //; op = op.split('*')[0] + //; + //; swizzle = lambda v: v + //; if op.endswith('.xx'): + //; swizzle = lambda v: ((v & 0xffff) | (v << 16)) & 0xffffffff; + //; elif op.endswith('.yy'): + //; swizzle = lambda v: (v >> 16) | (v & 0xffff0000); + //; elif op.endswith('.yx'): + //; swizzle = lambda v: ((v >> 16) | (v << 16)) & 0xffffffff; + //; op = op.rstrip('xy.') + //; + //; val = None + //; if op.startswith('0x'): + //; val = int(op[2:], 16) + //; elif op == '-1.0': + //; val = 0xbf800000 if is_int else 0xbC00 + //; elif op == '1.0': + //; val = 0x3f800000 if is_int else 0x3c00 + //; else: + //; val = int(op) & 0xffffffff + //; + //; return mods(swizzle(val)) + + //; # Check correctness + //; for i in range(36): + //; expected = globals()['expected%u' % i] + //; got = globals()['got%u' % i] + //; got_parsed = parse_op(got) + //; if got_parsed != expected: + //; raise Exception('Check %u failed: expected 0x%.8x, got 0x%.8x ("%s")' % (i, expected, got_parsed, got)) + + //; # Check that all literals are ones that cannot be encoded as inline constants + //; allowed_literals = [0x00004242, 0x0000fffe, 0x00308030, 0x0030ffff, 0x3c00ffff, + //; 0x42420000, 0x42424242, 0x4242c242, 0x4242ffff, 0x7ffefffe, + //; 0x80300030, 0xbeefdead, 0xc2424242, 0xdeadbeef, 0xfffe0000, + //; 0xfffe7ffe, 0xffff0030, 0xffff3c00, 0xffff4242] + //; if opcode == 'v_pk_add_u16': + //; allowed_literals.extend([0x00003c00, 0x3c000000, 0x3c003c00, 0x3c00bc00, 0xbc003c00]) + //; else: + //; allowed_literals.extend([0x00003f80, 0x3f800000]) + //; + //; for i in range(36): + //; got = globals()['got%u' % i] + //; if not got.startswith('0x'): + //; continue; + //; got = int(got[2:].rstrip(',').split('*')[0].split('.')[0], 16) + //; if got not in allowed_literals: + //; raise Exception('Literal check %u failed: 0x%.8x not in allowed literals' % (i, got)) + + unsigned idx = 0; + for (uint32_t constant : {0x3C00, 0x0030, 0xfffe, 0x4242}) { + vop3p_constant(&idx, op, swizzle, constant); + vop3p_constant(&idx, op, swizzle, constant | 0xffff0000); + vop3p_constant(&idx, op, swizzle, constant | (constant << 16)); + vop3p_constant(&idx, op, swizzle, constant << 16); + vop3p_constant(&idx, op, swizzle, (constant << 16) | 0x0000ffff); + vop3p_constant(&idx, op, swizzle, constant | ((constant ^ 0x8000) << 16)); + vop3p_constant(&idx, op, swizzle, (constant ^ 0x8000) | (constant << 16)); + } + + for (uint32_t constant : {0x3f800000u, 0xfffffffeu, 0x00000030u, 0xdeadbeefu}) { + uint32_t lo = constant & 0xffff; + uint32_t hi = constant >> 16; + vop3p_constant(&idx, op, swizzle, constant); + vop3p_constant(&idx, op, swizzle, hi | (lo << 16)); + } + + finish_opt_test(); + } + } +END_TEST + +BEGIN_TEST(optimize.fmamix_two_literals) + /* This test has to recreate literals sometimes because we don't combine them at all if there's + * at least one uncombined use. + */ + for (unsigned i = GFX10; i <= GFX10_3; i++) { + //>> v1: %a, v1: %b = p_startpgm + if (!setup_cs("v1 v1", (amd_gfx_level)i)) + continue; + + Temp a = inputs[0]; + Temp b = inputs[1]; + + Temp c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f))); + Temp c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f))); + Temp c_denorm = bld.copy(bld.def(v1), Operand::c32(0x387fc000)); + + //! v1: %res0 = v_fma_mix_f32 %a, lo(0x42003e00), hi(0x42003e00) + //! p_unit_test 0, %res0 + writeout(0, fma(a, c15, c30)); + + /* No need to use v_fma_mix_f32. */ + //! v1: %res1 = v_fmaak_f32 %a, %b, 0x40400000 + //! p_unit_test 1, %res1 + writeout(1, fma(a, b, c30)); + + /* Separate mul/add can become v_fma_mix_f32 if it's not precise. */ + //! v1: %res2 = v_fma_mix_f32 %a, lo(0x42003e00), hi(0x42003e00) + //! p_unit_test 2, %res2 + writeout(2, fadd(fmul(a, c15), c30)); + + //~gfx10! v1: %c15 = p_parallelcopy 0x3fc00000 + c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f))); + c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f))); + + /* v_fma_mix_f32 is a fused mul/add, so it can't be used for precise separate mul/add. */ + //~gfx10! v1: (precise)%res3 = v_madak_f32 %a, %c15, 0x40400000 + //~gfx10_3! v1: (precise)%res3_tmp = v_mul_f32 %a, 0x3fc00000 + //~gfx10_3! v1: %res3 = v_add_f32 %res3_tmp, 0x40400000 + //! p_unit_test 3, %res3 + writeout(3, fadd(bld.precise().vop2(aco_opcode::v_mul_f32, bld.def(v1), a, c15), c30)); + + //~gfx10! v1: (precise)%res4 = v_madak_f32 %1, %c16, 0x40400000 + //~gfx10_3! v1: %res4_tmp = v_mul_f32 %a, 0x3fc00000 + //~gfx10_3! v1: (precise)%res4 = v_add_f32 %res4_tmp, 0x40400000 + //! p_unit_test 4, %res4 + writeout(4, bld.precise().vop2(aco_opcode::v_add_f32, bld.def(v1), fmul(a, c15), c30)); + + /* Can't convert to fp16 if it will be flushed as a denormal. */ + //! v1: %res5 = v_fma_mix_f32 %1, lo(0x3ff3e00), hi(0x3ff3e00) + //! p_unit_test 5, %res5 + c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f))); + writeout(5, fma(a, c15, c_denorm)); + + //>> BB1 + //! /* logical preds: / linear preds: / kind: uniform, */ + program->next_fp_mode.denorm16_64 = fp_denorm_flush; + bld.reset(program->create_and_insert_block()); + + //~gfx10; del c15 + //! v1: %c15 = p_parallelcopy 0x3fc00000 + //! v1: %res6 = v_fmaak_f32 %a, %c15, 0x387fc000 + //! p_unit_test 6, %res6 + c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f))); + c_denorm = bld.copy(bld.def(v1), Operand::c32(0x387fc000)); + writeout(6, fma(a, c15, c_denorm)); + + /* Can't accept more than 3 unique fp16 literals. */ + //! v1: %c45 = p_parallelcopy 0x40900000 + //! v1: %res7 = v_fma_mix_f32 lo(0x42003e00), hi(0x42003e00), %c45 + //! p_unit_test 7, %res7 + Temp c45 = bld.copy(bld.def(v1), Operand::c32(fui(4.5f))); + writeout(7, fma(c15, c30, c45)); + + /* Modifiers must be preserved. */ + //! v1: %res8 = v_fma_mix_f32 -%a, lo(0x44804200), hi(0x44804200) + //! p_unit_test 8, %res8 + writeout(8, fma(fneg(a), c30, c45)); + + //! v1: %res9 = v_fma_mix_f32 lo(0x44804200), |%a|, hi(0x44804200) + //! p_unit_test 9, %res9 + writeout(9, fma(c30, fabs(a), c45)); + + //! v1: %res10 = v_fma_mix_f32 %a, lo(0x44804200), hi(0x44804200) clamp + //! p_unit_test 10, %res10 + writeout(10, fsat(fma(a, c30, c45))); + + /* Output modifiers are not supported by v_fma_mix_f32. */ + c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f))); + //; del c45 + //! v1: %c45 = p_parallelcopy 0x40900000 + //! v1: %res11 = v_fma_f32 %a, 0x40400000, %c45 *0.5 + //! p_unit_test 11, %res11 + c45 = bld.copy(bld.def(v1), Operand::c32(fui(4.5f))); + writeout(11, fmul(fma(a, c30, c45), bld.copy(bld.def(v1), Operand::c32(0x3f000000)))); + + /* Has a literal which can't be represented as fp16. */ + //! v1: %c03 = p_parallelcopy 0x3e99999a + //! v1: %res12 = v_fmaak_f32 %a, %c03, 0x40400000 + //! p_unit_test 12, %res12 + Temp c03 = bld.copy(bld.def(v1), Operand::c32(fui(0.3f))); + writeout(12, fma(a, c03, c30)); + + /* We should still use fmaak/fmamk if the two literals are identical. */ + //! v1: %res13 = v_fmaak_f32 0x40400000, %a, 0x40400000 + //! p_unit_test 13, %res13 + writeout(13, fma(a, c30, c30)); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.fma_opsel) + /* TODO make these work before GFX11 using SDWA. */ + for (unsigned i = GFX11; i <= GFX11; i++) { + //>> v2b: %a, v2b: %b, v1: %c, v1: %d, v1: %e = p_startpgm + if (!setup_cs("v2b v2b v1 v1 v1", (amd_gfx_level)i)) + continue; + + Temp a = inputs[0]; + Temp b = inputs[1]; + Temp c = inputs[2]; + Temp d = inputs[3]; + Temp e = inputs[4]; + Temp c_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), c, Operand::c32(1)); + Temp d_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), d, Operand::c32(1)); + Temp e_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), e, Operand::c32(1)); + + //! v2b: %res0 = v_fma_f16 %b, hi(%c), %a + //! p_unit_test 0, %res0 + writeout(0, fadd(fmul(b, c_hi), a)); + + //! v2b: %res1 = v_fma_f16 %a, %b, hi(%d) + //! p_unit_test 1, %res1 + writeout(1, fadd(fmul(a, b), d_hi)); + + //! v2b: %res2 = v_fma_f16 %a, %b, hi(%e) + //! p_unit_test 2, %res2 + writeout(2, fma(a, b, e_hi)); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.dpp_opsel) + //>> v1: %a, v1: %b = p_startpgm + if (!setup_cs("v1 v1", GFX11)) + return; + + Temp a = inputs[0]; + Temp b = inputs[1]; + + Temp dpp16 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); + Temp dpp16_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), dpp16, Operand::c32(1)); + Temp dpp8 = bld.vop1_dpp8(aco_opcode::v_mov_b32, bld.def(v1), a); + Temp dpp8_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), dpp8, Operand::c32(1)); + + Temp b_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(1)); + Temp b_lo = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(0)); + + //! v2b: %res0 = v_add_f16 hi(%a), hi(%b) row_mirror bound_ctrl:1 fi + //! p_unit_test 0, %res0 + writeout(0, fadd(dpp16_hi, b_hi)); + + //! v2b: %res1 = v_add_f16 hi(%a), %b dpp8:[0,0,0,0,0,0,0,0] fi + //! p_unit_test 1, %res1 + writeout(1, fadd(b_lo, dpp8_hi)); + + finish_opt_test(); +END_TEST + +BEGIN_TEST(optimize.apply_sgpr_swap_opsel) + //>> v1: %a, s1: %b = p_startpgm + if (!setup_cs("v1 s1", GFX11)) + return; + + Temp a = inputs[0]; + Temp b = inputs[1]; + + Temp b_vgpr = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), bld.copy(bld.def(v1), b), + Operand::c32(0)); + + Temp res0 = bld.tmp(v2b); + VALU_instruction& valu = bld.vop2(aco_opcode::v_sub_f16, Definition(res0), a, b_vgpr)->valu(); + valu.opsel[0] = true; + + //! v2b: %res0 = v_subrev_f16 %b, hi(%a) + //! p_unit_test 0, %res0 + writeout(0, res0); + + finish_opt_test(); +END_TEST + +BEGIN_TEST(optimize.combine_comparison_ordering) + //>> v1: %a, v1: %b, v1: %c = p_startpgm + if (!setup_cs("v1 v1 v1", GFX11)) + return; + + Temp a = inputs[0]; + Temp b = inputs[1]; + Temp c = inputs[2]; + + Temp a_unordered = bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), a, a); + Temp b_unordered = bld.vopc(aco_opcode::v_cmp_neq_f32, bld.def(bld.lm), b, b); + Temp unordered = + bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), a_unordered, b_unordered); + + Temp a_lt_a = bld.vopc(aco_opcode::v_cmp_lt_f32, bld.def(bld.lm), a, a); + Temp unordered_cmp = + bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), unordered, a_lt_a); + + //! s2: %res0_unordered = v_cmp_u_f32 %a, %b + //! s2: %res0_cmp = v_cmp_lt_f32 %a, %a + //! s2: %res0, s2: %_:scc = s_or_b64 %res0_unordered, %res0_cmp + //! p_unit_test 0, %res0 + writeout(0, unordered_cmp); + + Temp c_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), c, Operand::c32(1)); + + Temp c_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), c, c); + Temp c_hi_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), c_hi, c_hi); + unordered = + bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), c_unordered, c_hi_unordered); + + Temp c_lt_c_hi = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), c, c_hi); + unordered_cmp = + bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), unordered, c_lt_c_hi); + + //! s2: %res1 = v_cmp_nge_f16 %c, hi(%c) + //! p_unit_test 1, %res1 + writeout(1, unordered_cmp); + + finish_opt_test(); +END_TEST + +BEGIN_TEST(optimize.combine_comparison_ordering_opsel) + //>> v1: %a, v2b: %b = p_startpgm + if (!setup_cs("v1 v2b", GFX11)) + return; + + Temp a = inputs[0]; + Temp b = inputs[1]; + + Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1)); + + Temp ahi_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a_hi, a_hi); + Temp b_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), b, b); + Temp unordered = + bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), ahi_unordered, b_unordered); + + Temp ahi_lt_b = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), a_hi, b); + Temp unordered_cmp = + bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), unordered, ahi_lt_b); + + //! s2: %res0 = v_cmp_nge_f16 hi(%a), %b + //! p_unit_test 0, %res0 + writeout(0, unordered_cmp); + + Temp ahi_cmp_const = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), a_hi, + bld.copy(bld.def(v2b), Operand::c16(0x4400))); + Temp ahi_ucmp_const = + bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), ahi_unordered, ahi_cmp_const); + //! s2: %res1 = v_cmp_nle_f16 4.0, hi(%a) + //! p_unit_test 1, %res1 + writeout(1, ahi_ucmp_const); + + a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1)); + ahi_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a_hi, a_hi); + b_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), b, b); + unordered = + bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), ahi_unordered, b_unordered); + Temp alo_lt_b = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), a, b); + Temp noopt = bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), unordered, alo_lt_b); + //! s2: %u2 = v_cmp_u_f16 hi(%a), %b + //! s2: %cmp2 = v_cmp_lt_f16 %a, %b + //! s2: %res2, s2: %scc2:scc = s_or_b64 %u2, %cmp2 + //! p_unit_test 2, %res2 + writeout(2, noopt); + + Temp hi_neq_lo = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a, a_hi); + Temp a_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a, a); + noopt = bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), hi_neq_lo, a_unordered); + //! s2: %nan31 = v_cmp_neq_f16 %a, hi(%a) + //! s2: %nan32 = v_cmp_neq_f16 %a, %a + //! s2: %res3, s2: %scc3:scc = s_or_b64 %nan31, %nan32 + //! p_unit_test 3, %res3 + writeout(3, noopt); + + ahi_cmp_const = bld.vopc(aco_opcode::v_cmp_lt_f16, bld.def(bld.lm), a_hi, + bld.copy(bld.def(v2b), Operand::c16(0x4400))); + a_unordered = bld.vopc(aco_opcode::v_cmp_neq_f16, bld.def(bld.lm), a, a); + noopt = + bld.sop2(Builder::s_or, bld.def(bld.lm), bld.def(bld.lm, scc), a_unordered, ahi_cmp_const); + //! s2: %cmp4 = v_cmp_gt_f16 4.0, hi(%a) + //! s2: %nan4 = v_cmp_neq_f16 %a, %a + //! s2: %res4, s2: %scc4:scc = s_or_b64 %nan4, %cmp4 + //! p_unit_test 4, %res4 + writeout(4, noopt); + + finish_opt_test(); +END_TEST + +BEGIN_TEST(optimize.max3_opsel) + /* TODO make these work before GFX11 using SDWA. */ + for (unsigned i = GFX11; i <= GFX11; i++) { + //>> v1: %a, v1: %b, v2b: %c = p_startpgm + if (!setup_cs("v1 v1 v2b", GFX11)) + continue; + + Temp a = inputs[0]; + Temp b = inputs[1]; + Temp c = inputs[2]; + + Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1)); + Temp b_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(1)); + + //! v2b: %res0 = v_max3_f16 hi(%a), %c, hi(%b) + //! p_unit_test 0, %res0 + writeout(0, bld.vop2(aco_opcode::v_max_f16, bld.def(v2b), + bld.vop2(aco_opcode::v_max_f16, bld.def(v2b), a_hi, c), b_hi)); + + finish_opt_test(); + } +END_TEST + +BEGIN_TEST(optimize.neg_mul_opsel) + //>> v1: %a, v2b: %b = p_startpgm + if (!setup_cs("v1 v2b", GFX11)) + return; + + Temp a = inputs[0]; + Temp b = inputs[1]; + + Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1)); + + //! v2b: %res0 = v_mul_f16 -hi(%a), %b + //! p_unit_test 0, %res0 + writeout(0, fneg(fmul(a_hi, b))); + + //! v1: %res1 = v_fma_mix_f32 -hi(%a), lo(%b), -0 + //! p_unit_test 1, %res1 + writeout(1, fneg(fmul(f2f32(a_hi), f2f32(b)))); + + finish_opt_test(); +END_TEST + +BEGIN_TEST(optimize.vinterp_inreg_output_modifiers) + //>> v1: %a, v1: %b, v1: %c = p_startpgm + if (!setup_cs("v1 v1 v1", GFX11)) + return; + + //! v1: %res0 = v_interp_p2_f32_inreg %a, %b, %c clamp + //! p_unit_test 0, %res0 + Temp tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f32_inreg, bld.def(v1), inputs[0], + inputs[1], inputs[2]); + writeout(0, fsat(tmp)); + + //! v1: %res1 = v_fma_f32 %b, %a, %c *2 quad_perm:[2,2,2,2] fi + //! p_unit_test 1, %res1 + tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f32_inreg, bld.def(v1), inputs[1], inputs[0], + inputs[2]); + tmp = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp); + writeout(1, tmp); + + //! v2b: %res2 = v_interp_p2_f16_f32_inreg %a, %b, %c clamp + //! p_unit_test 2, %res2 + tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f16_f32_inreg, bld.def(v2b), inputs[0], + inputs[1], inputs[2]); + writeout(2, fsat(tmp)); + + //! v2b: %tmp3 = v_interp_p2_f16_f32_inreg %b, %a, %c + //! v2b: %res3 = v_mul_f16 2.0, %tmp3 + //! p_unit_test 3, %res3 + tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f16_f32_inreg, bld.def(v2b), inputs[1], + inputs[0], inputs[2]); + tmp = bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp); + writeout(3, tmp); + + //! v2b: %res4 = v_fma_mixlo_f16 %c, %b, %a quad_perm:[2,2,2,2] fi + //! p_unit_test 4, %res4 + tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f32_inreg, bld.def(v1), inputs[2], inputs[1], + inputs[0]); + writeout(4, f2f16(tmp)); + + finish_opt_test(); +END_TEST |