microsoft/compiler: use nir_shader_instructions_pass in dxil_nir_lower_double_math

No functional changes.

Signed-off-by: Marcin Ślusarz <marcin.slusarz@intel.com>
Reviewed-by: Jesse Natalie <jenatali@microsoft.com>
Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/12324>
This commit is contained in:
Marcin Ślusarz 2021-08-10 13:36:47 +02:00 committed by Marge Bot
parent b8de41902e
commit 65f9234f96

View file

@ -1243,69 +1243,62 @@ dxil_nir_split_clip_cull_distance(nir_shader *shader)
return params.new_var != NULL;
}
bool
dxil_nir_lower_double_math(nir_shader *shader)
static bool
dxil_nir_lower_double_math_instr(nir_builder *b,
nir_instr *instr,
UNUSED void *cb_data)
{
if (instr->type != nir_instr_type_alu)
return false;
nir_alu_instr *alu = nir_instr_as_alu(instr);
/* TODO: See if we can apply this explicitly to packs/unpacks that are then
* used as a double. As-is, if we had an app explicitly do a 64bit integer op,
* then try to bitcast to double (not expressible in HLSL, but it is in other
* source languages), this would unpack the integer and repack as a double, when
* we probably want to just send the bitcast through to the backend.
*/
b->cursor = nir_before_instr(&alu->instr);
bool progress = false;
nir_foreach_function(func, shader) {
bool func_progress = false;
if (!func->impl)
continue;
nir_builder b;
nir_builder_init(&b, func->impl);
nir_foreach_block(block, func->impl) {
nir_foreach_instr_safe(instr, block) {
if (instr->type != nir_instr_type_alu)
continue;
nir_alu_instr *alu = nir_instr_as_alu(instr);
/* TODO: See if we can apply this explicitly to packs/unpacks that are then
* used as a double. As-is, if we had an app explicitly do a 64bit integer op,
* then try to bitcast to double (not expressible in HLSL, but it is in other
* source languages), this would unpack the integer and repack as a double, when
* we probably want to just send the bitcast through to the backend.
*/
b.cursor = nir_before_instr(&alu->instr);
for (unsigned i = 0; i < nir_op_infos[alu->op].num_inputs; ++i) {
if (nir_alu_type_get_base_type(nir_op_infos[alu->op].input_types[i]) == nir_type_float &&
alu->src[i].src.ssa->bit_size == 64) {
nir_ssa_def *packed_double = nir_channel(&b, alu->src[i].src.ssa, alu->src[i].swizzle[0]);
nir_ssa_def *unpacked_double = nir_unpack_64_2x32(&b, packed_double);
nir_ssa_def *repacked_double = nir_pack_double_2x32_dxil(&b, unpacked_double);
nir_instr_rewrite_src_ssa(instr, &alu->src[i].src, repacked_double);
memset(alu->src[i].swizzle, 0, ARRAY_SIZE(alu->src[i].swizzle));
func_progress = true;
}
}
if (nir_alu_type_get_base_type(nir_op_infos[alu->op].output_type) == nir_type_float &&
alu->dest.dest.ssa.bit_size == 64) {
b.cursor = nir_after_instr(&alu->instr);
nir_ssa_def *packed_double = &alu->dest.dest.ssa;
nir_ssa_def *unpacked_double = nir_unpack_double_2x32_dxil(&b, packed_double);
nir_ssa_def *repacked_double = nir_pack_64_2x32(&b, unpacked_double);
nir_ssa_def_rewrite_uses_after(packed_double, repacked_double, unpacked_double->parent_instr);
func_progress = true;
}
}
for (unsigned i = 0; i < nir_op_infos[alu->op].num_inputs; ++i) {
if (nir_alu_type_get_base_type(nir_op_infos[alu->op].input_types[i]) == nir_type_float &&
alu->src[i].src.ssa->bit_size == 64) {
nir_ssa_def *packed_double = nir_channel(b, alu->src[i].src.ssa, alu->src[i].swizzle[0]);
nir_ssa_def *unpacked_double = nir_unpack_64_2x32(b, packed_double);
nir_ssa_def *repacked_double = nir_pack_double_2x32_dxil(b, unpacked_double);
nir_instr_rewrite_src_ssa(instr, &alu->src[i].src, repacked_double);
memset(alu->src[i].swizzle, 0, ARRAY_SIZE(alu->src[i].swizzle));
progress = true;
}
if (func_progress)
nir_metadata_preserve(func->impl, nir_metadata_block_index |
nir_metadata_dominance |
nir_metadata_loop_analysis);
else
nir_metadata_preserve(func->impl, nir_metadata_all);
progress |= func_progress;
}
if (nir_alu_type_get_base_type(nir_op_infos[alu->op].output_type) == nir_type_float &&
alu->dest.dest.ssa.bit_size == 64) {
b->cursor = nir_after_instr(&alu->instr);
nir_ssa_def *packed_double = &alu->dest.dest.ssa;
nir_ssa_def *unpacked_double = nir_unpack_double_2x32_dxil(b, packed_double);
nir_ssa_def *repacked_double = nir_pack_64_2x32(b, unpacked_double);
nir_ssa_def_rewrite_uses_after(packed_double, repacked_double, unpacked_double->parent_instr);
progress = true;
}
return progress;
}
bool
dxil_nir_lower_double_math(nir_shader *shader)
{
return nir_shader_instructions_pass(shader,
dxil_nir_lower_double_math_instr,
nir_metadata_block_index |
nir_metadata_dominance |
nir_metadata_loop_analysis,
NULL);
}
typedef struct {
gl_system_value *values;
uint32_t count;