diff --git a/src/amd/common/ac_nir.c b/src/amd/common/ac_nir.c index 105636b7be1..7a28128c0a4 100644 --- a/src/amd/common/ac_nir.c +++ b/src/amd/common/ac_nir.c @@ -620,7 +620,7 @@ ac_nir_map_io_location(unsigned location, * including a stride to the base and component offsets. */ nir_def * -ac_nir_calc_io_offset_mapped(nir_builder *b, +ac_nir_calc_io_off(nir_builder *b, nir_intrinsic_instr *intrin, nir_def *base_stride, unsigned component_stride, diff --git a/src/amd/common/ac_nir_helpers.h b/src/amd/common/ac_nir_helpers.h index 39a694af0c8..6926af47f6a 100644 --- a/src/amd/common/ac_nir_helpers.h +++ b/src/amd/common/ac_nir_helpers.h @@ -112,7 +112,7 @@ ac_nir_export_parameters(nir_builder *b, nir_def *(*outputs_16bit_hi)[4]); nir_def * -ac_nir_calc_io_offset_mapped(nir_builder *b, +ac_nir_calc_io_off(nir_builder *b, nir_intrinsic_instr *intrin, nir_def *base_stride, unsigned component_stride, diff --git a/src/amd/common/ac_nir_lower_esgs_io_to_mem.c b/src/amd/common/ac_nir_lower_esgs_io_to_mem.c index ae94f992dee..bb1e38337b7 100644 --- a/src/amd/common/ac_nir_lower_esgs_io_to_mem.c +++ b/src/amd/common/ac_nir_lower_esgs_io_to_mem.c @@ -161,7 +161,7 @@ lower_es_output_store(nir_builder *b, b->cursor = nir_before_instr(&intrin->instr); unsigned mapped = ac_nir_map_io_location(io_sem.location, st->gs_inputs_read, st->map_io); - nir_def *io_off = ac_nir_calc_io_offset_mapped(b, intrin, nir_imm_int(b, 16u), 4u, mapped); + nir_def *io_off = ac_nir_calc_io_off(b, intrin, nir_imm_int(b, 16u), 4u, mapped); nir_def *store_val = intrin->src[0].ssa; if (st->gfx_level <= GFX8) { @@ -292,7 +292,7 @@ gs_per_vertex_input_offset(nir_builder *b, unsigned base_stride = st->gfx_level >= GFX9 ? 1 : 64 /* Wave size on GFX6-8 */; const nir_io_semantics io_sem = nir_intrinsic_io_semantics(instr); unsigned mapped = ac_nir_map_io_location(io_sem.location, st->gs_inputs_read, st->map_io); - nir_def *io_off = ac_nir_calc_io_offset_mapped(b, instr, nir_imm_int(b, base_stride * 4u), base_stride, mapped); + nir_def *io_off = ac_nir_calc_io_off(b, instr, nir_imm_int(b, base_stride * 4u), base_stride, mapped); nir_def *off = nir_iadd(b, io_off, vertex_offset); return nir_imul_imm(b, off, 4u); } diff --git a/src/amd/common/ac_nir_lower_tess_io_to_mem.c b/src/amd/common/ac_nir_lower_tess_io_to_mem.c index 74a2648a791..4c6a3bb7862 100644 --- a/src/amd/common/ac_nir_lower_tess_io_to_mem.c +++ b/src/amd/common/ac_nir_lower_tess_io_to_mem.c @@ -287,7 +287,7 @@ lower_ls_output_store(nir_builder *b, nir_def *base_off_var = nir_imul(b, vertex_idx, nir_load_lshs_vertex_stride_amd(b)); unsigned mapped = ac_nir_map_io_location(io_sem.location, st->tcs_inputs_read, st->map_io); - nir_def *io_off = ac_nir_calc_io_offset_mapped(b, intrin, nir_imm_int(b, 16u), 4u, mapped); + nir_def *io_off = ac_nir_calc_io_off(b, intrin, nir_imm_int(b, 16u), 4u, mapped); unsigned write_mask = nir_intrinsic_write_mask(intrin); nir_def *off = nir_iadd_nuw(b, base_off_var, io_off); @@ -349,7 +349,7 @@ hs_per_vertex_input_lds_offset(nir_builder *b, const nir_io_semantics io_sem = nir_intrinsic_io_semantics(instr); const unsigned mapped = ac_nir_map_io_location(io_sem.location, st->tcs_inputs_read, st->map_io); - nir_def *io_offset = ac_nir_calc_io_offset_mapped(b, instr, nir_imm_int(b, 16u), 4u, mapped); + nir_def *io_offset = ac_nir_calc_io_off(b, instr, nir_imm_int(b, 16u), 4u, mapped); return nir_iadd_nuw(b, nir_iadd_nuw(b, tcs_in_current_patch_offset, vertex_index_off), io_offset); } @@ -402,7 +402,7 @@ hs_output_lds_offset(nir_builder *b, if (intrin) { const nir_io_semantics io_sem = nir_intrinsic_io_semantics(intrin); const unsigned mapped = hs_output_lds_map_io_location(b->shader, per_vertex, io_sem.location, st); - off = ac_nir_calc_io_offset_mapped(b, intrin, nir_imm_int(b, 16u), 4, mapped); + off = ac_nir_calc_io_off(b, intrin, nir_imm_int(b, 16u), 4, mapped); } else { off = nir_imm_int(b, 0); } @@ -483,7 +483,7 @@ hs_per_vertex_output_vmem_offset(nir_builder *b, nir_def *tcs_num_patches = nir_load_tcs_num_patches_amd(b); nir_def *attr_stride = nir_imul(b, tcs_num_patches, nir_imul_imm(b, out_vertices_per_patch, 16u)); nir_def *io_offset = - ac_nir_calc_io_offset_mapped(b, intrin, attr_stride, 4u, + ac_nir_calc_io_off(b, intrin, attr_stride, 4u, hs_output_vram_map_io_location(b->shader, true, io_sem.location, st)); nir_def *rel_patch_id = nir_load_tess_rel_patch_id_amd(b); @@ -506,7 +506,7 @@ hs_per_patch_output_vmem_offset(nir_builder *b, nir_def * off = intrin - ? ac_nir_calc_io_offset_mapped(b, intrin, nir_imul_imm(b, tcs_num_patches, 16u), 4u, + ? ac_nir_calc_io_off(b, intrin, nir_imul_imm(b, tcs_num_patches, 16u), 4u, hs_output_vram_map_io_location(b->shader, false, nir_intrinsic_io_semantics(intrin).location, st)) : nir_imm_int(b, 0);