r600/sfn: take address loads into account when scheduling

Also change a bit the instruction priority handling

Signed-off-by: Gert Wollny <gert.wollny@collabora.com>
Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/21347>
This commit is contained in:
Gert Wollny 2023-02-10 16:38:04 +01:00 committed by Marge Bot
parent 8e1cf2d439
commit d617052db6
2 changed files with 74 additions and 25 deletions

View file

@ -875,7 +875,8 @@ AluInstr::register_priority() const
if (m_dest) { if (m_dest) {
if (m_dest->has_flag(Register::ssa) && has_alu_flag(alu_write)) { if (m_dest->has_flag(Register::ssa) && has_alu_flag(alu_write)) {
if (m_dest->pin() != pin_group && m_dest->pin() != pin_chgr) if (m_dest->pin() != pin_group && m_dest->pin() != pin_chgr &&
!m_dest->addr())
priority--; priority--;
} else { } else {
// Arrays and registers are pre-allocated, hence scheduling // Arrays and registers are pre-allocated, hence scheduling
@ -886,14 +887,18 @@ AluInstr::register_priority() const
for (const auto s : m_src) { for (const auto s : m_src) {
auto r = s->as_register(); auto r = s->as_register();
if (r && r->has_flag(Register::ssa)) { if (r) {
int pending = 0; if (r->has_flag(Register::ssa)) {
for (auto b : r->uses()) { int pending = 0;
if (!b->is_scheduled()) for (auto b : r->uses()) {
++pending; if (!b->is_scheduled())
++pending;
}
if (pending == 1)
++priority;
} }
if (pending == 1) if (r->addr() && r->addr()->as_register())
++priority; priority += 2;
} }
if (s->as_uniform()) if (s->as_uniform())
++priority; ++priority;

View file

@ -476,6 +476,10 @@ BlockScheduler::schedule_alu(Shader::ShaderBlocks& out_blocks)
bool success = false; bool success = false;
AluGroup *group = nullptr; AluGroup *group = nullptr;
sfn_log << SfnLog::schedule << "Schedule alu with " <<
m_current_block->expected_ar_uses()
<< " pending AR loads\n";
bool has_alu_ready = !alu_vec_ready.empty() || !alu_trans_ready.empty(); bool has_alu_ready = !alu_vec_ready.empty() || !alu_trans_ready.empty();
bool has_lds_ready = bool has_lds_ready =
@ -494,20 +498,36 @@ BlockScheduler::schedule_alu(Shader::ShaderBlocks& out_blocks)
* fetch + read from queue has to be in the same ALU CF block */ * fetch + read from queue has to be in the same ALU CF block */
if (!alu_groups_ready.empty() && !has_lds_ready) { if (!alu_groups_ready.empty() && !has_lds_ready) {
group = *alu_groups_ready.begin(); group = *alu_groups_ready.begin();
if (!m_current_block->try_reserve_kcache(*group)) {
start_new_block(out_blocks, Block::alu);
m_current_block->set_instr_flag(Instr::force_cf);
}
if (!m_current_block->try_reserve_kcache(*group)) sfn_log << SfnLog::schedule << "try schedule " <<
unreachable("Scheduling a group in a new block should always succeed"); *group << "\n";
alu_groups_ready.erase(alu_groups_ready.begin());
sfn_log << SfnLog::schedule << "Schedule ALU group\n"; /* Only start a new CF if we have no pending AR reads */
success = true; if (m_current_block->try_reserve_kcache(*group)) {
} else if (has_alu_ready) { alu_groups_ready.erase(alu_groups_ready.begin());
success = true;
} else {
if (m_current_block->expected_ar_uses() == 0) {
start_new_block(out_blocks, Block::alu);
m_current_block->set_instr_flag(Instr::force_cf);
if (!m_current_block->try_reserve_kcache(*group))
unreachable("Scheduling a group in a new block should always succeed");
alu_groups_ready.erase(alu_groups_ready.begin());
sfn_log << SfnLog::schedule << "Schedule ALU group\n";
success = true;
} else {
sfn_log << SfnLog::schedule << "Don't add group because of " <<
m_current_block->expected_ar_uses()
<< "pending AR loads\n";
}
}
}
if (!group && has_alu_ready) {
group = new AluGroup(); group = new AluGroup();
sfn_log << SfnLog::schedule << "START new ALU group\n"; sfn_log << SfnLog::schedule << "START new ALU group\n";
} else { } else if (!success) {
return false; return false;
} }
@ -540,6 +560,10 @@ BlockScheduler::schedule_alu(Shader::ShaderBlocks& out_blocks)
// kcache constellations // kcache constellations
assert(!m_current_block->lds_group_active()); assert(!m_current_block->lds_group_active());
// AR is loaded but not all uses are done, we don't want
// to start a new CF here
assert(m_current_block->expected_ar_uses() ==0);
// kcache reservation failed, so we have to start a new CF // kcache reservation failed, so we have to start a new CF
start_new_block(out_blocks, Block::alu); start_new_block(out_blocks, Block::alu);
m_current_block->set_instr_flag(Instr::force_cf); m_current_block->set_instr_flag(Instr::force_cf);
@ -560,8 +584,9 @@ BlockScheduler::schedule_alu(Shader::ShaderBlocks& out_blocks)
if (group->has_lds_group_end()) if (group->has_lds_group_end())
m_current_block->lds_group_end(); m_current_block->lds_group_end();
if (group->has_kill_op()) { if (group->index_mode_load() || group->has_kill_op()) {
assert(!group->has_lds_group_start()); assert(!group->has_lds_group_start());
assert(m_current_block->expected_ar_uses() == 0);
start_new_block(out_blocks, Block::alu); start_new_block(out_blocks, Block::alu);
m_current_block->set_instr_flag(Instr::force_cf); m_current_block->set_instr_flag(Instr::force_cf);
} }
@ -669,6 +694,18 @@ BlockScheduler::schedule_alu_to_group_vec(AluGroup *group)
--m_lds_addr_count; --m_lds_addr_count;
} }
if ((*old_i)->num_ar_uses())
m_current_block->set_expected_ar_uses((*old_i)->num_ar_uses());
auto addr = std::get<0>((*old_i)->indirect_addr());
bool has_indirect_reg_load = addr != nullptr && addr->has_flag(Register::addr_or_idx);
if (std::get<0>((*old_i)->indirect_addr()) ||
(!(*old_i)->has_alu_flag(alu_is_lds) &&
((*old_i)->opcode() == op1_set_cf_idx0 ||
(*old_i)->opcode() == op1_set_cf_idx1)))
m_current_block->dec_expected_ar_uses();
alu_vec_ready.erase(old_i); alu_vec_ready.erase(old_i);
success = true; success = true;
sfn_log << SfnLog::schedule << " success\n"; sfn_log << SfnLog::schedule << " success\n";
@ -700,6 +737,10 @@ BlockScheduler::schedule_alu_to_group_trans(AluGroup *group,
if (group->add_trans_instructions(*i)) { if (group->add_trans_instructions(*i)) {
auto old_i = i; auto old_i = i;
++i; ++i;
auto addr = std::get<0>((*old_i)->indirect_addr());
if (addr && addr->has_flag(Register::addr_or_idx))
m_current_block->dec_expected_ar_uses();
readylist.erase(old_i); readylist.erase(old_i);
success = true; success = true;
sfn_log << SfnLog::schedule << " success\n"; sfn_log << SfnLog::schedule << " success\n";
@ -831,13 +872,16 @@ BlockScheduler::collect_ready_alu_vec(std::list<AluInstr *>& ready,
* vec-only instructions when scheduling to the vector slots * vec-only instructions when scheduling to the vector slots
* for everything else we look at the register use */ * for everything else we look at the register use */
if ((*i)->has_lds_access()) auto [addr, dummy1, dummy2] = (*i)->indirect_addr();
if ((*i)->has_lds_access()) {
priority = 100000; priority = 100000;
else if (AluGroup::has_t()) { } else if (addr) {
priority = 10000;
} else if (AluGroup::has_t()) {
auto opinfo = alu_ops.find((*i)->opcode()); auto opinfo = alu_ops.find((*i)->opcode());
assert(opinfo != alu_ops.end()); assert(opinfo != alu_ops.end());
if (opinfo->second.can_channel(AluOp::t, m_chip_class) && if (opinfo->second.can_channel(AluOp::t, m_chip_class))
!std::get<0>((*i)->indirect_addr()))
priority = -1; priority = -1;
} }
@ -861,7 +905,7 @@ BlockScheduler::collect_ready_alu_vec(std::list<AluInstr *>& ready,
}); });
for (auto& i : ready) for (auto& i : ready)
sfn_log << SfnLog::schedule << "V (S): " << *i << "\n"; sfn_log << SfnLog::schedule << "V (S): " << i->priority() << " " << *i << "\n";
return !ready.empty(); return !ready.empty();
} }