/* * Copyright © 2018 Valve Corporation * Copyright © 2018 Google * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. * */ #include "aco_ir.h" #include "util/u_math.h" #include #include namespace aco { RegisterDemand get_live_changes(aco_ptr& instr) { RegisterDemand changes; for (const Definition& def : instr->definitions) { if (!def.isTemp() || def.isKill()) continue; changes += def.getTemp(); } for (const Operand& op : instr->operands) { if (!op.isTemp() || !op.isFirstKill()) continue; changes -= op.getTemp(); } return changes; } RegisterDemand get_temp_registers(aco_ptr& instr) { RegisterDemand temp_registers; for (Definition def : instr->definitions) { if (!def.isTemp()) continue; if (def.isKill()) temp_registers += def.getTemp(); } for (Operand op : instr->operands) { if (op.isTemp() && op.isLateKill() && op.isFirstKill()) temp_registers += op.getTemp(); } return temp_registers; } RegisterDemand get_demand_before(RegisterDemand demand, aco_ptr& instr, aco_ptr& instr_before) { demand -= get_live_changes(instr); demand -= get_temp_registers(instr); if (instr_before) demand += get_temp_registers(instr_before); return demand; } namespace { struct PhiInfo { uint16_t logical_phi_sgpr_ops = 0; uint16_t linear_phi_ops = 0; uint16_t linear_phi_defs = 0; }; void process_live_temps_per_block(Program* program, live& lives, Block* block, unsigned& worklist, std::vector& phi_info) { std::vector& register_demand = lives.register_demand[block->index]; RegisterDemand new_demand; register_demand.resize(block->instructions.size()); RegisterDemand block_register_demand; IDSet live = lives.live_out[block->index]; /* initialize register demand */ for (unsigned t : live) new_demand += Temp(t, program->temp_rc[t]); new_demand.sgpr -= phi_info[block->index].logical_phi_sgpr_ops; /* traverse the instructions backwards */ int idx; for (idx = block->instructions.size() - 1; idx >= 0; idx--) { Instruction* insn = block->instructions[idx].get(); if (is_phi(insn)) break; register_demand[idx] = RegisterDemand(new_demand.vgpr, new_demand.sgpr); /* KILL */ for (Definition& definition : insn->definitions) { if (!definition.isTemp()) { continue; } if ((definition.isFixed() || definition.hasHint()) && definition.physReg() == vcc) program->needs_vcc = true; const Temp temp = definition.getTemp(); const size_t n = live.erase(temp.id()); if (n) { new_demand -= temp; definition.setKill(false); } else { register_demand[idx] += temp; definition.setKill(true); } } /* GEN */ if (insn->opcode == aco_opcode::p_logical_end) { new_demand.sgpr += phi_info[block->index].logical_phi_sgpr_ops; } else { /* we need to do this in a separate loop because the next one can * setKill() for several operands at once and we don't want to * overwrite that in a later iteration */ for (Operand& op : insn->operands) op.setKill(false); for (unsigned i = 0; i < insn->operands.size(); ++i) { Operand& operand = insn->operands[i]; if (!operand.isTemp()) continue; if (operand.isFixed() && operand.physReg() == vcc) program->needs_vcc = true; const Temp temp = operand.getTemp(); const bool inserted = live.insert(temp.id()).second; if (inserted) { operand.setFirstKill(true); for (unsigned j = i + 1; j < insn->operands.size(); ++j) { if (insn->operands[j].isTemp() && insn->operands[j].tempId() == operand.tempId()) { insn->operands[j].setFirstKill(false); insn->operands[j].setKill(true); } } if (operand.isLateKill()) register_demand[idx] += temp; new_demand += temp; } } } block_register_demand.update(register_demand[idx]); } /* update block's register demand for a last time */ block_register_demand.update(new_demand); if (program->progress < CompilationProgress::after_ra) block->register_demand = block_register_demand; /* handle phi definitions */ uint16_t linear_phi_defs = 0; int phi_idx = idx; while (phi_idx >= 0) { register_demand[phi_idx] = new_demand; Instruction* insn = block->instructions[phi_idx].get(); assert(is_phi(insn) && insn->definitions.size() == 1); if (!insn->definitions[0].isTemp()) { assert(insn->definitions[0].isFixed() && insn->definitions[0].physReg() == exec); phi_idx--; continue; } Definition& definition = insn->definitions[0]; if ((definition.isFixed() || definition.hasHint()) && definition.physReg() == vcc) program->needs_vcc = true; const Temp temp = definition.getTemp(); const size_t n = live.erase(temp.id()); if (n) definition.setKill(false); else definition.setKill(true); if (insn->opcode == aco_opcode::p_linear_phi) { assert(definition.getTemp().type() == RegType::sgpr); linear_phi_defs += definition.size(); } phi_idx--; } for (unsigned pred_idx : block->linear_preds) phi_info[pred_idx].linear_phi_defs = linear_phi_defs; /* now, we need to merge the live-ins into the live-out sets */ for (unsigned t : live) { RegClass rc = program->temp_rc[t]; std::vector& preds = rc.is_linear() ? block->linear_preds : block->logical_preds; #ifndef NDEBUG if (preds.empty()) aco_err(program, "Temporary never defined or are defined after use: %%%d in BB%d", t, block->index); #endif for (unsigned pred_idx : preds) { auto it = lives.live_out[pred_idx].insert(t); if (it.second) worklist = std::max(worklist, pred_idx + 1); } } /* handle phi operands */ phi_idx = idx; while (phi_idx >= 0) { Instruction* insn = block->instructions[phi_idx].get(); assert(is_phi(insn)); /* directly insert into the predecessors live-out set */ std::vector& preds = insn->opcode == aco_opcode::p_phi ? block->logical_preds : block->linear_preds; for (unsigned i = 0; i < preds.size(); ++i) { Operand& operand = insn->operands[i]; if (!operand.isTemp()) continue; if (operand.isFixed() && operand.physReg() == vcc) program->needs_vcc = true; /* check if we changed an already processed block */ const bool inserted = lives.live_out[preds[i]].insert(operand.tempId()).second; if (inserted) { worklist = std::max(worklist, preds[i] + 1); if (insn->opcode == aco_opcode::p_phi && operand.getTemp().type() == RegType::sgpr) { phi_info[preds[i]].logical_phi_sgpr_ops += operand.size(); } else if (insn->opcode == aco_opcode::p_linear_phi) { assert(operand.getTemp().type() == RegType::sgpr); phi_info[preds[i]].linear_phi_ops += operand.size(); } } /* set if the operand is killed by this (or another) phi instruction */ operand.setKill(!live.count(operand.tempId())); } phi_idx--; } assert(block->index != 0 || (new_demand == RegisterDemand() && live.empty())); } unsigned calc_waves_per_workgroup(Program* program) { /* When workgroup size is not known, just go with wave_size */ unsigned workgroup_size = program->workgroup_size == UINT_MAX ? program->wave_size : program->workgroup_size; return align(workgroup_size, program->wave_size) / program->wave_size; } } /* end namespace */ uint16_t get_extra_sgprs(Program* program) { if (program->chip_class >= GFX10) { assert(!program->needs_flat_scr); assert(!program->dev.xnack_enabled); return 0; } else if (program->chip_class >= GFX8) { if (program->needs_flat_scr) return 6; else if (program->dev.xnack_enabled) return 4; else if (program->needs_vcc) return 2; else return 0; } else { assert(!program->dev.xnack_enabled); if (program->needs_flat_scr) return 4; else if (program->needs_vcc) return 2; else return 0; } } uint16_t get_sgpr_alloc(Program* program, uint16_t addressable_sgprs) { uint16_t sgprs = addressable_sgprs + get_extra_sgprs(program); uint16_t granule = program->dev.sgpr_alloc_granule; return ALIGN_NPOT(std::max(sgprs, granule), granule); } uint16_t get_vgpr_alloc(Program* program, uint16_t addressable_vgprs) { assert(addressable_vgprs <= program->dev.vgpr_limit); uint16_t granule = program->dev.vgpr_alloc_granule; return align(std::max(addressable_vgprs, granule), granule); } unsigned round_down(unsigned a, unsigned b) { return a - (a % b); } uint16_t get_addr_sgpr_from_waves(Program* program, uint16_t waves) { /* it's not possible to allocate more than 128 SGPRs */ uint16_t sgprs = std::min(program->dev.physical_sgprs / waves, 128); sgprs = round_down(sgprs, program->dev.sgpr_alloc_granule); sgprs -= get_extra_sgprs(program); return std::min(sgprs, program->dev.sgpr_limit); } uint16_t get_addr_vgpr_from_waves(Program* program, uint16_t waves) { uint16_t vgprs = program->dev.physical_vgprs / waves & ~(program->dev.vgpr_alloc_granule - 1); vgprs -= program->config->num_shared_vgprs / 2; return std::min(vgprs, program->dev.vgpr_limit); } void calc_min_waves(Program* program) { unsigned waves_per_workgroup = calc_waves_per_workgroup(program); unsigned simd_per_cu_wgp = program->dev.simd_per_cu * (program->wgp_mode ? 2 : 1); program->min_waves = DIV_ROUND_UP(waves_per_workgroup, simd_per_cu_wgp); } void update_vgpr_sgpr_demand(Program* program, const RegisterDemand new_demand) { unsigned max_waves_per_simd = program->dev.max_wave64_per_simd * (64 / program->wave_size); unsigned simd_per_cu_wgp = program->dev.simd_per_cu * (program->wgp_mode ? 2 : 1); unsigned lds_limit = program->wgp_mode ? program->dev.lds_limit * 2 : program->dev.lds_limit; unsigned max_workgroups_per_cu_wgp = program->wgp_mode ? 32 : 16; assert(program->min_waves >= 1); uint16_t sgpr_limit = get_addr_sgpr_from_waves(program, program->min_waves); uint16_t vgpr_limit = get_addr_vgpr_from_waves(program, program->min_waves); /* this won't compile, register pressure reduction necessary */ if (new_demand.vgpr > vgpr_limit || new_demand.sgpr > sgpr_limit) { program->num_waves = 0; program->max_reg_demand = new_demand; } else { program->num_waves = program->dev.physical_sgprs / get_sgpr_alloc(program, new_demand.sgpr); uint16_t vgpr_demand = get_vgpr_alloc(program, new_demand.vgpr) + program->config->num_shared_vgprs / 2; program->num_waves = std::min(program->num_waves, program->dev.physical_vgprs / vgpr_demand); program->max_waves = max_waves_per_simd; /* adjust max_waves for workgroup and LDS limits */ unsigned waves_per_workgroup = calc_waves_per_workgroup(program); unsigned workgroups_per_cu_wgp = max_waves_per_simd * simd_per_cu_wgp / waves_per_workgroup; unsigned lds_per_workgroup = align(program->config->lds_size * program->dev.lds_encoding_granule, program->dev.lds_alloc_granule); if (program->stage == fragment_fs) { /* PS inputs are moved from PC (parameter cache) to LDS before PS waves are launched. * Each PS input occupies 3x vec4 of LDS space. See Figure 10.3 in GCN3 ISA manual. * These limit occupancy the same way as other stages' LDS usage does. */ unsigned lds_bytes_per_interp = 3 * 16; unsigned lds_param_bytes = lds_bytes_per_interp * program->info->ps.num_interp; lds_per_workgroup += align(lds_param_bytes, program->dev.lds_alloc_granule); } if (lds_per_workgroup) workgroups_per_cu_wgp = std::min(workgroups_per_cu_wgp, lds_limit / lds_per_workgroup); if (waves_per_workgroup > 1) workgroups_per_cu_wgp = std::min(workgroups_per_cu_wgp, max_workgroups_per_cu_wgp); /* in cases like waves_per_workgroup=3 or lds=65536 and * waves_per_workgroup=1, we want the maximum possible number of waves per * SIMD and not the minimum. so DIV_ROUND_UP is used */ program->max_waves = std::min( program->max_waves, DIV_ROUND_UP(workgroups_per_cu_wgp * waves_per_workgroup, simd_per_cu_wgp)); /* incorporate max_waves and calculate max_reg_demand */ program->num_waves = std::min(program->num_waves, program->max_waves); program->max_reg_demand.vgpr = get_addr_vgpr_from_waves(program, program->num_waves); program->max_reg_demand.sgpr = get_addr_sgpr_from_waves(program, program->num_waves); } } live live_var_analysis(Program* program) { live result; result.live_out.resize(program->blocks.size()); result.register_demand.resize(program->blocks.size()); unsigned worklist = program->blocks.size(); std::vector phi_info(program->blocks.size()); RegisterDemand new_demand; program->needs_vcc = false; /* this implementation assumes that the block idx corresponds to the block's position in * program->blocks vector */ while (worklist) { unsigned block_idx = --worklist; process_live_temps_per_block(program, result, &program->blocks[block_idx], worklist, phi_info); new_demand.update(program->blocks[block_idx].register_demand); } /* Handle branches: we will insert copies created for linear phis just before the branch. */ for (Block& block : program->blocks) { result.register_demand[block.index].back().sgpr += phi_info[block.index].linear_phi_defs; result.register_demand[block.index].back().sgpr -= phi_info[block.index].linear_phi_ops; } /* calculate the program's register demand and number of waves */ if (program->progress < CompilationProgress::after_ra) update_vgpr_sgpr_demand(program, new_demand); return result; } } // namespace aco