/* * Copyright (C) 2019 Alyssa Rosenzweig * Copyright (C) 2019 Collabora, Ltd. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include "compiler.h" /* Creates pipeline registers. This is a prepass run before the main register * allocator but after scheduling, once bundles are created. It works by * iterating the scheduled IR, checking if a value is ever used after the end * of the current bundle. If it is not, it is promoted to a bundle-specific * pipeline register. * * Pipeline registers are only written from the first two stages of the * pipeline (vmul/sadd) lasting the duration of the bundle only. There are two * 128-bit pipeline registers available (r24/r25). The upshot is that no actual * register allocation is needed; we can _always_ promote a value to a pipeline * register, liveness permitting. This greatly simplifies the logic of this * passing, negating the need for a proper RA like work registers. */ static bool mir_pipeline_ins( compiler_context *ctx, midgard_block *block, midgard_bundle *bundle, unsigned i, unsigned pipeline_count) { midgard_instruction *ins = bundle->instructions[i]; /* Our goal is to create a pipeline register. Pipeline registers are * created at the start of the bundle and are destroyed at the end. So * we conservatively require: * * 1. Each component read in the second stage is written in the first stage. * 2. The index is not live after the bundle. * 3. We're not a special index (writeout, conditionals, ..) * * Rationale: #1 ensures that there is no need to go before the * creation of the bundle, so the pipeline register can exist. #2 is * since the pipeline register will be destroyed at the end. This * ensures that nothing will try to read/write the pipeline register * once it is not live, and that there's no need to go earlier. */ unsigned node = ins->dest; unsigned read_mask = 0; if (node >= SSA_FIXED_MINIMUM) return false; if (node == ctx->blend_src1) return false; /* Analyze the bundle for a per-byte read mask */ for (unsigned j = 0; j < bundle->instruction_count; ++j) { midgard_instruction *q = bundle->instructions[j]; /* The fragment colour can't be pipelined (well, it is * pipelined in r0, but this is a delicate dance with * scheduling and RA, not for us to worry about) */ if (q->compact_branch && q->writeout && mir_has_arg(q, node)) return false; if (q->unit < UNIT_VADD) continue; read_mask |= mir_bytemask_of_read_components(q, node); } /* Now check what's written in the beginning stage */ for (unsigned j = 0; j < bundle->instruction_count; ++j) { midgard_instruction *q = bundle->instructions[j]; if (q->unit >= UNIT_VADD) break; if (q->dest != node) continue; /* Remove the written mask from the read requirements */ read_mask &= ~mir_bytemask(q); } /* Check for leftovers */ if (read_mask) return false; /* We want to know if we live after this bundle, so check if * we're live after the last instruction of the bundle */ midgard_instruction *end = bundle->instructions[ bundle->instruction_count - 1]; if (mir_is_live_after(ctx, block, end, ins->dest)) return false; /* We're only live in this bundle -- pipeline! */ unsigned preg = SSA_FIXED_REGISTER(24 + pipeline_count); for (unsigned j = 0; j < bundle->instruction_count; ++j) { midgard_instruction *q = bundle->instructions[j]; if (q->unit >= UNIT_VADD) mir_rewrite_index_src_single(q, node, preg); else mir_rewrite_index_dst_single(q, node, preg); } return true; } void mir_create_pipeline_registers(compiler_context *ctx) { mir_invalidate_liveness(ctx); mir_foreach_block(ctx, _block) { midgard_block *block = (midgard_block *) _block; mir_foreach_bundle_in_block(block, bundle) { if (!mir_is_alu_bundle(bundle)) continue; if (bundle->instruction_count < 2) continue; /* Only first 2 instructions could pipeline */ bool succ = mir_pipeline_ins(ctx, block, bundle, 0, 0); mir_pipeline_ins(ctx, block, bundle, 1, succ); } } }