mirror of
https://github.com/KhronosGroup/SPIRV-Tools
synced 2024-12-26 01:31:06 +00:00
659 lines
25 KiB
C++
659 lines
25 KiB
C++
// Copyright (c) 2017 The Khronos Group Inc.
|
|
// Copyright (c) 2017 Valve Corporation
|
|
// Copyright (c) 2017 LunarG Inc.
|
|
// Copyright (c) 2018 Google Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "source/opt/dead_branch_elim_pass.h"
|
|
|
|
#include <list>
|
|
#include <memory>
|
|
#include <vector>
|
|
|
|
#include "source/cfa.h"
|
|
#include "source/opt/ir_context.h"
|
|
#include "source/opt/iterator.h"
|
|
#include "source/opt/struct_cfg_analysis.h"
|
|
#include "source/util/make_unique.h"
|
|
|
|
namespace spvtools {
|
|
namespace opt {
|
|
|
|
namespace {
|
|
|
|
const uint32_t kBranchCondTrueLabIdInIdx = 1;
|
|
const uint32_t kBranchCondFalseLabIdInIdx = 2;
|
|
|
|
} // anonymous namespace
|
|
|
|
bool DeadBranchElimPass::GetConstCondition(uint32_t condId, bool* condVal) {
|
|
bool condIsConst;
|
|
Instruction* cInst = get_def_use_mgr()->GetDef(condId);
|
|
switch (cInst->opcode()) {
|
|
case SpvOpConstantNull:
|
|
case SpvOpConstantFalse: {
|
|
*condVal = false;
|
|
condIsConst = true;
|
|
} break;
|
|
case SpvOpConstantTrue: {
|
|
*condVal = true;
|
|
condIsConst = true;
|
|
} break;
|
|
case SpvOpLogicalNot: {
|
|
bool negVal;
|
|
condIsConst =
|
|
GetConstCondition(cInst->GetSingleWordInOperand(0), &negVal);
|
|
if (condIsConst) *condVal = !negVal;
|
|
} break;
|
|
default: { condIsConst = false; } break;
|
|
}
|
|
return condIsConst;
|
|
}
|
|
|
|
bool DeadBranchElimPass::GetConstInteger(uint32_t selId, uint32_t* selVal) {
|
|
Instruction* sInst = get_def_use_mgr()->GetDef(selId);
|
|
uint32_t typeId = sInst->type_id();
|
|
Instruction* typeInst = get_def_use_mgr()->GetDef(typeId);
|
|
if (!typeInst || (typeInst->opcode() != SpvOpTypeInt)) return false;
|
|
// TODO(greg-lunarg): Support non-32 bit ints
|
|
if (typeInst->GetSingleWordInOperand(0) != 32) return false;
|
|
if (sInst->opcode() == SpvOpConstant) {
|
|
*selVal = sInst->GetSingleWordInOperand(0);
|
|
return true;
|
|
} else if (sInst->opcode() == SpvOpConstantNull) {
|
|
*selVal = 0;
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
void DeadBranchElimPass::AddBranch(uint32_t labelId, BasicBlock* bp) {
|
|
assert(get_def_use_mgr()->GetDef(labelId) != nullptr);
|
|
std::unique_ptr<Instruction> newBranch(
|
|
new Instruction(context(), SpvOpBranch, 0, 0,
|
|
{{spv_operand_type_t::SPV_OPERAND_TYPE_ID, {labelId}}}));
|
|
context()->AnalyzeDefUse(&*newBranch);
|
|
context()->set_instr_block(&*newBranch, bp);
|
|
bp->AddInstruction(std::move(newBranch));
|
|
}
|
|
|
|
BasicBlock* DeadBranchElimPass::GetParentBlock(uint32_t id) {
|
|
return context()->get_instr_block(get_def_use_mgr()->GetDef(id));
|
|
}
|
|
|
|
bool DeadBranchElimPass::MarkLiveBlocks(
|
|
Function* func, std::unordered_set<BasicBlock*>* live_blocks) {
|
|
std::vector<std::pair<BasicBlock*, uint32_t>> conditions_to_simplify;
|
|
std::unordered_set<BasicBlock*> blocks_with_backedge;
|
|
std::vector<BasicBlock*> stack;
|
|
stack.push_back(&*func->begin());
|
|
bool modified = false;
|
|
while (!stack.empty()) {
|
|
BasicBlock* block = stack.back();
|
|
stack.pop_back();
|
|
|
|
// Live blocks doubles as visited set.
|
|
if (!live_blocks->insert(block).second) continue;
|
|
|
|
uint32_t cont_id = block->ContinueBlockIdIfAny();
|
|
if (cont_id != 0) {
|
|
AddBlocksWithBackEdge(cont_id, block->id(), block->MergeBlockIdIfAny(),
|
|
&blocks_with_backedge);
|
|
}
|
|
|
|
Instruction* terminator = block->terminator();
|
|
uint32_t live_lab_id = 0;
|
|
// Check if the terminator has a single valid successor.
|
|
if (terminator->opcode() == SpvOpBranchConditional) {
|
|
bool condVal;
|
|
if (GetConstCondition(terminator->GetSingleWordInOperand(0u), &condVal)) {
|
|
live_lab_id = terminator->GetSingleWordInOperand(
|
|
condVal ? kBranchCondTrueLabIdInIdx : kBranchCondFalseLabIdInIdx);
|
|
}
|
|
} else if (terminator->opcode() == SpvOpSwitch) {
|
|
uint32_t sel_val;
|
|
if (GetConstInteger(terminator->GetSingleWordInOperand(0u), &sel_val)) {
|
|
// Search switch operands for selector value, set live_lab_id to
|
|
// corresponding label, use default if not found.
|
|
uint32_t icnt = 0;
|
|
uint32_t case_val;
|
|
terminator->WhileEachInOperand(
|
|
[&icnt, &case_val, &sel_val, &live_lab_id](const uint32_t* idp) {
|
|
if (icnt == 1) {
|
|
// Start with default label.
|
|
live_lab_id = *idp;
|
|
} else if (icnt > 1) {
|
|
if (icnt % 2 == 0) {
|
|
case_val = *idp;
|
|
} else {
|
|
if (case_val == sel_val) {
|
|
live_lab_id = *idp;
|
|
return false;
|
|
}
|
|
}
|
|
}
|
|
++icnt;
|
|
return true;
|
|
});
|
|
}
|
|
}
|
|
|
|
// Don't simplify back edges unless it becomes a branch to the header. Every
|
|
// loop must have exactly one back edge to the loop header, so we cannot
|
|
// remove it.
|
|
bool simplify = false;
|
|
if (live_lab_id != 0) {
|
|
if (!blocks_with_backedge.count(block)) {
|
|
// This is not a back edge.
|
|
simplify = true;
|
|
} else {
|
|
const auto& struct_cfg_analysis = context()->GetStructuredCFGAnalysis();
|
|
uint32_t header_id = struct_cfg_analysis->ContainingLoop(block->id());
|
|
if (live_lab_id == header_id) {
|
|
// The new branch will be a branch to the header.
|
|
simplify = true;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (simplify) {
|
|
conditions_to_simplify.push_back({block, live_lab_id});
|
|
stack.push_back(GetParentBlock(live_lab_id));
|
|
} else {
|
|
// All successors are live.
|
|
const auto* const_block = block;
|
|
const_block->ForEachSuccessorLabel([&stack, this](const uint32_t label) {
|
|
stack.push_back(GetParentBlock(label));
|
|
});
|
|
}
|
|
}
|
|
|
|
// Traverse |conditions_to_simplify| in reverse order. This is done so that
|
|
// we simplify nested constructs before simplifying the constructs that
|
|
// contain them.
|
|
for (auto b = conditions_to_simplify.rbegin();
|
|
b != conditions_to_simplify.rend(); ++b) {
|
|
modified |= SimplifyBranch(b->first, b->second);
|
|
}
|
|
|
|
return modified;
|
|
}
|
|
|
|
bool DeadBranchElimPass::SimplifyBranch(BasicBlock* block,
|
|
uint32_t live_lab_id) {
|
|
Instruction* merge_inst = block->GetMergeInst();
|
|
Instruction* terminator = block->terminator();
|
|
if (merge_inst && merge_inst->opcode() == SpvOpSelectionMerge) {
|
|
if (merge_inst->NextNode()->opcode() == SpvOpSwitch &&
|
|
SwitchHasNestedBreak(block->id())) {
|
|
if (terminator->NumInOperands() == 2) {
|
|
// We cannot remove the branch, and it already has a single case, so no
|
|
// work to do.
|
|
return false;
|
|
}
|
|
// We have to keep the switch because it has a nest break, so we
|
|
// remove all cases except for the live one.
|
|
Instruction::OperandList new_operands;
|
|
new_operands.push_back(terminator->GetInOperand(0));
|
|
new_operands.push_back({SPV_OPERAND_TYPE_ID, {live_lab_id}});
|
|
terminator->SetInOperands(move(new_operands));
|
|
context()->UpdateDefUse(terminator);
|
|
} else {
|
|
// Check if the merge instruction is still needed because of a
|
|
// non-nested break from the construct. Move the merge instruction if
|
|
// it is still needed.
|
|
StructuredCFGAnalysis* cfg_analysis =
|
|
context()->GetStructuredCFGAnalysis();
|
|
Instruction* first_break = FindFirstExitFromSelectionMerge(
|
|
live_lab_id, merge_inst->GetSingleWordInOperand(0),
|
|
cfg_analysis->LoopMergeBlock(live_lab_id),
|
|
cfg_analysis->LoopContinueBlock(live_lab_id),
|
|
cfg_analysis->SwitchMergeBlock(live_lab_id));
|
|
|
|
AddBranch(live_lab_id, block);
|
|
context()->KillInst(terminator);
|
|
if (first_break == nullptr) {
|
|
context()->KillInst(merge_inst);
|
|
} else {
|
|
merge_inst->RemoveFromList();
|
|
first_break->InsertBefore(std::unique_ptr<Instruction>(merge_inst));
|
|
context()->set_instr_block(merge_inst,
|
|
context()->get_instr_block(first_break));
|
|
}
|
|
}
|
|
} else {
|
|
AddBranch(live_lab_id, block);
|
|
context()->KillInst(terminator);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
void DeadBranchElimPass::MarkUnreachableStructuredTargets(
|
|
const std::unordered_set<BasicBlock*>& live_blocks,
|
|
std::unordered_set<BasicBlock*>* unreachable_merges,
|
|
std::unordered_map<BasicBlock*, BasicBlock*>* unreachable_continues) {
|
|
for (auto block : live_blocks) {
|
|
if (auto merge_id = block->MergeBlockIdIfAny()) {
|
|
BasicBlock* merge_block = GetParentBlock(merge_id);
|
|
if (!live_blocks.count(merge_block)) {
|
|
unreachable_merges->insert(merge_block);
|
|
}
|
|
if (auto cont_id = block->ContinueBlockIdIfAny()) {
|
|
BasicBlock* cont_block = GetParentBlock(cont_id);
|
|
if (!live_blocks.count(cont_block)) {
|
|
(*unreachable_continues)[cont_block] = block;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
bool DeadBranchElimPass::FixPhiNodesInLiveBlocks(
|
|
Function* func, const std::unordered_set<BasicBlock*>& live_blocks,
|
|
const std::unordered_map<BasicBlock*, BasicBlock*>& unreachable_continues) {
|
|
bool modified = false;
|
|
for (auto& block : *func) {
|
|
if (live_blocks.count(&block)) {
|
|
for (auto iter = block.begin(); iter != block.end();) {
|
|
if (iter->opcode() != SpvOpPhi) {
|
|
break;
|
|
}
|
|
|
|
bool changed = false;
|
|
bool backedge_added = false;
|
|
Instruction* inst = &*iter;
|
|
std::vector<Operand> operands;
|
|
// Build a complete set of operands (not just input operands). Start
|
|
// with type and result id operands.
|
|
operands.push_back(inst->GetOperand(0u));
|
|
operands.push_back(inst->GetOperand(1u));
|
|
// Iterate through the incoming labels and determine which to keep
|
|
// and/or modify. If there in an unreachable continue block, there will
|
|
// be an edge from that block to the header. We need to keep it to
|
|
// maintain the structured control flow. If the header has more that 2
|
|
// incoming edges, then the OpPhi must have an entry for that edge.
|
|
// However, if there is only one other incoming edge, the OpPhi can be
|
|
// eliminated.
|
|
for (uint32_t i = 1; i < inst->NumInOperands(); i += 2) {
|
|
BasicBlock* inc = GetParentBlock(inst->GetSingleWordInOperand(i));
|
|
auto cont_iter = unreachable_continues.find(inc);
|
|
if (cont_iter != unreachable_continues.end() &&
|
|
cont_iter->second == &block && inst->NumInOperands() > 4) {
|
|
if (get_def_use_mgr()
|
|
->GetDef(inst->GetSingleWordInOperand(i - 1))
|
|
->opcode() == SpvOpUndef) {
|
|
// Already undef incoming value, no change necessary.
|
|
operands.push_back(inst->GetInOperand(i - 1));
|
|
operands.push_back(inst->GetInOperand(i));
|
|
backedge_added = true;
|
|
} else {
|
|
// Replace incoming value with undef if this phi exists in the
|
|
// loop header. Otherwise, this edge is not live since the
|
|
// unreachable continue block will be replaced with an
|
|
// unconditional branch to the header only.
|
|
operands.emplace_back(
|
|
SPV_OPERAND_TYPE_ID,
|
|
std::initializer_list<uint32_t>{Type2Undef(inst->type_id())});
|
|
operands.push_back(inst->GetInOperand(i));
|
|
changed = true;
|
|
backedge_added = true;
|
|
}
|
|
} else if (live_blocks.count(inc) && inc->IsSuccessor(&block)) {
|
|
// Keep live incoming edge.
|
|
operands.push_back(inst->GetInOperand(i - 1));
|
|
operands.push_back(inst->GetInOperand(i));
|
|
} else {
|
|
// Remove incoming edge.
|
|
changed = true;
|
|
}
|
|
}
|
|
|
|
if (changed) {
|
|
modified = true;
|
|
uint32_t continue_id = block.ContinueBlockIdIfAny();
|
|
if (!backedge_added && continue_id != 0 &&
|
|
unreachable_continues.count(GetParentBlock(continue_id)) &&
|
|
operands.size() > 4) {
|
|
// Changed the backedge to branch from the continue block instead
|
|
// of a successor of the continue block. Add an entry to the phi to
|
|
// provide an undef for the continue block. Since the successor of
|
|
// the continue must also be unreachable (dominated by the continue
|
|
// block), any entry for the original backedge has been removed
|
|
// from the phi operands.
|
|
operands.emplace_back(
|
|
SPV_OPERAND_TYPE_ID,
|
|
std::initializer_list<uint32_t>{Type2Undef(inst->type_id())});
|
|
operands.emplace_back(SPV_OPERAND_TYPE_ID,
|
|
std::initializer_list<uint32_t>{continue_id});
|
|
}
|
|
|
|
// Either replace the phi with a single value or rebuild the phi out
|
|
// of |operands|.
|
|
//
|
|
// We always have type and result id operands. So this phi has a
|
|
// single source if there are two more operands beyond those.
|
|
if (operands.size() == 4) {
|
|
// First input data operands is at index 2.
|
|
uint32_t replId = operands[2u].words[0];
|
|
context()->ReplaceAllUsesWith(inst->result_id(), replId);
|
|
iter = context()->KillInst(&*inst);
|
|
} else {
|
|
// We've rewritten the operands, so first instruct the def/use
|
|
// manager to forget uses in the phi before we replace them. After
|
|
// replacing operands update the def/use manager by re-analyzing
|
|
// the used ids in this phi.
|
|
get_def_use_mgr()->EraseUseRecordsOfOperandIds(inst);
|
|
inst->ReplaceOperands(operands);
|
|
get_def_use_mgr()->AnalyzeInstUse(inst);
|
|
++iter;
|
|
}
|
|
} else {
|
|
++iter;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return modified;
|
|
}
|
|
|
|
bool DeadBranchElimPass::EraseDeadBlocks(
|
|
Function* func, const std::unordered_set<BasicBlock*>& live_blocks,
|
|
const std::unordered_set<BasicBlock*>& unreachable_merges,
|
|
const std::unordered_map<BasicBlock*, BasicBlock*>& unreachable_continues) {
|
|
bool modified = false;
|
|
for (auto ebi = func->begin(); ebi != func->end();) {
|
|
if (unreachable_continues.count(&*ebi)) {
|
|
uint32_t cont_id = unreachable_continues.find(&*ebi)->second->id();
|
|
if (ebi->begin() != ebi->tail() ||
|
|
ebi->terminator()->opcode() != SpvOpBranch ||
|
|
ebi->terminator()->GetSingleWordInOperand(0u) != cont_id) {
|
|
// Make unreachable, but leave the label.
|
|
KillAllInsts(&*ebi, false);
|
|
// Add unconditional branch to header.
|
|
assert(unreachable_continues.count(&*ebi));
|
|
ebi->AddInstruction(MakeUnique<Instruction>(
|
|
context(), SpvOpBranch, 0, 0,
|
|
std::initializer_list<Operand>{{SPV_OPERAND_TYPE_ID, {cont_id}}}));
|
|
get_def_use_mgr()->AnalyzeInstUse(&*ebi->tail());
|
|
context()->set_instr_block(&*ebi->tail(), &*ebi);
|
|
modified = true;
|
|
}
|
|
++ebi;
|
|
} else if (unreachable_merges.count(&*ebi)) {
|
|
if (ebi->begin() != ebi->tail() ||
|
|
ebi->terminator()->opcode() != SpvOpUnreachable) {
|
|
// Make unreachable, but leave the label.
|
|
KillAllInsts(&*ebi, false);
|
|
// Add unreachable terminator.
|
|
ebi->AddInstruction(
|
|
MakeUnique<Instruction>(context(), SpvOpUnreachable, 0, 0,
|
|
std::initializer_list<Operand>{}));
|
|
context()->AnalyzeUses(ebi->terminator());
|
|
context()->set_instr_block(ebi->terminator(), &*ebi);
|
|
modified = true;
|
|
}
|
|
++ebi;
|
|
} else if (!live_blocks.count(&*ebi)) {
|
|
// Kill this block.
|
|
KillAllInsts(&*ebi);
|
|
ebi = ebi.Erase();
|
|
modified = true;
|
|
} else {
|
|
++ebi;
|
|
}
|
|
}
|
|
|
|
return modified;
|
|
}
|
|
|
|
bool DeadBranchElimPass::EliminateDeadBranches(Function* func) {
|
|
bool modified = false;
|
|
std::unordered_set<BasicBlock*> live_blocks;
|
|
modified |= MarkLiveBlocks(func, &live_blocks);
|
|
|
|
std::unordered_set<BasicBlock*> unreachable_merges;
|
|
std::unordered_map<BasicBlock*, BasicBlock*> unreachable_continues;
|
|
MarkUnreachableStructuredTargets(live_blocks, &unreachable_merges,
|
|
&unreachable_continues);
|
|
modified |= FixPhiNodesInLiveBlocks(func, live_blocks, unreachable_continues);
|
|
modified |= EraseDeadBlocks(func, live_blocks, unreachable_merges,
|
|
unreachable_continues);
|
|
|
|
return modified;
|
|
}
|
|
|
|
void DeadBranchElimPass::FixBlockOrder() {
|
|
context()->BuildInvalidAnalyses(IRContext::kAnalysisCFG |
|
|
IRContext::kAnalysisDominatorAnalysis);
|
|
// Reorders blocks according to DFS of dominator tree.
|
|
ProcessFunction reorder_dominators = [this](Function* function) {
|
|
DominatorAnalysis* dominators = context()->GetDominatorAnalysis(function);
|
|
std::vector<BasicBlock*> blocks;
|
|
for (auto iter = dominators->GetDomTree().begin();
|
|
iter != dominators->GetDomTree().end(); ++iter) {
|
|
if (iter->id() != 0) {
|
|
blocks.push_back(iter->bb_);
|
|
}
|
|
}
|
|
for (uint32_t i = 1; i < blocks.size(); ++i) {
|
|
function->MoveBasicBlockToAfter(blocks[i]->id(), blocks[i - 1]);
|
|
}
|
|
return true;
|
|
};
|
|
|
|
// Reorders blocks according to structured order.
|
|
ProcessFunction reorder_structured = [this](Function* function) {
|
|
std::list<BasicBlock*> order;
|
|
context()->cfg()->ComputeStructuredOrder(function, &*function->begin(),
|
|
&order);
|
|
std::vector<BasicBlock*> blocks;
|
|
for (auto block : order) {
|
|
blocks.push_back(block);
|
|
}
|
|
for (uint32_t i = 1; i < blocks.size(); ++i) {
|
|
function->MoveBasicBlockToAfter(blocks[i]->id(), blocks[i - 1]);
|
|
}
|
|
return true;
|
|
};
|
|
|
|
// Structured order is more intuitive so use it where possible.
|
|
if (context()->get_feature_mgr()->HasCapability(SpvCapabilityShader)) {
|
|
context()->ProcessReachableCallTree(reorder_structured);
|
|
} else {
|
|
context()->ProcessReachableCallTree(reorder_dominators);
|
|
}
|
|
}
|
|
|
|
Pass::Status DeadBranchElimPass::Process() {
|
|
// Do not process if module contains OpGroupDecorate. Additional
|
|
// support required in KillNamesAndDecorates().
|
|
// TODO(greg-lunarg): Add support for OpGroupDecorate
|
|
for (auto& ai : get_module()->annotations())
|
|
if (ai.opcode() == SpvOpGroupDecorate) return Status::SuccessWithoutChange;
|
|
// Process all entry point functions
|
|
ProcessFunction pfn = [this](Function* fp) {
|
|
return EliminateDeadBranches(fp);
|
|
};
|
|
bool modified = context()->ProcessReachableCallTree(pfn);
|
|
if (modified) FixBlockOrder();
|
|
return modified ? Status::SuccessWithChange : Status::SuccessWithoutChange;
|
|
}
|
|
|
|
Instruction* DeadBranchElimPass::FindFirstExitFromSelectionMerge(
|
|
uint32_t start_block_id, uint32_t merge_block_id, uint32_t loop_merge_id,
|
|
uint32_t loop_continue_id, uint32_t switch_merge_id) {
|
|
// To find the "first" exit, we follow branches looking for a conditional
|
|
// branch that is not in a nested construct and is not the header of a new
|
|
// construct. We follow the control flow from |start_block_id| to find the
|
|
// first one.
|
|
|
|
while (start_block_id != merge_block_id && start_block_id != loop_merge_id &&
|
|
start_block_id != loop_continue_id) {
|
|
BasicBlock* start_block = context()->get_instr_block(start_block_id);
|
|
Instruction* branch = start_block->terminator();
|
|
uint32_t next_block_id = 0;
|
|
switch (branch->opcode()) {
|
|
case SpvOpBranchConditional:
|
|
next_block_id = start_block->MergeBlockIdIfAny();
|
|
if (next_block_id == 0) {
|
|
// If a possible target is the |loop_merge_id| or |loop_continue_id|,
|
|
// which are not the current merge node, then we continue the search
|
|
// with the other target.
|
|
for (uint32_t i = 1; i < 3; i++) {
|
|
if (branch->GetSingleWordInOperand(i) == loop_merge_id &&
|
|
loop_merge_id != merge_block_id) {
|
|
next_block_id = branch->GetSingleWordInOperand(3 - i);
|
|
break;
|
|
}
|
|
if (branch->GetSingleWordInOperand(i) == loop_continue_id &&
|
|
loop_continue_id != merge_block_id) {
|
|
next_block_id = branch->GetSingleWordInOperand(3 - i);
|
|
break;
|
|
}
|
|
if (branch->GetSingleWordInOperand(i) == switch_merge_id &&
|
|
switch_merge_id != merge_block_id) {
|
|
next_block_id = branch->GetSingleWordInOperand(3 - i);
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (next_block_id == 0) {
|
|
return branch;
|
|
}
|
|
}
|
|
break;
|
|
case SpvOpSwitch:
|
|
next_block_id = start_block->MergeBlockIdIfAny();
|
|
if (next_block_id == 0) {
|
|
// A switch with no merge instructions can have at most 5 targets:
|
|
// a. |merge_block_id|
|
|
// b. |loop_merge_id|
|
|
// c. |loop_continue_id|
|
|
// d. |switch_merge_id|
|
|
// e. 1 block inside the current region.
|
|
//
|
|
// Note that because this is a switch, |merge_block_id| must equal
|
|
// |switch_merge_id|.
|
|
//
|
|
// This leads to a number of cases of what to do.
|
|
//
|
|
// 1. Does not jump to a block inside of the current construct. In
|
|
// this case, there is not conditional break, so we should return
|
|
// |nullptr|.
|
|
//
|
|
// 2. Jumps to |merge_block_id| and a block inside the current
|
|
// construct. In this case, this branch conditionally break to the
|
|
// end of the current construct, so return the current branch.
|
|
//
|
|
// 3. Otherwise, this branch may break, but not to the current merge
|
|
// block. So we continue with the block that is inside the loop.
|
|
bool found_break = false;
|
|
for (uint32_t i = 1; i < branch->NumInOperands(); i += 2) {
|
|
uint32_t target = branch->GetSingleWordInOperand(i);
|
|
if (target == merge_block_id) {
|
|
found_break = true;
|
|
} else if (target != loop_merge_id && target != loop_continue_id) {
|
|
next_block_id = branch->GetSingleWordInOperand(i);
|
|
}
|
|
}
|
|
|
|
if (next_block_id == 0) {
|
|
// Case 1.
|
|
return nullptr;
|
|
}
|
|
|
|
if (found_break) {
|
|
// Case 2.
|
|
return branch;
|
|
}
|
|
|
|
// The fall through is case 3.
|
|
}
|
|
break;
|
|
case SpvOpBranch:
|
|
// Need to check if this is the header of a loop nested in the
|
|
// selection construct.
|
|
next_block_id = start_block->MergeBlockIdIfAny();
|
|
if (next_block_id == 0) {
|
|
next_block_id = branch->GetSingleWordInOperand(0);
|
|
}
|
|
break;
|
|
default:
|
|
return nullptr;
|
|
}
|
|
start_block_id = next_block_id;
|
|
}
|
|
return nullptr;
|
|
}
|
|
|
|
void DeadBranchElimPass::AddBlocksWithBackEdge(
|
|
uint32_t cont_id, uint32_t header_id, uint32_t merge_id,
|
|
std::unordered_set<BasicBlock*>* blocks_with_back_edges) {
|
|
std::unordered_set<uint32_t> visited;
|
|
visited.insert(cont_id);
|
|
visited.insert(header_id);
|
|
visited.insert(merge_id);
|
|
|
|
std::vector<uint32_t> work_list;
|
|
work_list.push_back(cont_id);
|
|
|
|
while (!work_list.empty()) {
|
|
uint32_t bb_id = work_list.back();
|
|
work_list.pop_back();
|
|
|
|
BasicBlock* bb = context()->get_instr_block(bb_id);
|
|
|
|
bool has_back_edge = false;
|
|
bb->ForEachSuccessorLabel([header_id, &visited, &work_list,
|
|
&has_back_edge](uint32_t* succ_label_id) {
|
|
if (visited.insert(*succ_label_id).second) {
|
|
work_list.push_back(*succ_label_id);
|
|
}
|
|
if (*succ_label_id == header_id) {
|
|
has_back_edge = true;
|
|
}
|
|
});
|
|
|
|
if (has_back_edge) {
|
|
blocks_with_back_edges->insert(bb);
|
|
}
|
|
}
|
|
}
|
|
|
|
bool DeadBranchElimPass::SwitchHasNestedBreak(uint32_t switch_header_id) {
|
|
std::vector<BasicBlock*> block_in_construct;
|
|
BasicBlock* start_block = context()->get_instr_block(switch_header_id);
|
|
uint32_t merge_block_id = start_block->MergeBlockIdIfAny();
|
|
|
|
StructuredCFGAnalysis* cfg_analysis = context()->GetStructuredCFGAnalysis();
|
|
return !get_def_use_mgr()->WhileEachUser(
|
|
merge_block_id,
|
|
[this, cfg_analysis, switch_header_id](Instruction* inst) {
|
|
if (!inst->IsBranch()) {
|
|
return true;
|
|
}
|
|
|
|
BasicBlock* bb = context()->get_instr_block(inst);
|
|
if (bb->id() == switch_header_id) {
|
|
return true;
|
|
}
|
|
return (cfg_analysis->ContainingConstruct(inst) == switch_header_id &&
|
|
bb->GetMergeInst() == nullptr);
|
|
});
|
|
}
|
|
|
|
} // namespace opt
|
|
} // namespace spvtools
|