mirror of
https://github.com/KhronosGroup/SPIRV-Tools
synced 2024-11-22 19:50:05 +00:00
ca004da9f9
* Add knowledge of cooperative matrices Some optimizations are not aware of cooperative matrices, and either do nothing or assert. This commits fixes that up. * Add int tests, and a handle a couple more cases. * Add float tests, and a handle a couple more cases. * Add NV coop matrix as well.
1126 lines
40 KiB
C++
1126 lines
40 KiB
C++
// Copyright (c) 2017 The Khronos Group Inc.
|
|
// Copyright (c) 2017 Valve Corporation
|
|
// Copyright (c) 2017 LunarG Inc.
|
|
// Copyright (c) 2018-2021 Google LLC
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "source/opt/aggressive_dead_code_elim_pass.h"
|
|
|
|
#include <memory>
|
|
#include <stack>
|
|
|
|
#include "source/cfa.h"
|
|
#include "source/opt/eliminate_dead_functions_util.h"
|
|
#include "source/opt/ir_builder.h"
|
|
#include "source/opt/reflect.h"
|
|
#include "source/spirv_constant.h"
|
|
#include "source/util/string_utils.h"
|
|
|
|
namespace spvtools {
|
|
namespace opt {
|
|
namespace {
|
|
|
|
constexpr uint32_t kTypePointerStorageClassInIdx = 0;
|
|
constexpr uint32_t kEntryPointFunctionIdInIdx = 1;
|
|
constexpr uint32_t kSelectionMergeMergeBlockIdInIdx = 0;
|
|
constexpr uint32_t kLoopMergeContinueBlockIdInIdx = 1;
|
|
constexpr uint32_t kCopyMemoryTargetAddrInIdx = 0;
|
|
constexpr uint32_t kCopyMemorySourceAddrInIdx = 1;
|
|
constexpr uint32_t kLoadSourceAddrInIdx = 0;
|
|
constexpr uint32_t kDebugDeclareOperandVariableIndex = 5;
|
|
constexpr uint32_t kGlobalVariableVariableIndex = 12;
|
|
|
|
// Sorting functor to present annotation instructions in an easy-to-process
|
|
// order. The functor orders by opcode first and falls back on unique id
|
|
// ordering if both instructions have the same opcode.
|
|
//
|
|
// Desired priority:
|
|
// spv::Op::OpGroupDecorate
|
|
// spv::Op::OpGroupMemberDecorate
|
|
// spv::Op::OpDecorate
|
|
// spv::Op::OpMemberDecorate
|
|
// spv::Op::OpDecorateId
|
|
// spv::Op::OpDecorateStringGOOGLE
|
|
// spv::Op::OpDecorationGroup
|
|
struct DecorationLess {
|
|
bool operator()(const Instruction* lhs, const Instruction* rhs) const {
|
|
assert(lhs && rhs);
|
|
spv::Op lhsOp = lhs->opcode();
|
|
spv::Op rhsOp = rhs->opcode();
|
|
if (lhsOp != rhsOp) {
|
|
#define PRIORITY_CASE(opcode) \
|
|
if (lhsOp == opcode && rhsOp != opcode) return true; \
|
|
if (rhsOp == opcode && lhsOp != opcode) return false;
|
|
// OpGroupDecorate and OpGroupMember decorate are highest priority to
|
|
// eliminate dead targets early and simplify subsequent checks.
|
|
PRIORITY_CASE(spv::Op::OpGroupDecorate)
|
|
PRIORITY_CASE(spv::Op::OpGroupMemberDecorate)
|
|
PRIORITY_CASE(spv::Op::OpDecorate)
|
|
PRIORITY_CASE(spv::Op::OpMemberDecorate)
|
|
PRIORITY_CASE(spv::Op::OpDecorateId)
|
|
PRIORITY_CASE(spv::Op::OpDecorateStringGOOGLE)
|
|
// OpDecorationGroup is lowest priority to ensure use/def chains remain
|
|
// usable for instructions that target this group.
|
|
PRIORITY_CASE(spv::Op::OpDecorationGroup)
|
|
#undef PRIORITY_CASE
|
|
}
|
|
|
|
// Fall back to maintain total ordering (compare unique ids).
|
|
return *lhs < *rhs;
|
|
}
|
|
};
|
|
|
|
} // namespace
|
|
|
|
bool AggressiveDCEPass::IsVarOfStorage(uint32_t varId,
|
|
spv::StorageClass storageClass) {
|
|
if (varId == 0) return false;
|
|
const Instruction* varInst = get_def_use_mgr()->GetDef(varId);
|
|
const spv::Op op = varInst->opcode();
|
|
if (op != spv::Op::OpVariable) return false;
|
|
const uint32_t varTypeId = varInst->type_id();
|
|
const Instruction* varTypeInst = get_def_use_mgr()->GetDef(varTypeId);
|
|
if (varTypeInst->opcode() != spv::Op::OpTypePointer) return false;
|
|
return spv::StorageClass(varTypeInst->GetSingleWordInOperand(
|
|
kTypePointerStorageClassInIdx)) == storageClass;
|
|
}
|
|
|
|
bool AggressiveDCEPass::IsLocalVar(uint32_t varId, Function* func) {
|
|
if (IsVarOfStorage(varId, spv::StorageClass::Function)) {
|
|
return true;
|
|
}
|
|
|
|
if (!IsVarOfStorage(varId, spv::StorageClass::Private) &&
|
|
!IsVarOfStorage(varId, spv::StorageClass::Workgroup)) {
|
|
return false;
|
|
}
|
|
|
|
// For a variable in the Private or WorkGroup storage class, the variable will
|
|
// get a new instance for every call to an entry point. If the entry point
|
|
// does not have a call, then no other function can read or write to that
|
|
// instance of the variable.
|
|
return IsEntryPointWithNoCalls(func);
|
|
}
|
|
|
|
void AggressiveDCEPass::AddStores(Function* func, uint32_t ptrId) {
|
|
get_def_use_mgr()->ForEachUser(ptrId, [this, ptrId, func](Instruction* user) {
|
|
// If the user is not a part of |func|, skip it.
|
|
BasicBlock* blk = context()->get_instr_block(user);
|
|
if (blk && blk->GetParent() != func) return;
|
|
|
|
switch (user->opcode()) {
|
|
case spv::Op::OpAccessChain:
|
|
case spv::Op::OpInBoundsAccessChain:
|
|
case spv::Op::OpCopyObject:
|
|
this->AddStores(func, user->result_id());
|
|
break;
|
|
case spv::Op::OpLoad:
|
|
break;
|
|
case spv::Op::OpCopyMemory:
|
|
case spv::Op::OpCopyMemorySized:
|
|
if (user->GetSingleWordInOperand(kCopyMemoryTargetAddrInIdx) == ptrId) {
|
|
AddToWorklist(user);
|
|
}
|
|
break;
|
|
// If default, assume it stores e.g. frexp, modf, function call
|
|
case spv::Op::OpStore:
|
|
default:
|
|
AddToWorklist(user);
|
|
break;
|
|
}
|
|
});
|
|
}
|
|
|
|
bool AggressiveDCEPass::AllExtensionsSupported() const {
|
|
// If any extension not in allowlist, return false
|
|
for (auto& ei : get_module()->extensions()) {
|
|
const std::string extName = ei.GetInOperand(0).AsString();
|
|
if (extensions_allowlist_.find(extName) == extensions_allowlist_.end())
|
|
return false;
|
|
}
|
|
// Only allow NonSemantic.Shader.DebugInfo.100, we cannot safely optimise
|
|
// around unknown extended instruction sets even if they are non-semantic
|
|
for (auto& inst : context()->module()->ext_inst_imports()) {
|
|
assert(inst.opcode() == spv::Op::OpExtInstImport &&
|
|
"Expecting an import of an extension's instruction set.");
|
|
const std::string extension_name = inst.GetInOperand(0).AsString();
|
|
if (spvtools::utils::starts_with(extension_name, "NonSemantic.") &&
|
|
(extension_name != "NonSemantic.Shader.DebugInfo.100") &&
|
|
(extension_name != "NonSemantic.DebugPrintf")) {
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool AggressiveDCEPass::IsTargetDead(Instruction* inst) {
|
|
const uint32_t tId = inst->GetSingleWordInOperand(0);
|
|
Instruction* tInst = get_def_use_mgr()->GetDef(tId);
|
|
if (IsAnnotationInst(tInst->opcode())) {
|
|
// This must be a decoration group. We go through annotations in a specific
|
|
// order. So if this is not used by any group or group member decorates, it
|
|
// is dead.
|
|
assert(tInst->opcode() == spv::Op::OpDecorationGroup);
|
|
bool dead = true;
|
|
get_def_use_mgr()->ForEachUser(tInst, [&dead](Instruction* user) {
|
|
if (user->opcode() == spv::Op::OpGroupDecorate ||
|
|
user->opcode() == spv::Op::OpGroupMemberDecorate)
|
|
dead = false;
|
|
});
|
|
return dead;
|
|
}
|
|
return !IsLive(tInst);
|
|
}
|
|
|
|
void AggressiveDCEPass::ProcessLoad(Function* func, uint32_t varId) {
|
|
// Only process locals
|
|
if (!IsLocalVar(varId, func)) return;
|
|
// Return if already processed
|
|
if (live_local_vars_.find(varId) != live_local_vars_.end()) return;
|
|
// Mark all stores to varId as live
|
|
AddStores(func, varId);
|
|
// Cache varId as processed
|
|
live_local_vars_.insert(varId);
|
|
}
|
|
|
|
void AggressiveDCEPass::AddBranch(uint32_t labelId, BasicBlock* bp) {
|
|
std::unique_ptr<Instruction> newBranch(
|
|
new Instruction(context(), spv::Op::OpBranch, 0, 0,
|
|
{{spv_operand_type_t::SPV_OPERAND_TYPE_ID, {labelId}}}));
|
|
context()->AnalyzeDefUse(&*newBranch);
|
|
context()->set_instr_block(&*newBranch, bp);
|
|
bp->AddInstruction(std::move(newBranch));
|
|
}
|
|
|
|
void AggressiveDCEPass::AddBreaksAndContinuesToWorklist(
|
|
Instruction* mergeInst) {
|
|
assert(mergeInst->opcode() == spv::Op::OpSelectionMerge ||
|
|
mergeInst->opcode() == spv::Op::OpLoopMerge);
|
|
|
|
BasicBlock* header = context()->get_instr_block(mergeInst);
|
|
const uint32_t mergeId = mergeInst->GetSingleWordInOperand(0);
|
|
get_def_use_mgr()->ForEachUser(mergeId, [header, this](Instruction* user) {
|
|
if (!user->IsBranch()) return;
|
|
BasicBlock* block = context()->get_instr_block(user);
|
|
if (BlockIsInConstruct(header, block)) {
|
|
// This is a break from the loop.
|
|
AddToWorklist(user);
|
|
// Add branch's merge if there is one.
|
|
Instruction* userMerge = GetMergeInstruction(user);
|
|
if (userMerge != nullptr) AddToWorklist(userMerge);
|
|
}
|
|
});
|
|
|
|
if (mergeInst->opcode() != spv::Op::OpLoopMerge) {
|
|
return;
|
|
}
|
|
|
|
// For loops we need to find the continues as well.
|
|
const uint32_t contId =
|
|
mergeInst->GetSingleWordInOperand(kLoopMergeContinueBlockIdInIdx);
|
|
get_def_use_mgr()->ForEachUser(contId, [&contId, this](Instruction* user) {
|
|
spv::Op op = user->opcode();
|
|
if (op == spv::Op::OpBranchConditional || op == spv::Op::OpSwitch) {
|
|
// A conditional branch or switch can only be a continue if it does not
|
|
// have a merge instruction or its merge block is not the continue block.
|
|
Instruction* hdrMerge = GetMergeInstruction(user);
|
|
if (hdrMerge != nullptr &&
|
|
hdrMerge->opcode() == spv::Op::OpSelectionMerge) {
|
|
uint32_t hdrMergeId =
|
|
hdrMerge->GetSingleWordInOperand(kSelectionMergeMergeBlockIdInIdx);
|
|
if (hdrMergeId == contId) return;
|
|
// Need to mark merge instruction too
|
|
AddToWorklist(hdrMerge);
|
|
}
|
|
} else if (op == spv::Op::OpBranch) {
|
|
// An unconditional branch can only be a continue if it is not
|
|
// branching to its own merge block.
|
|
BasicBlock* blk = context()->get_instr_block(user);
|
|
Instruction* hdrBranch = GetHeaderBranch(blk);
|
|
if (hdrBranch == nullptr) return;
|
|
Instruction* hdrMerge = GetMergeInstruction(hdrBranch);
|
|
if (hdrMerge->opcode() == spv::Op::OpLoopMerge) return;
|
|
uint32_t hdrMergeId =
|
|
hdrMerge->GetSingleWordInOperand(kSelectionMergeMergeBlockIdInIdx);
|
|
if (contId == hdrMergeId) return;
|
|
} else {
|
|
return;
|
|
}
|
|
AddToWorklist(user);
|
|
});
|
|
}
|
|
|
|
bool AggressiveDCEPass::AggressiveDCE(Function* func) {
|
|
std::list<BasicBlock*> structured_order;
|
|
cfg()->ComputeStructuredOrder(func, &*func->begin(), &structured_order);
|
|
live_local_vars_.clear();
|
|
InitializeWorkList(func, structured_order);
|
|
ProcessWorkList(func);
|
|
return KillDeadInstructions(func, structured_order);
|
|
}
|
|
|
|
bool AggressiveDCEPass::KillDeadInstructions(
|
|
const Function* func, std::list<BasicBlock*>& structured_order) {
|
|
bool modified = false;
|
|
for (auto bi = structured_order.begin(); bi != structured_order.end();) {
|
|
uint32_t merge_block_id = 0;
|
|
(*bi)->ForEachInst([this, &modified, &merge_block_id](Instruction* inst) {
|
|
if (IsLive(inst)) return;
|
|
if (inst->opcode() == spv::Op::OpLabel) return;
|
|
// If dead instruction is selection merge, remember merge block
|
|
// for new branch at end of block
|
|
if (inst->opcode() == spv::Op::OpSelectionMerge ||
|
|
inst->opcode() == spv::Op::OpLoopMerge)
|
|
merge_block_id = inst->GetSingleWordInOperand(0);
|
|
to_kill_.push_back(inst);
|
|
modified = true;
|
|
});
|
|
// If a structured if or loop was deleted, add a branch to its merge
|
|
// block, and traverse to the merge block and continue processing there.
|
|
// We know the block still exists because the label is not deleted.
|
|
if (merge_block_id != 0) {
|
|
AddBranch(merge_block_id, *bi);
|
|
for (++bi; (*bi)->id() != merge_block_id; ++bi) {
|
|
}
|
|
|
|
auto merge_terminator = (*bi)->terminator();
|
|
if (merge_terminator->opcode() == spv::Op::OpUnreachable) {
|
|
// The merge was unreachable. This is undefined behaviour so just
|
|
// return (or return an undef). Then mark the new return as live.
|
|
auto func_ret_type_inst = get_def_use_mgr()->GetDef(func->type_id());
|
|
if (func_ret_type_inst->opcode() == spv::Op::OpTypeVoid) {
|
|
merge_terminator->SetOpcode(spv::Op::OpReturn);
|
|
} else {
|
|
// Find an undef for the return value and make sure it gets kept by
|
|
// the pass.
|
|
auto undef_id = Type2Undef(func->type_id());
|
|
auto undef = get_def_use_mgr()->GetDef(undef_id);
|
|
live_insts_.Set(undef->unique_id());
|
|
merge_terminator->SetOpcode(spv::Op::OpReturnValue);
|
|
merge_terminator->SetInOperands({{SPV_OPERAND_TYPE_ID, {undef_id}}});
|
|
get_def_use_mgr()->AnalyzeInstUse(merge_terminator);
|
|
}
|
|
live_insts_.Set(merge_terminator->unique_id());
|
|
}
|
|
} else {
|
|
Instruction* inst = (*bi)->terminator();
|
|
if (!IsLive(inst)) {
|
|
// If the terminator is not live, this block has no live instructions,
|
|
// and it will be unreachable.
|
|
AddUnreachable(*bi);
|
|
}
|
|
++bi;
|
|
}
|
|
}
|
|
return modified;
|
|
}
|
|
|
|
void AggressiveDCEPass::ProcessWorkList(Function* func) {
|
|
while (!worklist_.empty()) {
|
|
Instruction* live_inst = worklist_.front();
|
|
worklist_.pop();
|
|
AddOperandsToWorkList(live_inst);
|
|
MarkBlockAsLive(live_inst);
|
|
MarkLoadedVariablesAsLive(func, live_inst);
|
|
AddDecorationsToWorkList(live_inst);
|
|
AddDebugInstructionsToWorkList(live_inst);
|
|
}
|
|
}
|
|
|
|
void AggressiveDCEPass::AddDebugScopeToWorkList(const Instruction* inst) {
|
|
auto scope = inst->GetDebugScope();
|
|
auto lex_scope_id = scope.GetLexicalScope();
|
|
if (lex_scope_id != kNoDebugScope)
|
|
AddToWorklist(get_def_use_mgr()->GetDef(lex_scope_id));
|
|
auto inlined_at_id = scope.GetInlinedAt();
|
|
if (inlined_at_id != kNoInlinedAt)
|
|
AddToWorklist(get_def_use_mgr()->GetDef(inlined_at_id));
|
|
}
|
|
|
|
void AggressiveDCEPass::AddDebugInstructionsToWorkList(
|
|
const Instruction* inst) {
|
|
for (auto& line_inst : inst->dbg_line_insts()) {
|
|
if (line_inst.IsDebugLineInst()) {
|
|
AddOperandsToWorkList(&line_inst);
|
|
}
|
|
AddDebugScopeToWorkList(&line_inst);
|
|
}
|
|
AddDebugScopeToWorkList(inst);
|
|
}
|
|
|
|
void AggressiveDCEPass::AddDecorationsToWorkList(const Instruction* inst) {
|
|
// Add OpDecorateId instructions that apply to this instruction to the work
|
|
// list. We use the decoration manager to look through the group
|
|
// decorations to get to the OpDecorate* instructions themselves.
|
|
auto decorations =
|
|
get_decoration_mgr()->GetDecorationsFor(inst->result_id(), false);
|
|
for (Instruction* dec : decorations) {
|
|
// We only care about OpDecorateId instructions because the are the only
|
|
// decorations that will reference an id that will have to be kept live
|
|
// because of that use.
|
|
if (dec->opcode() != spv::Op::OpDecorateId) {
|
|
continue;
|
|
}
|
|
if (spv::Decoration(dec->GetSingleWordInOperand(1)) ==
|
|
spv::Decoration::HlslCounterBufferGOOGLE) {
|
|
// These decorations should not force the use id to be live. It will be
|
|
// removed if either the target or the in operand are dead.
|
|
continue;
|
|
}
|
|
AddToWorklist(dec);
|
|
}
|
|
}
|
|
|
|
void AggressiveDCEPass::MarkLoadedVariablesAsLive(Function* func,
|
|
Instruction* inst) {
|
|
std::vector<uint32_t> live_variables = GetLoadedVariables(inst);
|
|
for (uint32_t var_id : live_variables) {
|
|
ProcessLoad(func, var_id);
|
|
}
|
|
}
|
|
|
|
std::vector<uint32_t> AggressiveDCEPass::GetLoadedVariables(Instruction* inst) {
|
|
if (inst->opcode() == spv::Op::OpFunctionCall) {
|
|
return GetLoadedVariablesFromFunctionCall(inst);
|
|
}
|
|
uint32_t var_id = GetLoadedVariableFromNonFunctionCalls(inst);
|
|
if (var_id == 0) {
|
|
return {};
|
|
}
|
|
return {var_id};
|
|
}
|
|
|
|
uint32_t AggressiveDCEPass::GetLoadedVariableFromNonFunctionCalls(
|
|
Instruction* inst) {
|
|
std::vector<uint32_t> live_variables;
|
|
if (inst->IsAtomicWithLoad()) {
|
|
return GetVariableId(inst->GetSingleWordInOperand(kLoadSourceAddrInIdx));
|
|
}
|
|
|
|
switch (inst->opcode()) {
|
|
case spv::Op::OpLoad:
|
|
case spv::Op::OpImageTexelPointer:
|
|
return GetVariableId(inst->GetSingleWordInOperand(kLoadSourceAddrInIdx));
|
|
case spv::Op::OpCopyMemory:
|
|
case spv::Op::OpCopyMemorySized:
|
|
return GetVariableId(
|
|
inst->GetSingleWordInOperand(kCopyMemorySourceAddrInIdx));
|
|
default:
|
|
break;
|
|
}
|
|
|
|
switch (inst->GetCommonDebugOpcode()) {
|
|
case CommonDebugInfoDebugDeclare:
|
|
return inst->GetSingleWordOperand(kDebugDeclareOperandVariableIndex);
|
|
case CommonDebugInfoDebugValue: {
|
|
analysis::DebugInfoManager* debug_info_mgr =
|
|
context()->get_debug_info_mgr();
|
|
return debug_info_mgr->GetVariableIdOfDebugValueUsedForDeclare(inst);
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
std::vector<uint32_t> AggressiveDCEPass::GetLoadedVariablesFromFunctionCall(
|
|
const Instruction* inst) {
|
|
assert(inst->opcode() == spv::Op::OpFunctionCall);
|
|
std::vector<uint32_t> live_variables;
|
|
// NOTE: we should only be checking function call parameters here, not the
|
|
// function itself, however, `IsPtr` will trivially return false for
|
|
// OpFunction
|
|
inst->ForEachInId([this, &live_variables](const uint32_t* operand_id) {
|
|
if (!IsPtr(*operand_id)) return;
|
|
uint32_t var_id = GetVariableId(*operand_id);
|
|
live_variables.push_back(var_id);
|
|
});
|
|
return live_variables;
|
|
}
|
|
|
|
uint32_t AggressiveDCEPass::GetVariableId(uint32_t ptr_id) {
|
|
assert(IsPtr(ptr_id) &&
|
|
"Cannot get the variable when input is not a pointer.");
|
|
uint32_t varId = 0;
|
|
(void)GetPtr(ptr_id, &varId);
|
|
return varId;
|
|
}
|
|
|
|
void AggressiveDCEPass::MarkBlockAsLive(Instruction* inst) {
|
|
BasicBlock* basic_block = context()->get_instr_block(inst);
|
|
if (basic_block == nullptr) {
|
|
return;
|
|
}
|
|
|
|
// If we intend to keep this instruction, we need the block label and
|
|
// block terminator to have a valid block for the instruction.
|
|
AddToWorklist(basic_block->GetLabelInst());
|
|
|
|
// We need to mark the successors blocks that follow as live. If this is
|
|
// header of the merge construct, the construct may be folded, but we will
|
|
// definitely need the merge label. If it is not a construct, the terminator
|
|
// must be live, and the successor blocks will be marked as live when
|
|
// processing the terminator.
|
|
uint32_t merge_id = basic_block->MergeBlockIdIfAny();
|
|
if (merge_id == 0) {
|
|
AddToWorklist(basic_block->terminator());
|
|
} else {
|
|
AddToWorklist(context()->get_def_use_mgr()->GetDef(merge_id));
|
|
}
|
|
|
|
// Mark the structured control flow constructs that contains this block as
|
|
// live. If |inst| is an instruction in the loop header, then it is part of
|
|
// the loop, so the loop construct must be live. We exclude the label because
|
|
// it does not matter how many times it is executed. This could be extended
|
|
// to more instructions, but we will need it for now.
|
|
if (inst->opcode() != spv::Op::OpLabel)
|
|
MarkLoopConstructAsLiveIfLoopHeader(basic_block);
|
|
|
|
Instruction* next_branch_inst = GetBranchForNextHeader(basic_block);
|
|
if (next_branch_inst != nullptr) {
|
|
AddToWorklist(next_branch_inst);
|
|
Instruction* mergeInst = GetMergeInstruction(next_branch_inst);
|
|
AddToWorklist(mergeInst);
|
|
}
|
|
|
|
if (inst->opcode() == spv::Op::OpLoopMerge ||
|
|
inst->opcode() == spv::Op::OpSelectionMerge) {
|
|
AddBreaksAndContinuesToWorklist(inst);
|
|
}
|
|
}
|
|
void AggressiveDCEPass::MarkLoopConstructAsLiveIfLoopHeader(
|
|
BasicBlock* basic_block) {
|
|
// If this is the header for a loop, then loop structure needs to keep as well
|
|
// because the loop header is also part of the loop.
|
|
Instruction* merge_inst = basic_block->GetLoopMergeInst();
|
|
if (merge_inst != nullptr) {
|
|
AddToWorklist(basic_block->terminator());
|
|
AddToWorklist(merge_inst);
|
|
}
|
|
}
|
|
|
|
void AggressiveDCEPass::AddOperandsToWorkList(const Instruction* inst) {
|
|
inst->ForEachInId([this](const uint32_t* iid) {
|
|
Instruction* inInst = get_def_use_mgr()->GetDef(*iid);
|
|
AddToWorklist(inInst);
|
|
});
|
|
if (inst->type_id() != 0) {
|
|
AddToWorklist(get_def_use_mgr()->GetDef(inst->type_id()));
|
|
}
|
|
}
|
|
|
|
void AggressiveDCEPass::InitializeWorkList(
|
|
Function* func, std::list<BasicBlock*>& structured_order) {
|
|
AddToWorklist(&func->DefInst());
|
|
MarkFunctionParameterAsLive(func);
|
|
MarkFirstBlockAsLive(func);
|
|
|
|
// Add instructions with external side effects to the worklist. Also add
|
|
// branches that are not attached to a structured construct.
|
|
// TODO(s-perron): The handling of branch seems to be adhoc. This needs to be
|
|
// cleaned up.
|
|
for (auto& bi : structured_order) {
|
|
for (auto ii = bi->begin(); ii != bi->end(); ++ii) {
|
|
spv::Op op = ii->opcode();
|
|
if (ii->IsBranch()) {
|
|
continue;
|
|
}
|
|
switch (op) {
|
|
case spv::Op::OpStore: {
|
|
uint32_t var_id = 0;
|
|
(void)GetPtr(&*ii, &var_id);
|
|
if (!IsLocalVar(var_id, func)) AddToWorklist(&*ii);
|
|
} break;
|
|
case spv::Op::OpCopyMemory:
|
|
case spv::Op::OpCopyMemorySized: {
|
|
uint32_t var_id = 0;
|
|
uint32_t target_addr_id =
|
|
ii->GetSingleWordInOperand(kCopyMemoryTargetAddrInIdx);
|
|
(void)GetPtr(target_addr_id, &var_id);
|
|
if (!IsLocalVar(var_id, func)) AddToWorklist(&*ii);
|
|
} break;
|
|
case spv::Op::OpLoopMerge:
|
|
case spv::Op::OpSelectionMerge:
|
|
case spv::Op::OpUnreachable:
|
|
break;
|
|
default: {
|
|
// Function calls, atomics, function params, function returns, etc.
|
|
if (!ii->IsOpcodeSafeToDelete()) {
|
|
AddToWorklist(&*ii);
|
|
}
|
|
} break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void AggressiveDCEPass::InitializeModuleScopeLiveInstructions() {
|
|
// Keep all execution modes.
|
|
for (auto& exec : get_module()->execution_modes()) {
|
|
AddToWorklist(&exec);
|
|
}
|
|
// Keep all entry points.
|
|
for (auto& entry : get_module()->entry_points()) {
|
|
if (!preserve_interface_) {
|
|
live_insts_.Set(entry.unique_id());
|
|
// The actual function is live always.
|
|
AddToWorklist(
|
|
get_def_use_mgr()->GetDef(entry.GetSingleWordInOperand(1u)));
|
|
for (uint32_t i = 3; i < entry.NumInOperands(); ++i) {
|
|
auto* var = get_def_use_mgr()->GetDef(entry.GetSingleWordInOperand(i));
|
|
auto storage_class = var->GetSingleWordInOperand(0u);
|
|
// Vulkan support outputs without an associated input, but not inputs
|
|
// without an associated output. Don't remove outputs unless explicitly
|
|
// allowed.
|
|
if (!remove_outputs_ &&
|
|
spv::StorageClass(storage_class) == spv::StorageClass::Output) {
|
|
AddToWorklist(var);
|
|
}
|
|
}
|
|
} else {
|
|
AddToWorklist(&entry);
|
|
}
|
|
}
|
|
for (auto& anno : get_module()->annotations()) {
|
|
if (anno.opcode() == spv::Op::OpDecorate) {
|
|
// Keep workgroup size.
|
|
if (spv::Decoration(anno.GetSingleWordInOperand(1u)) ==
|
|
spv::Decoration::BuiltIn &&
|
|
spv::BuiltIn(anno.GetSingleWordInOperand(2u)) ==
|
|
spv::BuiltIn::WorkgroupSize) {
|
|
AddToWorklist(&anno);
|
|
}
|
|
|
|
if (context()->preserve_bindings()) {
|
|
// Keep all bindings.
|
|
if ((spv::Decoration(anno.GetSingleWordInOperand(1u)) ==
|
|
spv::Decoration::DescriptorSet) ||
|
|
(spv::Decoration(anno.GetSingleWordInOperand(1u)) ==
|
|
spv::Decoration::Binding)) {
|
|
AddToWorklist(&anno);
|
|
}
|
|
}
|
|
|
|
if (context()->preserve_spec_constants()) {
|
|
// Keep all specialization constant instructions
|
|
if (spv::Decoration(anno.GetSingleWordInOperand(1u)) ==
|
|
spv::Decoration::SpecId) {
|
|
AddToWorklist(&anno);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// For each DebugInfo GlobalVariable keep all operands except the Variable.
|
|
// Later, if the variable is killed with KillInst(), we will set the operand
|
|
// to DebugInfoNone. Create and save DebugInfoNone now for this possible
|
|
// later use. This is slightly unoptimal, but it avoids generating it during
|
|
// instruction killing when the module is not consistent.
|
|
bool debug_global_seen = false;
|
|
for (auto& dbg : get_module()->ext_inst_debuginfo()) {
|
|
if (dbg.GetCommonDebugOpcode() != CommonDebugInfoDebugGlobalVariable)
|
|
continue;
|
|
debug_global_seen = true;
|
|
dbg.ForEachInId([this](const uint32_t* iid) {
|
|
Instruction* in_inst = get_def_use_mgr()->GetDef(*iid);
|
|
if (in_inst->opcode() == spv::Op::OpVariable) return;
|
|
AddToWorklist(in_inst);
|
|
});
|
|
}
|
|
if (debug_global_seen) {
|
|
auto dbg_none = context()->get_debug_info_mgr()->GetDebugInfoNone();
|
|
AddToWorklist(dbg_none);
|
|
}
|
|
|
|
// Add top level DebugInfo to worklist
|
|
for (auto& dbg : get_module()->ext_inst_debuginfo()) {
|
|
auto op = dbg.GetShader100DebugOpcode();
|
|
if (op == NonSemanticShaderDebugInfo100DebugCompilationUnit ||
|
|
op == NonSemanticShaderDebugInfo100DebugEntryPoint ||
|
|
op == NonSemanticShaderDebugInfo100DebugSourceContinued) {
|
|
AddToWorklist(&dbg);
|
|
}
|
|
}
|
|
}
|
|
|
|
Pass::Status AggressiveDCEPass::ProcessImpl() {
|
|
// Current functionality assumes shader capability
|
|
// TODO(greg-lunarg): Handle additional capabilities
|
|
if (!context()->get_feature_mgr()->HasCapability(spv::Capability::Shader))
|
|
return Status::SuccessWithoutChange;
|
|
|
|
// Current functionality assumes relaxed logical addressing (see
|
|
// instruction.h)
|
|
// TODO(greg-lunarg): Handle non-logical addressing
|
|
if (context()->get_feature_mgr()->HasCapability(spv::Capability::Addresses))
|
|
return Status::SuccessWithoutChange;
|
|
|
|
// The variable pointer extension is no longer needed to use the capability,
|
|
// so we have to look for the capability.
|
|
if (context()->get_feature_mgr()->HasCapability(
|
|
spv::Capability::VariablePointersStorageBuffer))
|
|
return Status::SuccessWithoutChange;
|
|
|
|
// If any extensions in the module are not explicitly supported,
|
|
// return unmodified.
|
|
if (!AllExtensionsSupported()) return Status::SuccessWithoutChange;
|
|
|
|
// Eliminate Dead functions.
|
|
bool modified = EliminateDeadFunctions();
|
|
|
|
InitializeModuleScopeLiveInstructions();
|
|
|
|
// Run |AggressiveDCE| on the remaining functions. The order does not matter,
|
|
// since |AggressiveDCE| is intra-procedural. This can mean that function
|
|
// will become dead if all function call to them are removed. These dead
|
|
// function will still be in the module after this pass. We expect this to be
|
|
// rare.
|
|
for (Function& fp : *context()->module()) {
|
|
modified |= AggressiveDCE(&fp);
|
|
}
|
|
|
|
// If the decoration manager is kept live then the context will try to keep it
|
|
// up to date. ADCE deals with group decorations by changing the operands in
|
|
// |OpGroupDecorate| instruction directly without informing the decoration
|
|
// manager. This can put it in an invalid state which will cause an error
|
|
// when the context tries to update it. To avoid this problem invalidate
|
|
// the decoration manager upfront.
|
|
//
|
|
// We kill it at now because it is used when processing the entry point
|
|
// functions.
|
|
context()->InvalidateAnalyses(IRContext::Analysis::kAnalysisDecorations);
|
|
|
|
// Process module-level instructions. Now that all live instructions have
|
|
// been marked, it is safe to remove dead global values.
|
|
modified |= ProcessGlobalValues();
|
|
|
|
assert((to_kill_.empty() || modified) &&
|
|
"A dead instruction was identified, but no change recorded.");
|
|
|
|
// Kill all dead instructions.
|
|
for (auto inst : to_kill_) {
|
|
context()->KillInst(inst);
|
|
}
|
|
|
|
// Cleanup all CFG including all unreachable blocks.
|
|
for (Function& fp : *context()->module()) {
|
|
modified |= CFGCleanup(&fp);
|
|
}
|
|
|
|
return modified ? Status::SuccessWithChange : Status::SuccessWithoutChange;
|
|
}
|
|
|
|
bool AggressiveDCEPass::EliminateDeadFunctions() {
|
|
// Identify live functions first. Those that are not live
|
|
// are dead.
|
|
std::unordered_set<const Function*> live_function_set;
|
|
ProcessFunction mark_live = [&live_function_set](Function* fp) {
|
|
live_function_set.insert(fp);
|
|
return false;
|
|
};
|
|
context()->ProcessReachableCallTree(mark_live);
|
|
|
|
bool modified = false;
|
|
for (auto funcIter = get_module()->begin();
|
|
funcIter != get_module()->end();) {
|
|
if (live_function_set.count(&*funcIter) == 0) {
|
|
modified = true;
|
|
funcIter =
|
|
eliminatedeadfunctionsutil::EliminateFunction(context(), &funcIter);
|
|
} else {
|
|
++funcIter;
|
|
}
|
|
}
|
|
|
|
return modified;
|
|
}
|
|
|
|
bool AggressiveDCEPass::ProcessGlobalValues() {
|
|
// Remove debug and annotation statements referencing dead instructions.
|
|
// This must be done before killing the instructions, otherwise there are
|
|
// dead objects in the def/use database.
|
|
bool modified = false;
|
|
Instruction* instruction = &*get_module()->debug2_begin();
|
|
while (instruction) {
|
|
if (instruction->opcode() != spv::Op::OpName) {
|
|
instruction = instruction->NextNode();
|
|
continue;
|
|
}
|
|
|
|
if (IsTargetDead(instruction)) {
|
|
instruction = context()->KillInst(instruction);
|
|
modified = true;
|
|
} else {
|
|
instruction = instruction->NextNode();
|
|
}
|
|
}
|
|
|
|
// This code removes all unnecessary decorations safely (see #1174). It also
|
|
// does so in a more efficient manner than deleting them only as the targets
|
|
// are deleted.
|
|
std::vector<Instruction*> annotations;
|
|
for (auto& inst : get_module()->annotations()) annotations.push_back(&inst);
|
|
std::sort(annotations.begin(), annotations.end(), DecorationLess());
|
|
for (auto annotation : annotations) {
|
|
switch (annotation->opcode()) {
|
|
case spv::Op::OpDecorate:
|
|
case spv::Op::OpMemberDecorate:
|
|
case spv::Op::OpDecorateStringGOOGLE:
|
|
case spv::Op::OpMemberDecorateStringGOOGLE:
|
|
if (IsTargetDead(annotation)) {
|
|
context()->KillInst(annotation);
|
|
modified = true;
|
|
}
|
|
break;
|
|
case spv::Op::OpDecorateId:
|
|
if (IsTargetDead(annotation)) {
|
|
context()->KillInst(annotation);
|
|
modified = true;
|
|
} else {
|
|
if (spv::Decoration(annotation->GetSingleWordInOperand(1)) ==
|
|
spv::Decoration::HlslCounterBufferGOOGLE) {
|
|
// HlslCounterBuffer will reference an id other than the target.
|
|
// If that id is dead, then the decoration can be removed as well.
|
|
uint32_t counter_buffer_id = annotation->GetSingleWordInOperand(2);
|
|
Instruction* counter_buffer_inst =
|
|
get_def_use_mgr()->GetDef(counter_buffer_id);
|
|
if (!IsLive(counter_buffer_inst)) {
|
|
context()->KillInst(annotation);
|
|
modified = true;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
case spv::Op::OpGroupDecorate: {
|
|
// Go through the targets of this group decorate. Remove each dead
|
|
// target. If all targets are dead, remove this decoration.
|
|
bool dead = true;
|
|
bool removed_operand = false;
|
|
for (uint32_t i = 1; i < annotation->NumOperands();) {
|
|
Instruction* opInst =
|
|
get_def_use_mgr()->GetDef(annotation->GetSingleWordOperand(i));
|
|
if (!IsLive(opInst)) {
|
|
// Don't increment |i|.
|
|
annotation->RemoveOperand(i);
|
|
modified = true;
|
|
removed_operand = true;
|
|
} else {
|
|
i++;
|
|
dead = false;
|
|
}
|
|
}
|
|
if (dead) {
|
|
context()->KillInst(annotation);
|
|
modified = true;
|
|
} else if (removed_operand) {
|
|
context()->UpdateDefUse(annotation);
|
|
}
|
|
break;
|
|
}
|
|
case spv::Op::OpGroupMemberDecorate: {
|
|
// Go through the targets of this group member decorate. Remove each
|
|
// dead target (and member index). If all targets are dead, remove this
|
|
// decoration.
|
|
bool dead = true;
|
|
bool removed_operand = false;
|
|
for (uint32_t i = 1; i < annotation->NumOperands();) {
|
|
Instruction* opInst =
|
|
get_def_use_mgr()->GetDef(annotation->GetSingleWordOperand(i));
|
|
if (!IsLive(opInst)) {
|
|
// Don't increment |i|.
|
|
annotation->RemoveOperand(i + 1);
|
|
annotation->RemoveOperand(i);
|
|
modified = true;
|
|
removed_operand = true;
|
|
} else {
|
|
i += 2;
|
|
dead = false;
|
|
}
|
|
}
|
|
if (dead) {
|
|
context()->KillInst(annotation);
|
|
modified = true;
|
|
} else if (removed_operand) {
|
|
context()->UpdateDefUse(annotation);
|
|
}
|
|
break;
|
|
}
|
|
case spv::Op::OpDecorationGroup:
|
|
// By the time we hit decoration groups we've checked everything that
|
|
// can target them. So if they have no uses they must be dead.
|
|
if (get_def_use_mgr()->NumUsers(annotation) == 0) {
|
|
context()->KillInst(annotation);
|
|
modified = true;
|
|
}
|
|
break;
|
|
default:
|
|
assert(false);
|
|
break;
|
|
}
|
|
}
|
|
|
|
for (auto& dbg : get_module()->ext_inst_debuginfo()) {
|
|
if (IsLive(&dbg)) continue;
|
|
// Save GlobalVariable if its variable is live, otherwise null out variable
|
|
// index
|
|
if (dbg.GetCommonDebugOpcode() == CommonDebugInfoDebugGlobalVariable) {
|
|
auto var_id = dbg.GetSingleWordOperand(kGlobalVariableVariableIndex);
|
|
Instruction* var_inst = get_def_use_mgr()->GetDef(var_id);
|
|
if (IsLive(var_inst)) continue;
|
|
context()->ForgetUses(&dbg);
|
|
dbg.SetOperand(
|
|
kGlobalVariableVariableIndex,
|
|
{context()->get_debug_info_mgr()->GetDebugInfoNone()->result_id()});
|
|
context()->AnalyzeUses(&dbg);
|
|
continue;
|
|
}
|
|
to_kill_.push_back(&dbg);
|
|
modified = true;
|
|
}
|
|
|
|
// Since ADCE is disabled for non-shaders, we don't check for export linkage
|
|
// attributes here.
|
|
for (auto& val : get_module()->types_values()) {
|
|
if (!IsLive(&val)) {
|
|
// Save forwarded pointer if pointer is live since closure does not mark
|
|
// this live as it does not have a result id. This is a little too
|
|
// conservative since it is not known if the structure type that needed
|
|
// it is still live. TODO(greg-lunarg): Only save if needed.
|
|
if (val.opcode() == spv::Op::OpTypeForwardPointer) {
|
|
uint32_t ptr_ty_id = val.GetSingleWordInOperand(0);
|
|
Instruction* ptr_ty_inst = get_def_use_mgr()->GetDef(ptr_ty_id);
|
|
if (IsLive(ptr_ty_inst)) continue;
|
|
}
|
|
to_kill_.push_back(&val);
|
|
modified = true;
|
|
}
|
|
}
|
|
|
|
if (!preserve_interface_) {
|
|
// Remove the dead interface variables from the entry point interface list.
|
|
for (auto& entry : get_module()->entry_points()) {
|
|
std::vector<Operand> new_operands;
|
|
for (uint32_t i = 0; i < entry.NumInOperands(); ++i) {
|
|
if (i < 3) {
|
|
// Execution model, function id and name are always valid.
|
|
new_operands.push_back(entry.GetInOperand(i));
|
|
} else {
|
|
auto* var =
|
|
get_def_use_mgr()->GetDef(entry.GetSingleWordInOperand(i));
|
|
if (IsLive(var)) {
|
|
new_operands.push_back(entry.GetInOperand(i));
|
|
}
|
|
}
|
|
}
|
|
if (new_operands.size() != entry.NumInOperands()) {
|
|
entry.SetInOperands(std::move(new_operands));
|
|
get_def_use_mgr()->UpdateDefUse(&entry);
|
|
}
|
|
}
|
|
}
|
|
|
|
return modified;
|
|
}
|
|
|
|
Pass::Status AggressiveDCEPass::Process() {
|
|
// Initialize extensions allowlist
|
|
InitExtensions();
|
|
return ProcessImpl();
|
|
}
|
|
|
|
void AggressiveDCEPass::InitExtensions() {
|
|
extensions_allowlist_.clear();
|
|
|
|
// clang-format off
|
|
extensions_allowlist_.insert({
|
|
"SPV_AMD_shader_explicit_vertex_parameter",
|
|
"SPV_AMD_shader_trinary_minmax",
|
|
"SPV_AMD_gcn_shader",
|
|
"SPV_KHR_shader_ballot",
|
|
"SPV_AMD_shader_ballot",
|
|
"SPV_AMD_gpu_shader_half_float",
|
|
"SPV_KHR_shader_draw_parameters",
|
|
"SPV_KHR_subgroup_vote",
|
|
"SPV_KHR_8bit_storage",
|
|
"SPV_KHR_16bit_storage",
|
|
"SPV_KHR_device_group",
|
|
"SPV_KHR_multiview",
|
|
"SPV_NVX_multiview_per_view_attributes",
|
|
"SPV_NV_viewport_array2",
|
|
"SPV_NV_stereo_view_rendering",
|
|
"SPV_NV_sample_mask_override_coverage",
|
|
"SPV_NV_geometry_shader_passthrough",
|
|
"SPV_AMD_texture_gather_bias_lod",
|
|
"SPV_KHR_storage_buffer_storage_class",
|
|
// SPV_KHR_variable_pointers
|
|
// Currently do not support extended pointer expressions
|
|
"SPV_AMD_gpu_shader_int16",
|
|
"SPV_KHR_post_depth_coverage",
|
|
"SPV_KHR_shader_atomic_counter_ops",
|
|
"SPV_EXT_shader_stencil_export",
|
|
"SPV_EXT_shader_viewport_index_layer",
|
|
"SPV_AMD_shader_image_load_store_lod",
|
|
"SPV_AMD_shader_fragment_mask",
|
|
"SPV_EXT_fragment_fully_covered",
|
|
"SPV_AMD_gpu_shader_half_float_fetch",
|
|
"SPV_GOOGLE_decorate_string",
|
|
"SPV_GOOGLE_hlsl_functionality1",
|
|
"SPV_GOOGLE_user_type",
|
|
"SPV_NV_shader_subgroup_partitioned",
|
|
"SPV_EXT_demote_to_helper_invocation",
|
|
"SPV_EXT_descriptor_indexing",
|
|
"SPV_NV_fragment_shader_barycentric",
|
|
"SPV_NV_compute_shader_derivatives",
|
|
"SPV_NV_shader_image_footprint",
|
|
"SPV_NV_shading_rate",
|
|
"SPV_NV_mesh_shader",
|
|
"SPV_EXT_mesh_shader",
|
|
"SPV_NV_ray_tracing",
|
|
"SPV_KHR_ray_tracing",
|
|
"SPV_KHR_ray_query",
|
|
"SPV_EXT_fragment_invocation_density",
|
|
"SPV_EXT_physical_storage_buffer",
|
|
"SPV_KHR_physical_storage_buffer",
|
|
"SPV_KHR_terminate_invocation",
|
|
"SPV_KHR_shader_clock",
|
|
"SPV_KHR_vulkan_memory_model",
|
|
"SPV_KHR_subgroup_uniform_control_flow",
|
|
"SPV_KHR_integer_dot_product",
|
|
"SPV_EXT_shader_image_int64",
|
|
"SPV_KHR_non_semantic_info",
|
|
"SPV_KHR_uniform_group_instructions",
|
|
"SPV_KHR_fragment_shader_barycentric",
|
|
"SPV_NV_bindless_texture",
|
|
"SPV_EXT_shader_atomic_float_add",
|
|
"SPV_EXT_fragment_shader_interlock",
|
|
"SPV_NV_compute_shader_derivatives",
|
|
"SPV_NV_cooperative_matrix",
|
|
"SPV_KHR_cooperative_matrix"
|
|
});
|
|
// clang-format on
|
|
}
|
|
|
|
Instruction* AggressiveDCEPass::GetHeaderBranch(BasicBlock* blk) {
|
|
if (blk == nullptr) {
|
|
return nullptr;
|
|
}
|
|
BasicBlock* header_block = GetHeaderBlock(blk);
|
|
if (header_block == nullptr) {
|
|
return nullptr;
|
|
}
|
|
return header_block->terminator();
|
|
}
|
|
|
|
BasicBlock* AggressiveDCEPass::GetHeaderBlock(BasicBlock* blk) const {
|
|
if (blk == nullptr) {
|
|
return nullptr;
|
|
}
|
|
|
|
BasicBlock* header_block = nullptr;
|
|
if (blk->IsLoopHeader()) {
|
|
header_block = blk;
|
|
} else {
|
|
uint32_t header =
|
|
context()->GetStructuredCFGAnalysis()->ContainingConstruct(blk->id());
|
|
header_block = context()->get_instr_block(header);
|
|
}
|
|
return header_block;
|
|
}
|
|
|
|
Instruction* AggressiveDCEPass::GetMergeInstruction(Instruction* inst) {
|
|
BasicBlock* bb = context()->get_instr_block(inst);
|
|
if (bb == nullptr) {
|
|
return nullptr;
|
|
}
|
|
return bb->GetMergeInst();
|
|
}
|
|
|
|
Instruction* AggressiveDCEPass::GetBranchForNextHeader(BasicBlock* blk) {
|
|
if (blk == nullptr) {
|
|
return nullptr;
|
|
}
|
|
|
|
if (blk->IsLoopHeader()) {
|
|
uint32_t header =
|
|
context()->GetStructuredCFGAnalysis()->ContainingConstruct(blk->id());
|
|
blk = context()->get_instr_block(header);
|
|
}
|
|
return GetHeaderBranch(blk);
|
|
}
|
|
|
|
void AggressiveDCEPass::MarkFunctionParameterAsLive(const Function* func) {
|
|
func->ForEachParam(
|
|
[this](const Instruction* param) {
|
|
AddToWorklist(const_cast<Instruction*>(param));
|
|
},
|
|
false);
|
|
}
|
|
|
|
bool AggressiveDCEPass::BlockIsInConstruct(BasicBlock* header_block,
|
|
BasicBlock* bb) {
|
|
if (bb == nullptr || header_block == nullptr) {
|
|
return false;
|
|
}
|
|
|
|
uint32_t current_header = bb->id();
|
|
while (current_header != 0) {
|
|
if (current_header == header_block->id()) return true;
|
|
current_header = context()->GetStructuredCFGAnalysis()->ContainingConstruct(
|
|
current_header);
|
|
}
|
|
return false;
|
|
}
|
|
|
|
bool AggressiveDCEPass::IsEntryPointWithNoCalls(Function* func) {
|
|
auto cached_result = entry_point_with_no_calls_cache_.find(func->result_id());
|
|
if (cached_result != entry_point_with_no_calls_cache_.end()) {
|
|
return cached_result->second;
|
|
}
|
|
bool result = IsEntryPoint(func) && !HasCall(func);
|
|
entry_point_with_no_calls_cache_[func->result_id()] = result;
|
|
return result;
|
|
}
|
|
|
|
bool AggressiveDCEPass::IsEntryPoint(Function* func) {
|
|
for (const Instruction& entry_point : get_module()->entry_points()) {
|
|
uint32_t entry_point_id =
|
|
entry_point.GetSingleWordInOperand(kEntryPointFunctionIdInIdx);
|
|
if (entry_point_id == func->result_id()) {
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
bool AggressiveDCEPass::HasCall(Function* func) {
|
|
return !func->WhileEachInst([](Instruction* inst) {
|
|
return inst->opcode() != spv::Op::OpFunctionCall;
|
|
});
|
|
}
|
|
|
|
void AggressiveDCEPass::MarkFirstBlockAsLive(Function* func) {
|
|
BasicBlock* first_block = &*func->begin();
|
|
MarkBlockAsLive(first_block->GetLabelInst());
|
|
}
|
|
|
|
void AggressiveDCEPass::AddUnreachable(BasicBlock*& block) {
|
|
InstructionBuilder builder(
|
|
context(), block,
|
|
IRContext::kAnalysisInstrToBlockMapping | IRContext::kAnalysisDefUse);
|
|
builder.AddUnreachable();
|
|
}
|
|
|
|
} // namespace opt
|
|
} // namespace spvtools
|