mirror of
https://github.com/KhronosGroup/SPIRV-Tools
synced 2024-11-24 04:20:13 +00:00
345 lines
13 KiB
C++
345 lines
13 KiB
C++
// Copyright (c) 2017 Google Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
// Validates correctness of atomic SPIR-V instructions.
|
|
|
|
#include "source/val/validate.h"
|
|
|
|
#include "source/diagnostic.h"
|
|
#include "source/opcode.h"
|
|
#include "source/spirv_target_env.h"
|
|
#include "source/util/bitutils.h"
|
|
#include "source/val/instruction.h"
|
|
#include "source/val/validate_memory_semantics.h"
|
|
#include "source/val/validate_scopes.h"
|
|
#include "source/val/validation_state.h"
|
|
|
|
namespace {
|
|
|
|
bool IsStorageClassAllowedByUniversalRules(uint32_t storage_class) {
|
|
switch (storage_class) {
|
|
case SpvStorageClassUniform:
|
|
case SpvStorageClassStorageBuffer:
|
|
case SpvStorageClassWorkgroup:
|
|
case SpvStorageClassCrossWorkgroup:
|
|
case SpvStorageClassGeneric:
|
|
case SpvStorageClassAtomicCounter:
|
|
case SpvStorageClassImage:
|
|
case SpvStorageClassFunction:
|
|
case SpvStorageClassPhysicalStorageBufferEXT:
|
|
return true;
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
} // namespace
|
|
|
|
namespace spvtools {
|
|
namespace val {
|
|
|
|
// Validates correctness of atomic instructions.
|
|
spv_result_t AtomicsPass(ValidationState_t& _, const Instruction* inst) {
|
|
const SpvOp opcode = inst->opcode();
|
|
const uint32_t result_type = inst->type_id();
|
|
bool is_atomic_float_opcode = false;
|
|
if (opcode == SpvOpAtomicLoad || opcode == SpvOpAtomicStore ||
|
|
opcode == SpvOpAtomicFAddEXT || opcode == SpvOpAtomicExchange) {
|
|
is_atomic_float_opcode = true;
|
|
}
|
|
switch (opcode) {
|
|
case SpvOpAtomicLoad:
|
|
case SpvOpAtomicStore:
|
|
case SpvOpAtomicExchange:
|
|
case SpvOpAtomicFAddEXT:
|
|
case SpvOpAtomicCompareExchange:
|
|
case SpvOpAtomicCompareExchangeWeak:
|
|
case SpvOpAtomicIIncrement:
|
|
case SpvOpAtomicIDecrement:
|
|
case SpvOpAtomicIAdd:
|
|
case SpvOpAtomicISub:
|
|
case SpvOpAtomicSMin:
|
|
case SpvOpAtomicUMin:
|
|
case SpvOpAtomicSMax:
|
|
case SpvOpAtomicUMax:
|
|
case SpvOpAtomicAnd:
|
|
case SpvOpAtomicOr:
|
|
case SpvOpAtomicXor:
|
|
case SpvOpAtomicFlagTestAndSet:
|
|
case SpvOpAtomicFlagClear: {
|
|
if (_.HasCapability(SpvCapabilityKernel) &&
|
|
(opcode == SpvOpAtomicLoad || opcode == SpvOpAtomicExchange ||
|
|
opcode == SpvOpAtomicCompareExchange)) {
|
|
if (!_.IsFloatScalarType(result_type) &&
|
|
!_.IsIntScalarType(result_type)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be int or float scalar type";
|
|
}
|
|
} else if (opcode == SpvOpAtomicFlagTestAndSet) {
|
|
if (!_.IsBoolScalarType(result_type)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be bool scalar type";
|
|
}
|
|
} else if (opcode == SpvOpAtomicFlagClear || opcode == SpvOpAtomicStore) {
|
|
assert(result_type == 0);
|
|
} else {
|
|
if (_.IsFloatScalarType(result_type)) {
|
|
if (is_atomic_float_opcode) {
|
|
if (opcode == SpvOpAtomicFAddEXT) {
|
|
if ((_.GetBitWidth(result_type) == 32) &&
|
|
(!_.HasCapability(SpvCapabilityAtomicFloat32AddEXT))) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": float add atomics require the AtomicFloat32AddEXT "
|
|
"capability";
|
|
}
|
|
if ((_.GetBitWidth(result_type) == 64) &&
|
|
(!_.HasCapability(SpvCapabilityAtomicFloat64AddEXT))) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": float add atomics require the AtomicFloat64AddEXT "
|
|
"capability";
|
|
}
|
|
}
|
|
} else {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be int scalar type";
|
|
}
|
|
} else if (_.IsIntScalarType(result_type) &&
|
|
opcode == SpvOpAtomicFAddEXT) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be float scalar type";
|
|
} else if (!_.IsFloatScalarType(result_type) &&
|
|
!_.IsIntScalarType(result_type)) {
|
|
switch (opcode) {
|
|
case SpvOpAtomicFAddEXT:
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be float scalar type";
|
|
case SpvOpAtomicIIncrement:
|
|
case SpvOpAtomicIDecrement:
|
|
case SpvOpAtomicIAdd:
|
|
case SpvOpAtomicISub:
|
|
case SpvOpAtomicSMin:
|
|
case SpvOpAtomicSMax:
|
|
case SpvOpAtomicUMin:
|
|
case SpvOpAtomicUMax:
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be integer scalar type";
|
|
default:
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be int or float scalar type";
|
|
}
|
|
}
|
|
|
|
if (spvIsVulkanEnv(_.context()->target_env) &&
|
|
_.GetBitWidth(result_type) != 32) {
|
|
switch (opcode) {
|
|
case SpvOpAtomicSMin:
|
|
case SpvOpAtomicUMin:
|
|
case SpvOpAtomicSMax:
|
|
case SpvOpAtomicUMax:
|
|
case SpvOpAtomicAnd:
|
|
case SpvOpAtomicOr:
|
|
case SpvOpAtomicXor:
|
|
case SpvOpAtomicIAdd:
|
|
case SpvOpAtomicISub:
|
|
case SpvOpAtomicFAddEXT:
|
|
case SpvOpAtomicLoad:
|
|
case SpvOpAtomicStore:
|
|
case SpvOpAtomicExchange:
|
|
case SpvOpAtomicIIncrement:
|
|
case SpvOpAtomicIDecrement:
|
|
case SpvOpAtomicCompareExchangeWeak:
|
|
case SpvOpAtomicCompareExchange: {
|
|
if (_.GetBitWidth(result_type) == 64 &&
|
|
_.IsIntScalarType(result_type) &&
|
|
!_.HasCapability(SpvCapabilityInt64Atomics))
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": 64-bit atomics require the Int64Atomics "
|
|
"capability";
|
|
} break;
|
|
default:
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": according to the Vulkan spec atomic Result Type "
|
|
"needs "
|
|
"to be a 32-bit int scalar type";
|
|
}
|
|
}
|
|
}
|
|
|
|
uint32_t operand_index =
|
|
opcode == SpvOpAtomicFlagClear || opcode == SpvOpAtomicStore ? 0 : 2;
|
|
const uint32_t pointer_type = _.GetOperandTypeId(inst, operand_index++);
|
|
|
|
uint32_t data_type = 0;
|
|
uint32_t storage_class = 0;
|
|
if (!_.GetPointerTypeInfo(pointer_type, &data_type, &storage_class)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Pointer to be of type OpTypePointer";
|
|
}
|
|
|
|
// Validate storage class against universal rules
|
|
if (!IsStorageClassAllowedByUniversalRules(storage_class)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": storage class forbidden by universal validation rules.";
|
|
}
|
|
|
|
// Then Shader rules
|
|
if (_.HasCapability(SpvCapabilityShader)) {
|
|
if (storage_class == SpvStorageClassFunction) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": Function storage class forbidden when the Shader "
|
|
"capability is declared.";
|
|
}
|
|
}
|
|
|
|
// And finally OpenCL environment rules
|
|
if (spvIsOpenCLEnv(_.context()->target_env)) {
|
|
if ((storage_class != SpvStorageClassFunction) &&
|
|
(storage_class != SpvStorageClassWorkgroup) &&
|
|
(storage_class != SpvStorageClassCrossWorkgroup) &&
|
|
(storage_class != SpvStorageClassGeneric)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": storage class must be Function, Workgroup, "
|
|
"CrossWorkGroup or Generic in the OpenCL environment.";
|
|
}
|
|
|
|
if (_.context()->target_env == SPV_ENV_OPENCL_1_2) {
|
|
if (storage_class == SpvStorageClassGeneric) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< "Storage class cannot be Generic in OpenCL 1.2 "
|
|
"environment";
|
|
}
|
|
}
|
|
}
|
|
|
|
if (opcode == SpvOpAtomicFlagTestAndSet ||
|
|
opcode == SpvOpAtomicFlagClear) {
|
|
if (!_.IsIntScalarType(data_type) || _.GetBitWidth(data_type) != 32) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Pointer to point to a value of 32-bit int type";
|
|
}
|
|
} else if (opcode == SpvOpAtomicStore) {
|
|
if (!_.IsFloatScalarType(data_type) && !_.IsIntScalarType(data_type)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Pointer to be a pointer to int or float "
|
|
<< "scalar type";
|
|
}
|
|
} else {
|
|
if (data_type != result_type) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Pointer to point to a value of type Result "
|
|
"Type";
|
|
}
|
|
}
|
|
|
|
auto memory_scope = inst->GetOperandAs<const uint32_t>(operand_index++);
|
|
if (auto error = ValidateMemoryScope(_, inst, memory_scope)) {
|
|
return error;
|
|
}
|
|
|
|
const auto equal_semantics_index = operand_index++;
|
|
if (auto error = ValidateMemorySemantics(_, inst, equal_semantics_index))
|
|
return error;
|
|
|
|
if (opcode == SpvOpAtomicCompareExchange ||
|
|
opcode == SpvOpAtomicCompareExchangeWeak) {
|
|
const auto unequal_semantics_index = operand_index++;
|
|
if (auto error =
|
|
ValidateMemorySemantics(_, inst, unequal_semantics_index))
|
|
return error;
|
|
|
|
// Volatile bits must match for equal and unequal semantics. Previous
|
|
// checks guarantee they are 32-bit constants, but we need to recheck
|
|
// whether they are evaluatable constants.
|
|
bool is_int32 = false;
|
|
bool is_equal_const = false;
|
|
bool is_unequal_const = false;
|
|
uint32_t equal_value = 0;
|
|
uint32_t unequal_value = 0;
|
|
std::tie(is_int32, is_equal_const, equal_value) = _.EvalInt32IfConst(
|
|
inst->GetOperandAs<uint32_t>(equal_semantics_index));
|
|
std::tie(is_int32, is_unequal_const, unequal_value) =
|
|
_.EvalInt32IfConst(
|
|
inst->GetOperandAs<uint32_t>(unequal_semantics_index));
|
|
if (is_equal_const && is_unequal_const &&
|
|
((equal_value & SpvMemorySemanticsVolatileMask) ^
|
|
(unequal_value & SpvMemorySemanticsVolatileMask))) {
|
|
return _.diag(SPV_ERROR_INVALID_ID, inst)
|
|
<< "Volatile mask setting must match for Equal and Unequal "
|
|
"memory semantics";
|
|
}
|
|
}
|
|
|
|
if (opcode == SpvOpAtomicStore) {
|
|
const uint32_t value_type = _.GetOperandTypeId(inst, 3);
|
|
if (value_type != data_type) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Value type and the type pointed to by "
|
|
"Pointer to be the same";
|
|
}
|
|
} else if (opcode != SpvOpAtomicLoad && opcode != SpvOpAtomicIIncrement &&
|
|
opcode != SpvOpAtomicIDecrement &&
|
|
opcode != SpvOpAtomicFlagTestAndSet &&
|
|
opcode != SpvOpAtomicFlagClear) {
|
|
const uint32_t value_type = _.GetOperandTypeId(inst, operand_index++);
|
|
if (value_type != result_type) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Value to be of type Result Type";
|
|
}
|
|
}
|
|
|
|
if (opcode == SpvOpAtomicCompareExchange ||
|
|
opcode == SpvOpAtomicCompareExchangeWeak) {
|
|
const uint32_t comparator_type =
|
|
_.GetOperandTypeId(inst, operand_index++);
|
|
if (comparator_type != result_type) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Comparator to be of type Result Type";
|
|
}
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return SPV_SUCCESS;
|
|
}
|
|
|
|
} // namespace val
|
|
} // namespace spvtools
|