mirror of
https://github.com/KhronosGroup/Vulkan-Hpp
synced 2024-11-21 20:00:11 +00:00
Update Vulkan-Headers to v1.3.298 (#1971)
Co-authored-by: GitHub <noreply@github.com>
This commit is contained in:
parent
56fdc0b070
commit
8592ed9743
@ -1 +1 @@
|
|||||||
Subproject commit 14345dab231912ee9601136e96ca67a6e1f632e7
|
Subproject commit d91597a82f881d473887b560a03a7edf2720b72c
|
@ -63,7 +63,7 @@ extern "C" __declspec( dllimport ) FARPROC __stdcall GetProcAddress( HINSTANCE h
|
|||||||
# include <span>
|
# include <span>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
static_assert( VK_HEADER_VERSION == 297, "Wrong VK_HEADER_VERSION!" );
|
static_assert( VK_HEADER_VERSION == 298, "Wrong VK_HEADER_VERSION!" );
|
||||||
|
|
||||||
// <tuple> includes <sys/sysmacros.h> through some other header
|
// <tuple> includes <sys/sysmacros.h> through some other header
|
||||||
// this results in major(x) being resolved to gnu_dev_major(x)
|
// this results in major(x) being resolved to gnu_dev_major(x)
|
||||||
@ -3632,28 +3632,36 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
return ::vkGetExecutionGraphPipelineNodeIndexAMDX( device, executionGraph, pNodeInfo, pNodeIndex );
|
return ::vkGetExecutionGraphPipelineNodeIndexAMDX( device, executionGraph, pNodeInfo, pNodeIndex );
|
||||||
}
|
}
|
||||||
|
|
||||||
void vkCmdInitializeGraphScratchMemoryAMDX( VkCommandBuffer commandBuffer, VkDeviceAddress scratch ) const VULKAN_HPP_NOEXCEPT
|
void vkCmdInitializeGraphScratchMemoryAMDX( VkCommandBuffer commandBuffer,
|
||||||
|
VkPipeline executionGraph,
|
||||||
|
VkDeviceAddress scratch,
|
||||||
|
VkDeviceSize scratchSize ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
return ::vkCmdInitializeGraphScratchMemoryAMDX( commandBuffer, scratch );
|
return ::vkCmdInitializeGraphScratchMemoryAMDX( commandBuffer, executionGraph, scratch, scratchSize );
|
||||||
}
|
}
|
||||||
|
|
||||||
void vkCmdDispatchGraphAMDX( VkCommandBuffer commandBuffer,
|
void vkCmdDispatchGraphAMDX( VkCommandBuffer commandBuffer,
|
||||||
VkDeviceAddress scratch,
|
VkDeviceAddress scratch,
|
||||||
|
VkDeviceSize scratchSize,
|
||||||
const VkDispatchGraphCountInfoAMDX * pCountInfo ) const VULKAN_HPP_NOEXCEPT
|
const VkDispatchGraphCountInfoAMDX * pCountInfo ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
return ::vkCmdDispatchGraphAMDX( commandBuffer, scratch, pCountInfo );
|
return ::vkCmdDispatchGraphAMDX( commandBuffer, scratch, scratchSize, pCountInfo );
|
||||||
}
|
}
|
||||||
|
|
||||||
void vkCmdDispatchGraphIndirectAMDX( VkCommandBuffer commandBuffer,
|
void vkCmdDispatchGraphIndirectAMDX( VkCommandBuffer commandBuffer,
|
||||||
VkDeviceAddress scratch,
|
VkDeviceAddress scratch,
|
||||||
|
VkDeviceSize scratchSize,
|
||||||
const VkDispatchGraphCountInfoAMDX * pCountInfo ) const VULKAN_HPP_NOEXCEPT
|
const VkDispatchGraphCountInfoAMDX * pCountInfo ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
return ::vkCmdDispatchGraphIndirectAMDX( commandBuffer, scratch, pCountInfo );
|
return ::vkCmdDispatchGraphIndirectAMDX( commandBuffer, scratch, scratchSize, pCountInfo );
|
||||||
}
|
}
|
||||||
|
|
||||||
void vkCmdDispatchGraphIndirectCountAMDX( VkCommandBuffer commandBuffer, VkDeviceAddress scratch, VkDeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT
|
void vkCmdDispatchGraphIndirectCountAMDX( VkCommandBuffer commandBuffer,
|
||||||
|
VkDeviceAddress scratch,
|
||||||
|
VkDeviceSize scratchSize,
|
||||||
|
VkDeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
return ::vkCmdDispatchGraphIndirectCountAMDX( commandBuffer, scratch, countInfo );
|
return ::vkCmdDispatchGraphIndirectCountAMDX( commandBuffer, scratch, scratchSize, countInfo );
|
||||||
}
|
}
|
||||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
|
|
||||||
|
@ -7025,9 +7025,12 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
|
|
||||||
enum class PipelineCreateFlagBits2KHR : VkPipelineCreateFlags2KHR
|
enum class PipelineCreateFlagBits2KHR : VkPipelineCreateFlags2KHR
|
||||||
{
|
{
|
||||||
eDisableOptimization = VK_PIPELINE_CREATE_2_DISABLE_OPTIMIZATION_BIT_KHR,
|
eDisableOptimization = VK_PIPELINE_CREATE_2_DISABLE_OPTIMIZATION_BIT_KHR,
|
||||||
eAllowDerivatives = VK_PIPELINE_CREATE_2_ALLOW_DERIVATIVES_BIT_KHR,
|
eAllowDerivatives = VK_PIPELINE_CREATE_2_ALLOW_DERIVATIVES_BIT_KHR,
|
||||||
eDerivative = VK_PIPELINE_CREATE_2_DERIVATIVE_BIT_KHR,
|
eDerivative = VK_PIPELINE_CREATE_2_DERIVATIVE_BIT_KHR,
|
||||||
|
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||||
|
eExecutionGraphAMDX = VK_PIPELINE_CREATE_2_EXECUTION_GRAPH_BIT_AMDX,
|
||||||
|
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
eEnableLegacyDitheringEXT = VK_PIPELINE_CREATE_2_ENABLE_LEGACY_DITHERING_BIT_EXT,
|
eEnableLegacyDitheringEXT = VK_PIPELINE_CREATE_2_ENABLE_LEGACY_DITHERING_BIT_EXT,
|
||||||
eViewIndexFromDeviceIndex = VK_PIPELINE_CREATE_2_VIEW_INDEX_FROM_DEVICE_INDEX_BIT_KHR,
|
eViewIndexFromDeviceIndex = VK_PIPELINE_CREATE_2_VIEW_INDEX_FROM_DEVICE_INDEX_BIT_KHR,
|
||||||
eDispatchBase = VK_PIPELINE_CREATE_2_DISPATCH_BASE_BIT_KHR,
|
eDispatchBase = VK_PIPELINE_CREATE_2_DISPATCH_BASE_BIT_KHR,
|
||||||
@ -7068,8 +7071,11 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
{
|
{
|
||||||
static VULKAN_HPP_CONST_OR_CONSTEXPR bool isBitmask = true;
|
static VULKAN_HPP_CONST_OR_CONSTEXPR bool isBitmask = true;
|
||||||
static VULKAN_HPP_CONST_OR_CONSTEXPR PipelineCreateFlags2KHR allFlags =
|
static VULKAN_HPP_CONST_OR_CONSTEXPR PipelineCreateFlags2KHR allFlags =
|
||||||
PipelineCreateFlagBits2KHR::eDisableOptimization | PipelineCreateFlagBits2KHR::eAllowDerivatives | PipelineCreateFlagBits2KHR::eDerivative |
|
PipelineCreateFlagBits2KHR::eDisableOptimization | PipelineCreateFlagBits2KHR::eAllowDerivatives | PipelineCreateFlagBits2KHR::eDerivative
|
||||||
PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT | PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex |
|
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||||
|
| PipelineCreateFlagBits2KHR::eExecutionGraphAMDX
|
||||||
|
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
|
| PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT | PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex |
|
||||||
PipelineCreateFlagBits2KHR::eDispatchBase | PipelineCreateFlagBits2KHR::eDeferCompileNV | PipelineCreateFlagBits2KHR::eCaptureStatistics |
|
PipelineCreateFlagBits2KHR::eDispatchBase | PipelineCreateFlagBits2KHR::eDeferCompileNV | PipelineCreateFlagBits2KHR::eCaptureStatistics |
|
||||||
PipelineCreateFlagBits2KHR::eCaptureInternalRepresentations | PipelineCreateFlagBits2KHR::eFailOnPipelineCompileRequired |
|
PipelineCreateFlagBits2KHR::eCaptureInternalRepresentations | PipelineCreateFlagBits2KHR::eFailOnPipelineCompileRequired |
|
||||||
PipelineCreateFlagBits2KHR::eEarlyReturnOnFailure | PipelineCreateFlagBits2KHR::eLinkTimeOptimizationEXT |
|
PipelineCreateFlagBits2KHR::eEarlyReturnOnFailure | PipelineCreateFlagBits2KHR::eLinkTimeOptimizationEXT |
|
||||||
|
@ -993,17 +993,15 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||||
{ "VK_AMDX_shader_enqueue",
|
{ "VK_AMDX_shader_enqueue",
|
||||||
{ { "VK_VERSION_1_0",
|
{ { "VK_VERSION_1_0",
|
||||||
{ {
|
|
||||||
"VK_KHR_get_physical_device_properties2",
|
|
||||||
} } },
|
|
||||||
{ "VK_VERSION_1_1",
|
|
||||||
{ {
|
{ {
|
||||||
"VK_KHR_synchronization2",
|
"VK_KHR_synchronization2",
|
||||||
|
"VK_KHR_spirv_1_4",
|
||||||
|
"VK_EXT_extended_dynamic_state",
|
||||||
} } },
|
} } },
|
||||||
{ "VK_VERSION_1_3",
|
{ "VK_VERSION_1_3",
|
||||||
{ {
|
{ {
|
||||||
|
"VK_KHR_maintenance5",
|
||||||
"VK_KHR_pipeline_library",
|
"VK_KHR_pipeline_library",
|
||||||
"VK_KHR_spirv_1_4",
|
|
||||||
} } } } },
|
} } } } },
|
||||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
{ "VK_EXT_inline_uniform_block",
|
{ "VK_EXT_inline_uniform_block",
|
||||||
|
@ -15575,27 +15575,35 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||||
|
|
||||||
template <typename Dispatch>
|
template <typename Dispatch>
|
||||||
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||||
d.vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ) );
|
d.vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||||
|
static_cast<VkPipeline>( executionGraph ),
|
||||||
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ) );
|
||||||
}
|
}
|
||||||
|
|
||||||
template <typename Dispatch>
|
template <typename Dispatch>
|
||||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
||||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||||
d.vkCmdDispatchGraphAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
d.vkCmdDispatchGraphAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||||
static_cast<VkDeviceAddress>( scratch ),
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ),
|
||||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( pCountInfo ) );
|
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( pCountInfo ) );
|
||||||
}
|
}
|
||||||
|
|
||||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||||
template <typename Dispatch>
|
template <typename Dispatch>
|
||||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
||||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
@ -15604,24 +15612,30 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
VULKAN_HPP_ASSERT( d.vkCmdDispatchGraphAMDX && "Function <vkCmdDispatchGraphAMDX> requires <VK_AMDX_shader_enqueue>" );
|
VULKAN_HPP_ASSERT( d.vkCmdDispatchGraphAMDX && "Function <vkCmdDispatchGraphAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||||
# endif
|
# endif
|
||||||
|
|
||||||
d.vkCmdDispatchGraphAMDX( m_commandBuffer, static_cast<VkDeviceAddress>( scratch ), reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
d.vkCmdDispatchGraphAMDX( m_commandBuffer,
|
||||||
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ),
|
||||||
|
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||||
}
|
}
|
||||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||||
|
|
||||||
template <typename Dispatch>
|
template <typename Dispatch>
|
||||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
||||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||||
d.vkCmdDispatchGraphIndirectAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
d.vkCmdDispatchGraphIndirectAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||||
static_cast<VkDeviceAddress>( scratch ),
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ),
|
||||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( pCountInfo ) );
|
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( pCountInfo ) );
|
||||||
}
|
}
|
||||||
|
|
||||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||||
template <typename Dispatch>
|
template <typename Dispatch>
|
||||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
||||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
@ -15630,19 +15644,24 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
VULKAN_HPP_ASSERT( d.vkCmdDispatchGraphIndirectAMDX && "Function <vkCmdDispatchGraphIndirectAMDX> requires <VK_AMDX_shader_enqueue>" );
|
VULKAN_HPP_ASSERT( d.vkCmdDispatchGraphIndirectAMDX && "Function <vkCmdDispatchGraphIndirectAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||||
# endif
|
# endif
|
||||||
|
|
||||||
d.vkCmdDispatchGraphIndirectAMDX(
|
d.vkCmdDispatchGraphIndirectAMDX( m_commandBuffer,
|
||||||
m_commandBuffer, static_cast<VkDeviceAddress>( scratch ), reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ),
|
||||||
|
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||||
}
|
}
|
||||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||||
|
|
||||||
template <typename Dispatch>
|
template <typename Dispatch>
|
||||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo,
|
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo,
|
||||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||||
d.vkCmdDispatchGraphIndirectCountAMDX(
|
d.vkCmdDispatchGraphIndirectCountAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||||
static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ), static_cast<VkDeviceAddress>( countInfo ) );
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ),
|
||||||
|
static_cast<VkDeviceAddress>( countInfo ) );
|
||||||
}
|
}
|
||||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
|
|
||||||
|
@ -6078,33 +6078,40 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
//=== VK_AMDX_shader_enqueue ===
|
//=== VK_AMDX_shader_enqueue ===
|
||||||
|
|
||||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||||
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||||
|
|
||||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||||
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
||||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||||
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
||||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||||
|
|
||||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||||
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
||||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||||
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
||||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||||
|
|
||||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||||
void dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
void dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo,
|
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo,
|
||||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
|
@ -4626,7 +4626,9 @@ namespace std
|
|||||||
std::size_t seed = 0;
|
std::size_t seed = 0;
|
||||||
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.sType );
|
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.sType );
|
||||||
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.pNext );
|
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.pNext );
|
||||||
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.size );
|
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.minSize );
|
||||||
|
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.maxSize );
|
||||||
|
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.sizeGranularity );
|
||||||
return seed;
|
return seed;
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
@ -11865,6 +11867,7 @@ namespace std
|
|||||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.sType );
|
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.sType );
|
||||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.pNext );
|
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.pNext );
|
||||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.shaderEnqueue );
|
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.shaderEnqueue );
|
||||||
|
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.shaderMeshEnqueue );
|
||||||
return seed;
|
return seed;
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
@ -11885,6 +11888,11 @@ namespace std
|
|||||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphShaderPayloadSize );
|
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphShaderPayloadSize );
|
||||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphShaderPayloadCount );
|
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphShaderPayloadCount );
|
||||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.executionGraphDispatchAddressAlignment );
|
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.executionGraphDispatchAddressAlignment );
|
||||||
|
for ( size_t i = 0; i < 3; ++i )
|
||||||
|
{
|
||||||
|
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphWorkgroupCount[i] );
|
||||||
|
}
|
||||||
|
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphWorkgroups );
|
||||||
return seed;
|
return seed;
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
@ -5947,15 +5947,20 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||||
//=== VK_AMDX_shader_enqueue ===
|
//=== VK_AMDX_shader_enqueue ===
|
||||||
|
|
||||||
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch ) const VULKAN_HPP_NOEXCEPT;
|
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize ) const VULKAN_HPP_NOEXCEPT;
|
||||||
|
|
||||||
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT;
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||||
|
|
||||||
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT;
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||||
|
|
||||||
void dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
void dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT;
|
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
|
|
||||||
@ -18376,43 +18381,55 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
return nodeIndex;
|
return nodeIndex;
|
||||||
}
|
}
|
||||||
|
|
||||||
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch ) const VULKAN_HPP_NOEXCEPT
|
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX &&
|
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX &&
|
||||||
"Function <vkCmdInitializeGraphScratchMemoryAMDX> requires <VK_AMDX_shader_enqueue>" );
|
"Function <vkCmdInitializeGraphScratchMemoryAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||||
|
|
||||||
getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ) );
|
getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||||
|
static_cast<VkPipeline>( executionGraph ),
|
||||||
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ) );
|
||||||
}
|
}
|
||||||
|
|
||||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphAMDX && "Function <vkCmdDispatchGraphAMDX> requires <VK_AMDX_shader_enqueue>" );
|
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphAMDX && "Function <vkCmdDispatchGraphAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||||
|
|
||||||
getDispatcher()->vkCmdDispatchGraphAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
getDispatcher()->vkCmdDispatchGraphAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||||
static_cast<VkDeviceAddress>( scratch ),
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ),
|
||||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||||
}
|
}
|
||||||
|
|
||||||
VULKAN_HPP_INLINE void
|
VULKAN_HPP_INLINE void
|
||||||
CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT
|
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphIndirectAMDX && "Function <vkCmdDispatchGraphIndirectAMDX> requires <VK_AMDX_shader_enqueue>" );
|
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphIndirectAMDX && "Function <vkCmdDispatchGraphIndirectAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||||
|
|
||||||
getDispatcher()->vkCmdDispatchGraphIndirectAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
getDispatcher()->vkCmdDispatchGraphIndirectAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||||
static_cast<VkDeviceAddress>( scratch ),
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ),
|
||||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||||
}
|
}
|
||||||
|
|
||||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT
|
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX &&
|
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX &&
|
||||||
"Function <vkCmdDispatchGraphIndirectCountAMDX> requires <VK_AMDX_shader_enqueue>" );
|
"Function <vkCmdDispatchGraphIndirectCountAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||||
|
|
||||||
getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX(
|
getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||||
static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ), static_cast<VkDeviceAddress>( countInfo ) );
|
static_cast<VkDeviceAddress>( scratch ),
|
||||||
|
static_cast<VkDeviceSize>( scratchSize ),
|
||||||
|
static_cast<VkDeviceAddress>( countInfo ) );
|
||||||
}
|
}
|
||||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
|
|
||||||
|
@ -34505,9 +34505,14 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
static VULKAN_HPP_CONST_OR_CONSTEXPR StructureType structureType = StructureType::eExecutionGraphPipelineScratchSizeAMDX;
|
static VULKAN_HPP_CONST_OR_CONSTEXPR StructureType structureType = StructureType::eExecutionGraphPipelineScratchSizeAMDX;
|
||||||
|
|
||||||
# if !defined( VULKAN_HPP_NO_STRUCT_CONSTRUCTORS )
|
# if !defined( VULKAN_HPP_NO_STRUCT_CONSTRUCTORS )
|
||||||
VULKAN_HPP_CONSTEXPR ExecutionGraphPipelineScratchSizeAMDX( VULKAN_HPP_NAMESPACE::DeviceSize size_ = {}, void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
VULKAN_HPP_CONSTEXPR ExecutionGraphPipelineScratchSizeAMDX( VULKAN_HPP_NAMESPACE::DeviceSize minSize_ = {},
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize maxSize_ = {},
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize sizeGranularity_ = {},
|
||||||
|
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||||
: pNext{ pNext_ }
|
: pNext{ pNext_ }
|
||||||
, size{ size_ }
|
, minSize{ minSize_ }
|
||||||
|
, maxSize{ maxSize_ }
|
||||||
|
, sizeGranularity{ sizeGranularity_ }
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -34534,9 +34539,21 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
return *this;
|
return *this;
|
||||||
}
|
}
|
||||||
|
|
||||||
VULKAN_HPP_CONSTEXPR_14 ExecutionGraphPipelineScratchSizeAMDX & setSize( VULKAN_HPP_NAMESPACE::DeviceSize size_ ) VULKAN_HPP_NOEXCEPT
|
VULKAN_HPP_CONSTEXPR_14 ExecutionGraphPipelineScratchSizeAMDX & setMinSize( VULKAN_HPP_NAMESPACE::DeviceSize minSize_ ) VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
size = size_;
|
minSize = minSize_;
|
||||||
|
return *this;
|
||||||
|
}
|
||||||
|
|
||||||
|
VULKAN_HPP_CONSTEXPR_14 ExecutionGraphPipelineScratchSizeAMDX & setMaxSize( VULKAN_HPP_NAMESPACE::DeviceSize maxSize_ ) VULKAN_HPP_NOEXCEPT
|
||||||
|
{
|
||||||
|
maxSize = maxSize_;
|
||||||
|
return *this;
|
||||||
|
}
|
||||||
|
|
||||||
|
VULKAN_HPP_CONSTEXPR_14 ExecutionGraphPipelineScratchSizeAMDX & setSizeGranularity( VULKAN_HPP_NAMESPACE::DeviceSize sizeGranularity_ ) VULKAN_HPP_NOEXCEPT
|
||||||
|
{
|
||||||
|
sizeGranularity = sizeGranularity_;
|
||||||
return *this;
|
return *this;
|
||||||
}
|
}
|
||||||
# endif /*VULKAN_HPP_NO_STRUCT_SETTERS*/
|
# endif /*VULKAN_HPP_NO_STRUCT_SETTERS*/
|
||||||
@ -34555,11 +34572,15 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
# if 14 <= VULKAN_HPP_CPP_VERSION
|
# if 14 <= VULKAN_HPP_CPP_VERSION
|
||||||
auto
|
auto
|
||||||
# else
|
# else
|
||||||
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &, void * const &, VULKAN_HPP_NAMESPACE::DeviceSize const &>
|
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &,
|
||||||
|
void * const &,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize const &,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize const &,
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize const &>
|
||||||
# endif
|
# endif
|
||||||
reflect() const VULKAN_HPP_NOEXCEPT
|
reflect() const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
return std::tie( sType, pNext, size );
|
return std::tie( sType, pNext, minSize, maxSize, sizeGranularity );
|
||||||
}
|
}
|
||||||
# endif
|
# endif
|
||||||
|
|
||||||
@ -34571,7 +34592,8 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
# if defined( VULKAN_HPP_USE_REFLECT )
|
# if defined( VULKAN_HPP_USE_REFLECT )
|
||||||
return this->reflect() == rhs.reflect();
|
return this->reflect() == rhs.reflect();
|
||||||
# else
|
# else
|
||||||
return ( sType == rhs.sType ) && ( pNext == rhs.pNext ) && ( size == rhs.size );
|
return ( sType == rhs.sType ) && ( pNext == rhs.pNext ) && ( minSize == rhs.minSize ) && ( maxSize == rhs.maxSize ) &&
|
||||||
|
( sizeGranularity == rhs.sizeGranularity );
|
||||||
# endif
|
# endif
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -34582,9 +34604,11 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
# endif
|
# endif
|
||||||
|
|
||||||
public:
|
public:
|
||||||
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::eExecutionGraphPipelineScratchSizeAMDX;
|
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::eExecutionGraphPipelineScratchSizeAMDX;
|
||||||
void * pNext = {};
|
void * pNext = {};
|
||||||
VULKAN_HPP_NAMESPACE::DeviceSize size = {};
|
VULKAN_HPP_NAMESPACE::DeviceSize minSize = {};
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize maxSize = {};
|
||||||
|
VULKAN_HPP_NAMESPACE::DeviceSize sizeGranularity = {};
|
||||||
};
|
};
|
||||||
|
|
||||||
template <>
|
template <>
|
||||||
@ -39792,7 +39816,7 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT indirectCommandsLayout_ = {},
|
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT indirectCommandsLayout_ = {},
|
||||||
uint32_t maxSequenceCount_ = {},
|
uint32_t maxSequenceCount_ = {},
|
||||||
uint32_t maxDrawCount_ = {},
|
uint32_t maxDrawCount_ = {},
|
||||||
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
const void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||||
: pNext{ pNext_ }
|
: pNext{ pNext_ }
|
||||||
, indirectExecutionSet{ indirectExecutionSet_ }
|
, indirectExecutionSet{ indirectExecutionSet_ }
|
||||||
, indirectCommandsLayout{ indirectCommandsLayout_ }
|
, indirectCommandsLayout{ indirectCommandsLayout_ }
|
||||||
@ -39818,7 +39842,7 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
}
|
}
|
||||||
|
|
||||||
#if !defined( VULKAN_HPP_NO_STRUCT_SETTERS )
|
#if !defined( VULKAN_HPP_NO_STRUCT_SETTERS )
|
||||||
VULKAN_HPP_CONSTEXPR_14 GeneratedCommandsMemoryRequirementsInfoEXT & setPNext( void * pNext_ ) VULKAN_HPP_NOEXCEPT
|
VULKAN_HPP_CONSTEXPR_14 GeneratedCommandsMemoryRequirementsInfoEXT & setPNext( const void * pNext_ ) VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
pNext = pNext_;
|
pNext = pNext_;
|
||||||
return *this;
|
return *this;
|
||||||
@ -39866,7 +39890,7 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
auto
|
auto
|
||||||
# else
|
# else
|
||||||
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &,
|
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &,
|
||||||
void * const &,
|
const void * const &,
|
||||||
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT const &,
|
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT const &,
|
||||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT const &,
|
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT const &,
|
||||||
uint32_t const &,
|
uint32_t const &,
|
||||||
@ -39899,7 +39923,7 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
|
|
||||||
public:
|
public:
|
||||||
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::eGeneratedCommandsMemoryRequirementsInfoEXT;
|
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::eGeneratedCommandsMemoryRequirementsInfoEXT;
|
||||||
void * pNext = {};
|
const void * pNext = {};
|
||||||
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT indirectExecutionSet = {};
|
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT indirectExecutionSet = {};
|
||||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT indirectCommandsLayout = {};
|
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT indirectCommandsLayout = {};
|
||||||
uint32_t maxSequenceCount = {};
|
uint32_t maxSequenceCount = {};
|
||||||
@ -87116,10 +87140,12 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
static VULKAN_HPP_CONST_OR_CONSTEXPR StructureType structureType = StructureType::ePhysicalDeviceShaderEnqueueFeaturesAMDX;
|
static VULKAN_HPP_CONST_OR_CONSTEXPR StructureType structureType = StructureType::ePhysicalDeviceShaderEnqueueFeaturesAMDX;
|
||||||
|
|
||||||
# if !defined( VULKAN_HPP_NO_STRUCT_CONSTRUCTORS )
|
# if !defined( VULKAN_HPP_NO_STRUCT_CONSTRUCTORS )
|
||||||
VULKAN_HPP_CONSTEXPR PhysicalDeviceShaderEnqueueFeaturesAMDX( VULKAN_HPP_NAMESPACE::Bool32 shaderEnqueue_ = {},
|
VULKAN_HPP_CONSTEXPR PhysicalDeviceShaderEnqueueFeaturesAMDX( VULKAN_HPP_NAMESPACE::Bool32 shaderEnqueue_ = {},
|
||||||
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
VULKAN_HPP_NAMESPACE::Bool32 shaderMeshEnqueue_ = {},
|
||||||
|
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||||
: pNext{ pNext_ }
|
: pNext{ pNext_ }
|
||||||
, shaderEnqueue{ shaderEnqueue_ }
|
, shaderEnqueue{ shaderEnqueue_ }
|
||||||
|
, shaderMeshEnqueue{ shaderMeshEnqueue_ }
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -87151,6 +87177,13 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
shaderEnqueue = shaderEnqueue_;
|
shaderEnqueue = shaderEnqueue_;
|
||||||
return *this;
|
return *this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueueFeaturesAMDX &
|
||||||
|
setShaderMeshEnqueue( VULKAN_HPP_NAMESPACE::Bool32 shaderMeshEnqueue_ ) VULKAN_HPP_NOEXCEPT
|
||||||
|
{
|
||||||
|
shaderMeshEnqueue = shaderMeshEnqueue_;
|
||||||
|
return *this;
|
||||||
|
}
|
||||||
# endif /*VULKAN_HPP_NO_STRUCT_SETTERS*/
|
# endif /*VULKAN_HPP_NO_STRUCT_SETTERS*/
|
||||||
|
|
||||||
operator VkPhysicalDeviceShaderEnqueueFeaturesAMDX const &() const VULKAN_HPP_NOEXCEPT
|
operator VkPhysicalDeviceShaderEnqueueFeaturesAMDX const &() const VULKAN_HPP_NOEXCEPT
|
||||||
@ -87167,11 +87200,11 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
# if 14 <= VULKAN_HPP_CPP_VERSION
|
# if 14 <= VULKAN_HPP_CPP_VERSION
|
||||||
auto
|
auto
|
||||||
# else
|
# else
|
||||||
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &, void * const &, VULKAN_HPP_NAMESPACE::Bool32 const &>
|
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &, void * const &, VULKAN_HPP_NAMESPACE::Bool32 const &, VULKAN_HPP_NAMESPACE::Bool32 const &>
|
||||||
# endif
|
# endif
|
||||||
reflect() const VULKAN_HPP_NOEXCEPT
|
reflect() const VULKAN_HPP_NOEXCEPT
|
||||||
{
|
{
|
||||||
return std::tie( sType, pNext, shaderEnqueue );
|
return std::tie( sType, pNext, shaderEnqueue, shaderMeshEnqueue );
|
||||||
}
|
}
|
||||||
# endif
|
# endif
|
||||||
|
|
||||||
@ -87183,7 +87216,7 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
# if defined( VULKAN_HPP_USE_REFLECT )
|
# if defined( VULKAN_HPP_USE_REFLECT )
|
||||||
return this->reflect() == rhs.reflect();
|
return this->reflect() == rhs.reflect();
|
||||||
# else
|
# else
|
||||||
return ( sType == rhs.sType ) && ( pNext == rhs.pNext ) && ( shaderEnqueue == rhs.shaderEnqueue );
|
return ( sType == rhs.sType ) && ( pNext == rhs.pNext ) && ( shaderEnqueue == rhs.shaderEnqueue ) && ( shaderMeshEnqueue == rhs.shaderMeshEnqueue );
|
||||||
# endif
|
# endif
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -87194,9 +87227,10 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
# endif
|
# endif
|
||||||
|
|
||||||
public:
|
public:
|
||||||
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::ePhysicalDeviceShaderEnqueueFeaturesAMDX;
|
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::ePhysicalDeviceShaderEnqueueFeaturesAMDX;
|
||||||
void * pNext = {};
|
void * pNext = {};
|
||||||
VULKAN_HPP_NAMESPACE::Bool32 shaderEnqueue = {};
|
VULKAN_HPP_NAMESPACE::Bool32 shaderEnqueue = {};
|
||||||
|
VULKAN_HPP_NAMESPACE::Bool32 shaderMeshEnqueue = {};
|
||||||
};
|
};
|
||||||
|
|
||||||
template <>
|
template <>
|
||||||
@ -87215,22 +87249,26 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
static VULKAN_HPP_CONST_OR_CONSTEXPR StructureType structureType = StructureType::ePhysicalDeviceShaderEnqueuePropertiesAMDX;
|
static VULKAN_HPP_CONST_OR_CONSTEXPR StructureType structureType = StructureType::ePhysicalDeviceShaderEnqueuePropertiesAMDX;
|
||||||
|
|
||||||
# if !defined( VULKAN_HPP_NO_STRUCT_CONSTRUCTORS )
|
# if !defined( VULKAN_HPP_NO_STRUCT_CONSTRUCTORS )
|
||||||
VULKAN_HPP_CONSTEXPR PhysicalDeviceShaderEnqueuePropertiesAMDX( uint32_t maxExecutionGraphDepth_ = {},
|
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueuePropertiesAMDX( uint32_t maxExecutionGraphDepth_ = {},
|
||||||
uint32_t maxExecutionGraphShaderOutputNodes_ = {},
|
uint32_t maxExecutionGraphShaderOutputNodes_ = {},
|
||||||
uint32_t maxExecutionGraphShaderPayloadSize_ = {},
|
uint32_t maxExecutionGraphShaderPayloadSize_ = {},
|
||||||
uint32_t maxExecutionGraphShaderPayloadCount_ = {},
|
uint32_t maxExecutionGraphShaderPayloadCount_ = {},
|
||||||
uint32_t executionGraphDispatchAddressAlignment_ = {},
|
uint32_t executionGraphDispatchAddressAlignment_ = {},
|
||||||
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
std::array<uint32_t, 3> const & maxExecutionGraphWorkgroupCount_ = {},
|
||||||
|
uint32_t maxExecutionGraphWorkgroups_ = {},
|
||||||
|
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||||
: pNext{ pNext_ }
|
: pNext{ pNext_ }
|
||||||
, maxExecutionGraphDepth{ maxExecutionGraphDepth_ }
|
, maxExecutionGraphDepth{ maxExecutionGraphDepth_ }
|
||||||
, maxExecutionGraphShaderOutputNodes{ maxExecutionGraphShaderOutputNodes_ }
|
, maxExecutionGraphShaderOutputNodes{ maxExecutionGraphShaderOutputNodes_ }
|
||||||
, maxExecutionGraphShaderPayloadSize{ maxExecutionGraphShaderPayloadSize_ }
|
, maxExecutionGraphShaderPayloadSize{ maxExecutionGraphShaderPayloadSize_ }
|
||||||
, maxExecutionGraphShaderPayloadCount{ maxExecutionGraphShaderPayloadCount_ }
|
, maxExecutionGraphShaderPayloadCount{ maxExecutionGraphShaderPayloadCount_ }
|
||||||
, executionGraphDispatchAddressAlignment{ executionGraphDispatchAddressAlignment_ }
|
, executionGraphDispatchAddressAlignment{ executionGraphDispatchAddressAlignment_ }
|
||||||
|
, maxExecutionGraphWorkgroupCount{ maxExecutionGraphWorkgroupCount_ }
|
||||||
|
, maxExecutionGraphWorkgroups{ maxExecutionGraphWorkgroups_ }
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
VULKAN_HPP_CONSTEXPR PhysicalDeviceShaderEnqueuePropertiesAMDX( PhysicalDeviceShaderEnqueuePropertiesAMDX const & rhs ) VULKAN_HPP_NOEXCEPT = default;
|
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueuePropertiesAMDX( PhysicalDeviceShaderEnqueuePropertiesAMDX const & rhs ) VULKAN_HPP_NOEXCEPT = default;
|
||||||
|
|
||||||
PhysicalDeviceShaderEnqueuePropertiesAMDX( VkPhysicalDeviceShaderEnqueuePropertiesAMDX const & rhs ) VULKAN_HPP_NOEXCEPT
|
PhysicalDeviceShaderEnqueuePropertiesAMDX( VkPhysicalDeviceShaderEnqueuePropertiesAMDX const & rhs ) VULKAN_HPP_NOEXCEPT
|
||||||
: PhysicalDeviceShaderEnqueuePropertiesAMDX( *reinterpret_cast<PhysicalDeviceShaderEnqueuePropertiesAMDX const *>( &rhs ) )
|
: PhysicalDeviceShaderEnqueuePropertiesAMDX( *reinterpret_cast<PhysicalDeviceShaderEnqueuePropertiesAMDX const *>( &rhs ) )
|
||||||
@ -87286,6 +87324,20 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
executionGraphDispatchAddressAlignment = executionGraphDispatchAddressAlignment_;
|
executionGraphDispatchAddressAlignment = executionGraphDispatchAddressAlignment_;
|
||||||
return *this;
|
return *this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueuePropertiesAMDX &
|
||||||
|
setMaxExecutionGraphWorkgroupCount( std::array<uint32_t, 3> maxExecutionGraphWorkgroupCount_ ) VULKAN_HPP_NOEXCEPT
|
||||||
|
{
|
||||||
|
maxExecutionGraphWorkgroupCount = maxExecutionGraphWorkgroupCount_;
|
||||||
|
return *this;
|
||||||
|
}
|
||||||
|
|
||||||
|
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueuePropertiesAMDX &
|
||||||
|
setMaxExecutionGraphWorkgroups( uint32_t maxExecutionGraphWorkgroups_ ) VULKAN_HPP_NOEXCEPT
|
||||||
|
{
|
||||||
|
maxExecutionGraphWorkgroups = maxExecutionGraphWorkgroups_;
|
||||||
|
return *this;
|
||||||
|
}
|
||||||
# endif /*VULKAN_HPP_NO_STRUCT_SETTERS*/
|
# endif /*VULKAN_HPP_NO_STRUCT_SETTERS*/
|
||||||
|
|
||||||
operator VkPhysicalDeviceShaderEnqueuePropertiesAMDX const &() const VULKAN_HPP_NOEXCEPT
|
operator VkPhysicalDeviceShaderEnqueuePropertiesAMDX const &() const VULKAN_HPP_NOEXCEPT
|
||||||
@ -87308,6 +87360,8 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
uint32_t const &,
|
uint32_t const &,
|
||||||
uint32_t const &,
|
uint32_t const &,
|
||||||
uint32_t const &,
|
uint32_t const &,
|
||||||
|
uint32_t const &,
|
||||||
|
VULKAN_HPP_NAMESPACE::ArrayWrapper1D<uint32_t, 3> const &,
|
||||||
uint32_t const &>
|
uint32_t const &>
|
||||||
# endif
|
# endif
|
||||||
reflect() const VULKAN_HPP_NOEXCEPT
|
reflect() const VULKAN_HPP_NOEXCEPT
|
||||||
@ -87318,7 +87372,9 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
maxExecutionGraphShaderOutputNodes,
|
maxExecutionGraphShaderOutputNodes,
|
||||||
maxExecutionGraphShaderPayloadSize,
|
maxExecutionGraphShaderPayloadSize,
|
||||||
maxExecutionGraphShaderPayloadCount,
|
maxExecutionGraphShaderPayloadCount,
|
||||||
executionGraphDispatchAddressAlignment );
|
executionGraphDispatchAddressAlignment,
|
||||||
|
maxExecutionGraphWorkgroupCount,
|
||||||
|
maxExecutionGraphWorkgroups );
|
||||||
}
|
}
|
||||||
# endif
|
# endif
|
||||||
|
|
||||||
@ -87334,7 +87390,8 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
( maxExecutionGraphShaderOutputNodes == rhs.maxExecutionGraphShaderOutputNodes ) &&
|
( maxExecutionGraphShaderOutputNodes == rhs.maxExecutionGraphShaderOutputNodes ) &&
|
||||||
( maxExecutionGraphShaderPayloadSize == rhs.maxExecutionGraphShaderPayloadSize ) &&
|
( maxExecutionGraphShaderPayloadSize == rhs.maxExecutionGraphShaderPayloadSize ) &&
|
||||||
( maxExecutionGraphShaderPayloadCount == rhs.maxExecutionGraphShaderPayloadCount ) &&
|
( maxExecutionGraphShaderPayloadCount == rhs.maxExecutionGraphShaderPayloadCount ) &&
|
||||||
( executionGraphDispatchAddressAlignment == rhs.executionGraphDispatchAddressAlignment );
|
( executionGraphDispatchAddressAlignment == rhs.executionGraphDispatchAddressAlignment ) &&
|
||||||
|
( maxExecutionGraphWorkgroupCount == rhs.maxExecutionGraphWorkgroupCount ) && ( maxExecutionGraphWorkgroups == rhs.maxExecutionGraphWorkgroups );
|
||||||
# endif
|
# endif
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -87345,13 +87402,15 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
# endif
|
# endif
|
||||||
|
|
||||||
public:
|
public:
|
||||||
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::ePhysicalDeviceShaderEnqueuePropertiesAMDX;
|
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::ePhysicalDeviceShaderEnqueuePropertiesAMDX;
|
||||||
void * pNext = {};
|
void * pNext = {};
|
||||||
uint32_t maxExecutionGraphDepth = {};
|
uint32_t maxExecutionGraphDepth = {};
|
||||||
uint32_t maxExecutionGraphShaderOutputNodes = {};
|
uint32_t maxExecutionGraphShaderOutputNodes = {};
|
||||||
uint32_t maxExecutionGraphShaderPayloadSize = {};
|
uint32_t maxExecutionGraphShaderPayloadSize = {};
|
||||||
uint32_t maxExecutionGraphShaderPayloadCount = {};
|
uint32_t maxExecutionGraphShaderPayloadCount = {};
|
||||||
uint32_t executionGraphDispatchAddressAlignment = {};
|
uint32_t executionGraphDispatchAddressAlignment = {};
|
||||||
|
VULKAN_HPP_NAMESPACE::ArrayWrapper1D<uint32_t, 3> maxExecutionGraphWorkgroupCount = {};
|
||||||
|
uint32_t maxExecutionGraphWorkgroups = {};
|
||||||
};
|
};
|
||||||
|
|
||||||
template <>
|
template <>
|
||||||
|
@ -3385,6 +3385,10 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
result += "AllowDerivatives | ";
|
result += "AllowDerivatives | ";
|
||||||
if ( value & PipelineCreateFlagBits2KHR::eDerivative )
|
if ( value & PipelineCreateFlagBits2KHR::eDerivative )
|
||||||
result += "Derivative | ";
|
result += "Derivative | ";
|
||||||
|
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||||
|
if ( value & PipelineCreateFlagBits2KHR::eExecutionGraphAMDX )
|
||||||
|
result += "ExecutionGraphAMDX | ";
|
||||||
|
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
if ( value & PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT )
|
if ( value & PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT )
|
||||||
result += "EnableLegacyDitheringEXT | ";
|
result += "EnableLegacyDitheringEXT | ";
|
||||||
if ( value & PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex )
|
if ( value & PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex )
|
||||||
@ -8919,6 +8923,9 @@ namespace VULKAN_HPP_NAMESPACE
|
|||||||
case PipelineCreateFlagBits2KHR::eDisableOptimization: return "DisableOptimization";
|
case PipelineCreateFlagBits2KHR::eDisableOptimization: return "DisableOptimization";
|
||||||
case PipelineCreateFlagBits2KHR::eAllowDerivatives: return "AllowDerivatives";
|
case PipelineCreateFlagBits2KHR::eAllowDerivatives: return "AllowDerivatives";
|
||||||
case PipelineCreateFlagBits2KHR::eDerivative: return "Derivative";
|
case PipelineCreateFlagBits2KHR::eDerivative: return "Derivative";
|
||||||
|
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||||
|
case PipelineCreateFlagBits2KHR::eExecutionGraphAMDX: return "ExecutionGraphAMDX";
|
||||||
|
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||||
case PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT: return "EnableLegacyDitheringEXT";
|
case PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT: return "EnableLegacyDitheringEXT";
|
||||||
case PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex: return "ViewIndexFromDeviceIndex";
|
case PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex: return "ViewIndexFromDeviceIndex";
|
||||||
case PipelineCreateFlagBits2KHR::eDispatchBase: return "DispatchBase";
|
case PipelineCreateFlagBits2KHR::eDispatchBase: return "DispatchBase";
|
||||||
|
Loading…
Reference in New Issue
Block a user