2023-08-12 20:10:16 +00:00
|
|
|
#include "config.h"
|
|
|
|
|
|
|
|
#include "gskvulkanframeprivate.h"
|
|
|
|
|
|
|
|
#include "gskgpuopprivate.h"
|
2023-08-21 00:18:37 +00:00
|
|
|
#include "gskvulkanbufferprivate.h"
|
2023-10-03 19:04:21 +00:00
|
|
|
#include "gskvulkandescriptorsprivate.h"
|
2023-08-12 20:10:16 +00:00
|
|
|
#include "gskvulkandeviceprivate.h"
|
2023-08-21 00:18:37 +00:00
|
|
|
#include "gskvulkanimageprivate.h"
|
2023-08-12 20:10:16 +00:00
|
|
|
|
|
|
|
#include "gdk/gdkdisplayprivate.h"
|
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
#define GDK_ARRAY_NAME gsk_descriptor_image_infos
|
|
|
|
#define GDK_ARRAY_TYPE_NAME GskDescriptorImageInfos
|
|
|
|
#define GDK_ARRAY_ELEMENT_TYPE VkDescriptorImageInfo
|
|
|
|
#define GDK_ARRAY_BY_VALUE 1
|
|
|
|
#define GDK_ARRAY_PREALLOC 128
|
|
|
|
#define GDK_ARRAY_NO_MEMSET 1
|
|
|
|
#include "gdk/gdkarrayimpl.c"
|
|
|
|
|
|
|
|
#define GDK_ARRAY_NAME gsk_descriptor_buffer_infos
|
|
|
|
#define GDK_ARRAY_TYPE_NAME GskDescriptorBufferInfos
|
|
|
|
#define GDK_ARRAY_ELEMENT_TYPE VkDescriptorBufferInfo
|
|
|
|
#define GDK_ARRAY_BY_VALUE 1
|
|
|
|
#define GDK_ARRAY_PREALLOC 32
|
|
|
|
#define GDK_ARRAY_NO_MEMSET 1
|
|
|
|
#include "gdk/gdkarrayimpl.c"
|
|
|
|
|
2023-08-12 20:10:16 +00:00
|
|
|
struct _GskVulkanFrame
|
|
|
|
{
|
|
|
|
GskGpuFrame parent_instance;
|
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
GskDescriptorImageInfos descriptor_images;
|
|
|
|
GskDescriptorBufferInfos descriptor_buffers;
|
|
|
|
|
2023-08-12 20:10:16 +00:00
|
|
|
VkFence vk_fence;
|
|
|
|
VkCommandBuffer vk_command_buffer;
|
2023-08-21 00:18:37 +00:00
|
|
|
VkDescriptorPool vk_descriptor_pool;
|
2023-08-12 20:10:16 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
struct _GskVulkanFrameClass
|
|
|
|
{
|
|
|
|
GskGpuFrameClass parent_class;
|
|
|
|
};
|
|
|
|
|
|
|
|
G_DEFINE_TYPE (GskVulkanFrame, gsk_vulkan_frame, GSK_TYPE_GPU_FRAME)
|
|
|
|
|
|
|
|
static gboolean
|
|
|
|
gsk_vulkan_frame_is_busy (GskGpuFrame *frame)
|
|
|
|
{
|
|
|
|
GskVulkanFrame *self = GSK_VULKAN_FRAME (frame);
|
|
|
|
VkDevice device;
|
|
|
|
|
|
|
|
device = gsk_vulkan_device_get_vk_device (GSK_VULKAN_DEVICE (gsk_gpu_frame_get_device (frame)));
|
|
|
|
|
|
|
|
return vkGetFenceStatus (device, self->vk_fence) == VK_NOT_READY;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
gsk_vulkan_frame_setup (GskGpuFrame *frame)
|
|
|
|
{
|
|
|
|
GskVulkanFrame *self = GSK_VULKAN_FRAME (frame);
|
|
|
|
GskVulkanDevice *device;
|
|
|
|
VkDevice vk_device;
|
|
|
|
VkCommandPool vk_command_pool;
|
|
|
|
|
|
|
|
device = GSK_VULKAN_DEVICE (gsk_gpu_frame_get_device (frame));
|
|
|
|
vk_device = gsk_vulkan_device_get_vk_device (device);
|
|
|
|
vk_command_pool = gsk_vulkan_device_get_vk_command_pool (device);
|
|
|
|
|
|
|
|
GSK_VK_CHECK (vkAllocateCommandBuffers, vk_device,
|
|
|
|
&(VkCommandBufferAllocateInfo) {
|
|
|
|
.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO,
|
|
|
|
.commandPool = vk_command_pool,
|
|
|
|
.level = VK_COMMAND_BUFFER_LEVEL_PRIMARY,
|
|
|
|
.commandBufferCount = 1,
|
|
|
|
},
|
|
|
|
&self->vk_command_buffer);
|
|
|
|
|
|
|
|
GSK_VK_CHECK (vkCreateFence, vk_device,
|
|
|
|
&(VkFenceCreateInfo) {
|
|
|
|
.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO,
|
|
|
|
.flags = VK_FENCE_CREATE_SIGNALED_BIT
|
|
|
|
},
|
|
|
|
NULL,
|
|
|
|
&self->vk_fence);
|
2023-08-21 00:18:37 +00:00
|
|
|
|
|
|
|
GSK_VK_CHECK (vkCreateDescriptorPool, vk_device,
|
|
|
|
&(VkDescriptorPoolCreateInfo) {
|
|
|
|
.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_POOL_CREATE_INFO,
|
|
|
|
.flags = VK_DESCRIPTOR_POOL_CREATE_UPDATE_AFTER_BIND_BIT,
|
|
|
|
.maxSets = GSK_VULKAN_N_DESCRIPTOR_SETS,
|
|
|
|
.poolSizeCount = GSK_VULKAN_N_DESCRIPTOR_SETS,
|
|
|
|
.pPoolSizes = (VkDescriptorPoolSize[GSK_VULKAN_N_DESCRIPTOR_SETS]) {
|
|
|
|
{
|
|
|
|
.type = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER,
|
|
|
|
.descriptorCount = gsk_vulkan_device_get_max_descriptors (device),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.type = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER,
|
|
|
|
.descriptorCount = gsk_vulkan_device_get_max_descriptors (device),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
NULL,
|
|
|
|
&self->vk_descriptor_pool);
|
|
|
|
|
2023-08-12 20:10:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
gsk_vulkan_frame_cleanup (GskGpuFrame *frame)
|
|
|
|
{
|
|
|
|
GskVulkanFrame *self = GSK_VULKAN_FRAME (frame);
|
2023-08-21 00:18:37 +00:00
|
|
|
VkDevice vk_device;
|
2023-08-12 20:10:16 +00:00
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
vk_device = gsk_vulkan_device_get_vk_device (GSK_VULKAN_DEVICE (gsk_gpu_frame_get_device (frame)));
|
2023-08-12 20:10:16 +00:00
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
GSK_VK_CHECK (vkWaitForFences, vk_device,
|
2023-08-12 20:10:16 +00:00
|
|
|
1,
|
|
|
|
&self->vk_fence,
|
|
|
|
VK_TRUE,
|
|
|
|
INT64_MAX);
|
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
GSK_VK_CHECK (vkResetFences, vk_device,
|
2023-08-12 20:10:16 +00:00
|
|
|
1,
|
|
|
|
&self->vk_fence);
|
|
|
|
|
|
|
|
GSK_VK_CHECK (vkResetCommandBuffer, self->vk_command_buffer,
|
|
|
|
0);
|
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
GSK_VK_CHECK (vkResetDescriptorPool, vk_device,
|
|
|
|
self->vk_descriptor_pool,
|
|
|
|
0);
|
|
|
|
gsk_descriptor_image_infos_set_size (&self->descriptor_images, 0);
|
|
|
|
gsk_descriptor_buffer_infos_set_size (&self->descriptor_buffers, 0);
|
|
|
|
|
2023-08-12 20:10:16 +00:00
|
|
|
GSK_GPU_FRAME_CLASS (gsk_vulkan_frame_parent_class)->cleanup (frame);
|
|
|
|
}
|
|
|
|
|
2023-10-03 19:04:21 +00:00
|
|
|
guint32
|
|
|
|
gsk_vulkan_frame_add_image (GskVulkanFrame *self,
|
|
|
|
GskGpuImage *image,
|
|
|
|
GskGpuSampler sampler)
|
2023-08-21 00:18:37 +00:00
|
|
|
{
|
|
|
|
GskVulkanDevice *device;
|
|
|
|
guint32 result;
|
|
|
|
|
2023-10-03 19:04:21 +00:00
|
|
|
device = GSK_VULKAN_DEVICE (gsk_gpu_frame_get_device (GSK_GPU_FRAME (self)));
|
2023-08-21 00:18:37 +00:00
|
|
|
|
|
|
|
result = gsk_descriptor_image_infos_get_size (&self->descriptor_images);
|
|
|
|
g_assert (result < gsk_vulkan_device_get_max_descriptors (device));
|
|
|
|
|
|
|
|
gsk_descriptor_image_infos_append (&self->descriptor_images,
|
|
|
|
&(VkDescriptorImageInfo) {
|
|
|
|
.sampler = gsk_vulkan_device_get_vk_sampler (device, sampler),
|
|
|
|
.imageView = gsk_vulkan_image_get_vk_image_view (GSK_VULKAN_IMAGE (image)),
|
|
|
|
.imageLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL
|
|
|
|
});
|
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2023-08-30 19:57:41 +00:00
|
|
|
static guint32
|
|
|
|
gsk_vulkan_frame_get_buffer_descriptor (GskVulkanFrame *self,
|
|
|
|
GskGpuBuffer *buffer)
|
|
|
|
{
|
|
|
|
guint32 result;
|
|
|
|
|
|
|
|
result = gsk_descriptor_buffer_infos_get_size (&self->descriptor_buffers);
|
|
|
|
gsk_descriptor_buffer_infos_append (&self->descriptor_buffers,
|
|
|
|
&(VkDescriptorBufferInfo) {
|
|
|
|
.buffer = gsk_vulkan_buffer_get_vk_buffer (GSK_VULKAN_BUFFER (buffer)),
|
|
|
|
.offset = 0,
|
|
|
|
.range = VK_WHOLE_SIZE
|
|
|
|
});
|
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
static void
|
|
|
|
gsk_vulkan_frame_prepare_descriptor_sets (GskVulkanFrame *self)
|
|
|
|
{
|
|
|
|
GskVulkanDevice *device;
|
|
|
|
VkDevice vk_device;
|
|
|
|
VkWriteDescriptorSet write_descriptor_sets[GSK_VULKAN_N_DESCRIPTOR_SETS];
|
|
|
|
gsize n_descriptor_sets;
|
|
|
|
VkDescriptorSet descriptor_sets[GSK_VULKAN_N_DESCRIPTOR_SETS];
|
|
|
|
|
|
|
|
device = GSK_VULKAN_DEVICE (gsk_gpu_frame_get_device (GSK_GPU_FRAME (self)));
|
|
|
|
vk_device = gsk_vulkan_device_get_vk_device (device);
|
|
|
|
|
|
|
|
GSK_VK_CHECK (vkAllocateDescriptorSets, vk_device,
|
|
|
|
&(VkDescriptorSetAllocateInfo) {
|
|
|
|
.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_ALLOCATE_INFO,
|
|
|
|
.descriptorPool = self->vk_descriptor_pool,
|
|
|
|
.descriptorSetCount = GSK_VULKAN_N_DESCRIPTOR_SETS,
|
|
|
|
.pSetLayouts = (VkDescriptorSetLayout[GSK_VULKAN_N_DESCRIPTOR_SETS]) {
|
|
|
|
gsk_vulkan_device_get_vk_image_set_layout (device),
|
|
|
|
gsk_vulkan_device_get_vk_buffer_set_layout (device),
|
|
|
|
},
|
|
|
|
.pNext = &(VkDescriptorSetVariableDescriptorCountAllocateInfo) {
|
|
|
|
.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_VARIABLE_DESCRIPTOR_COUNT_ALLOCATE_INFO,
|
|
|
|
.descriptorSetCount = GSK_VULKAN_N_DESCRIPTOR_SETS,
|
|
|
|
.pDescriptorCounts = (uint32_t[GSK_VULKAN_N_DESCRIPTOR_SETS]) {
|
|
|
|
MAX (1, gsk_descriptor_image_infos_get_size (&self->descriptor_images)),
|
|
|
|
MAX (1, gsk_descriptor_buffer_infos_get_size (&self->descriptor_buffers))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
descriptor_sets);
|
|
|
|
|
|
|
|
n_descriptor_sets = 0;
|
|
|
|
if (gsk_descriptor_image_infos_get_size (&self->descriptor_images) > 0)
|
|
|
|
{
|
|
|
|
write_descriptor_sets[n_descriptor_sets++] = (VkWriteDescriptorSet) {
|
|
|
|
.sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET,
|
|
|
|
.dstSet = descriptor_sets[GSK_VULKAN_IMAGE_SET_LAYOUT],
|
|
|
|
.dstBinding = 0,
|
|
|
|
.dstArrayElement = 0,
|
|
|
|
.descriptorCount = gsk_descriptor_image_infos_get_size (&self->descriptor_images),
|
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER,
|
|
|
|
.pImageInfo = gsk_descriptor_image_infos_get_data (&self->descriptor_images)
|
|
|
|
};
|
|
|
|
}
|
|
|
|
if (gsk_descriptor_buffer_infos_get_size (&self->descriptor_buffers) > 0)
|
|
|
|
{
|
|
|
|
write_descriptor_sets[n_descriptor_sets++] = (VkWriteDescriptorSet) {
|
|
|
|
.sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET,
|
|
|
|
.dstSet = descriptor_sets[GSK_VULKAN_BUFFER_SET_LAYOUT],
|
|
|
|
.dstBinding = 0,
|
|
|
|
.dstArrayElement = 0,
|
|
|
|
.descriptorCount = gsk_descriptor_buffer_infos_get_size (&self->descriptor_buffers),
|
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER,
|
|
|
|
.pBufferInfo = gsk_descriptor_buffer_infos_get_data (&self->descriptor_buffers)
|
|
|
|
};
|
|
|
|
}
|
|
|
|
|
|
|
|
vkUpdateDescriptorSets (vk_device,
|
|
|
|
n_descriptor_sets,
|
|
|
|
write_descriptor_sets,
|
|
|
|
0, NULL);
|
|
|
|
|
|
|
|
vkCmdBindDescriptorSets (self->vk_command_buffer,
|
|
|
|
VK_PIPELINE_BIND_POINT_GRAPHICS,
|
|
|
|
gsk_vulkan_device_get_vk_pipeline_layout (device),
|
|
|
|
0,
|
|
|
|
GSK_VULKAN_N_DESCRIPTOR_SETS,
|
|
|
|
descriptor_sets,
|
|
|
|
0,
|
|
|
|
NULL);
|
|
|
|
}
|
|
|
|
|
2023-10-03 19:04:21 +00:00
|
|
|
static GskGpuDescriptors *
|
|
|
|
gsk_vulkan_frame_create_descriptors (GskGpuFrame *frame)
|
|
|
|
{
|
|
|
|
return GSK_GPU_DESCRIPTORS (gsk_vulkan_descriptors_new (GSK_VULKAN_FRAME (frame)));
|
|
|
|
}
|
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
static GskGpuBuffer *
|
|
|
|
gsk_vulkan_frame_create_vertex_buffer (GskGpuFrame *frame,
|
|
|
|
gsize size)
|
|
|
|
{
|
|
|
|
return gsk_vulkan_buffer_new_vertex (GSK_VULKAN_DEVICE (gsk_gpu_frame_get_device (frame)), size);
|
|
|
|
}
|
|
|
|
|
2023-08-30 19:57:41 +00:00
|
|
|
static GskGpuBuffer *
|
|
|
|
gsk_vulkan_frame_create_storage_buffer (GskGpuFrame *frame,
|
|
|
|
gsize size)
|
|
|
|
{
|
|
|
|
return gsk_vulkan_buffer_new_storage (GSK_VULKAN_DEVICE (gsk_gpu_frame_get_device (frame)), size);
|
|
|
|
}
|
|
|
|
|
2023-08-12 20:10:16 +00:00
|
|
|
static void
|
2023-08-21 00:18:37 +00:00
|
|
|
gsk_vulkan_frame_submit (GskGpuFrame *frame,
|
|
|
|
GskGpuBuffer *vertex_buffer,
|
2023-08-30 19:57:41 +00:00
|
|
|
GskGpuBuffer *storage_buffer,
|
2023-08-21 00:18:37 +00:00
|
|
|
GskGpuOp *op)
|
2023-08-12 20:10:16 +00:00
|
|
|
{
|
|
|
|
GskVulkanFrame *self = GSK_VULKAN_FRAME (frame);
|
|
|
|
|
2023-08-30 19:57:41 +00:00
|
|
|
if (storage_buffer)
|
|
|
|
{
|
|
|
|
G_GNUC_UNUSED guint32 descriptor;
|
|
|
|
descriptor = gsk_vulkan_frame_get_buffer_descriptor (self, storage_buffer);
|
|
|
|
g_assert (descriptor == 0);
|
|
|
|
}
|
|
|
|
|
2023-08-12 20:10:16 +00:00
|
|
|
GSK_VK_CHECK (vkBeginCommandBuffer, self->vk_command_buffer,
|
|
|
|
&(VkCommandBufferBeginInfo) {
|
|
|
|
.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO,
|
|
|
|
.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT,
|
|
|
|
});
|
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
gsk_vulkan_frame_prepare_descriptor_sets (self);
|
|
|
|
|
|
|
|
if (vertex_buffer)
|
|
|
|
vkCmdBindVertexBuffers (self->vk_command_buffer,
|
|
|
|
0,
|
|
|
|
1,
|
|
|
|
(VkBuffer[1]) {
|
|
|
|
gsk_vulkan_buffer_get_vk_buffer (GSK_VULKAN_BUFFER (vertex_buffer))
|
|
|
|
},
|
|
|
|
(VkDeviceSize[1]) { 0 });
|
|
|
|
|
2023-08-12 20:10:16 +00:00
|
|
|
while (op)
|
|
|
|
{
|
2023-08-21 00:18:37 +00:00
|
|
|
op = gsk_gpu_op_vk_command (op, frame, VK_NULL_HANDLE, VK_FORMAT_UNDEFINED, self->vk_command_buffer);
|
2023-08-12 20:10:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
GSK_VK_CHECK (vkEndCommandBuffer, self->vk_command_buffer);
|
|
|
|
|
|
|
|
GSK_VK_CHECK (vkQueueSubmit, gsk_vulkan_device_get_vk_queue (GSK_VULKAN_DEVICE (gsk_gpu_frame_get_device (frame))),
|
|
|
|
1,
|
|
|
|
&(VkSubmitInfo) {
|
|
|
|
.sType = VK_STRUCTURE_TYPE_SUBMIT_INFO,
|
|
|
|
.commandBufferCount = 1,
|
|
|
|
.pCommandBuffers = &self->vk_command_buffer,
|
|
|
|
},
|
|
|
|
self->vk_fence);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
gsk_vulkan_frame_finalize (GObject *object)
|
|
|
|
{
|
|
|
|
GskVulkanFrame *self = GSK_VULKAN_FRAME (object);
|
|
|
|
GskVulkanDevice *device;
|
|
|
|
VkDevice vk_device;
|
|
|
|
VkCommandPool vk_command_pool;
|
|
|
|
|
|
|
|
device = GSK_VULKAN_DEVICE (gsk_gpu_frame_get_device (GSK_GPU_FRAME (self)));
|
|
|
|
vk_device = gsk_vulkan_device_get_vk_device (device);
|
|
|
|
vk_command_pool = gsk_vulkan_device_get_vk_command_pool (device);
|
|
|
|
|
2023-08-21 00:18:37 +00:00
|
|
|
vkDestroyDescriptorPool (vk_device,
|
|
|
|
self->vk_descriptor_pool,
|
|
|
|
NULL);
|
|
|
|
gsk_descriptor_image_infos_clear (&self->descriptor_images);
|
|
|
|
gsk_descriptor_buffer_infos_clear (&self->descriptor_buffers);
|
|
|
|
|
2023-08-12 20:10:16 +00:00
|
|
|
vkFreeCommandBuffers (vk_device,
|
|
|
|
vk_command_pool,
|
|
|
|
1, &self->vk_command_buffer);
|
|
|
|
vkDestroyFence (vk_device,
|
|
|
|
self->vk_fence,
|
|
|
|
NULL);
|
|
|
|
|
|
|
|
G_OBJECT_CLASS (gsk_vulkan_frame_parent_class)->finalize (object);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
gsk_vulkan_frame_class_init (GskVulkanFrameClass *klass)
|
|
|
|
{
|
|
|
|
GskGpuFrameClass *gpu_frame_class = GSK_GPU_FRAME_CLASS (klass);
|
|
|
|
GObjectClass *object_class = G_OBJECT_CLASS (klass);
|
|
|
|
|
|
|
|
gpu_frame_class->is_busy = gsk_vulkan_frame_is_busy;
|
|
|
|
gpu_frame_class->setup = gsk_vulkan_frame_setup;
|
|
|
|
gpu_frame_class->cleanup = gsk_vulkan_frame_cleanup;
|
2023-10-03 19:04:21 +00:00
|
|
|
gpu_frame_class->create_descriptors = gsk_vulkan_frame_create_descriptors;
|
2023-08-21 00:18:37 +00:00
|
|
|
gpu_frame_class->create_vertex_buffer = gsk_vulkan_frame_create_vertex_buffer;
|
2023-08-30 19:57:41 +00:00
|
|
|
gpu_frame_class->create_storage_buffer = gsk_vulkan_frame_create_storage_buffer;
|
2023-08-12 20:10:16 +00:00
|
|
|
gpu_frame_class->submit = gsk_vulkan_frame_submit;
|
|
|
|
|
|
|
|
object_class->finalize = gsk_vulkan_frame_finalize;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
gsk_vulkan_frame_init (GskVulkanFrame *self)
|
|
|
|
{
|
2023-08-21 00:18:37 +00:00
|
|
|
gsk_descriptor_image_infos_init (&self->descriptor_images);
|
|
|
|
gsk_descriptor_buffer_infos_init (&self->descriptor_buffers);
|
2023-08-12 20:10:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
VkFence
|
|
|
|
gsk_vulkan_frame_get_vk_fence (GskVulkanFrame *self)
|
|
|
|
{
|
|
|
|
return self->vk_fence;
|
|
|
|
}
|