mirror of
https://github.com/SimoneN64/Kaizen.git
synced 2025-04-02 10:41:53 -04:00
687 lines
23 KiB
C++
687 lines
23 KiB
C++
/* Copyright (c) 2017-2023 Hans-Kristian Arntzen
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining
|
|
* a copy of this software and associated documentation files (the
|
|
* "Software"), to deal in the Software without restriction, including
|
|
* without limitation the rights to use, copy, modify, merge, publish,
|
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
|
* permit persons to whom the Software is furnished to do so, subject to
|
|
* the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be
|
|
* included in all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
|
|
* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
|
|
* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
|
|
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
|
|
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#define NOMINMAX
|
|
#include "shader.hpp"
|
|
#include "device.hpp"
|
|
#ifdef GRANITE_VULKAN_SPIRV_CROSS
|
|
#include "spirv_cross.hpp"
|
|
using namespace spirv_cross;
|
|
#endif
|
|
|
|
using namespace Util;
|
|
|
|
namespace Vulkan
|
|
{
|
|
void ImmutableSamplerBank::hash(Util::Hasher &h, const ImmutableSamplerBank *sampler_bank)
|
|
{
|
|
h.u32(0);
|
|
if (sampler_bank)
|
|
{
|
|
unsigned index = 0;
|
|
for (auto &set : sampler_bank->samplers)
|
|
{
|
|
for (auto *binding : set)
|
|
{
|
|
if (binding)
|
|
{
|
|
h.u32(index);
|
|
h.u64(binding->get_hash());
|
|
}
|
|
index++;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
PipelineLayout::PipelineLayout(Hash hash, Device *device_, const CombinedResourceLayout &layout_,
|
|
const ImmutableSamplerBank *immutable_samplers)
|
|
: IntrusiveHashMapEnabled<PipelineLayout>(hash)
|
|
, device(device_)
|
|
, layout(layout_)
|
|
{
|
|
VkDescriptorSetLayout layouts[VULKAN_NUM_DESCRIPTOR_SETS] = {};
|
|
unsigned num_sets = 0;
|
|
for (unsigned i = 0; i < VULKAN_NUM_DESCRIPTOR_SETS; i++)
|
|
{
|
|
set_allocators[i] = device->request_descriptor_set_allocator(layout.sets[i], layout.stages_for_bindings[i],
|
|
immutable_samplers ? immutable_samplers->samplers[i] : nullptr);
|
|
layouts[i] = set_allocators[i]->get_layout_for_pool();
|
|
if (layout.descriptor_set_mask & (1u << i))
|
|
{
|
|
num_sets = i + 1;
|
|
|
|
// Assume the last set index in layout is the highest frequency update one, make that push descriptor if possible.
|
|
// Only one descriptor set can be push descriptor.
|
|
bool has_push_layout = set_allocators[i]->get_layout_for_push() != VK_NULL_HANDLE;
|
|
if (has_push_layout)
|
|
push_set_index = i;
|
|
}
|
|
}
|
|
|
|
if (push_set_index != UINT32_MAX)
|
|
layouts[push_set_index] = set_allocators[push_set_index]->get_layout_for_push();
|
|
|
|
if (num_sets > VULKAN_NUM_DESCRIPTOR_SETS)
|
|
LOGE("Number of sets %u exceeds limit of %u.\n", num_sets, VULKAN_NUM_DESCRIPTOR_SETS);
|
|
|
|
VkPipelineLayoutCreateInfo info = { VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO };
|
|
if (num_sets)
|
|
{
|
|
info.setLayoutCount = num_sets;
|
|
info.pSetLayouts = layouts;
|
|
}
|
|
|
|
if (layout.push_constant_range.stageFlags != 0)
|
|
{
|
|
info.pushConstantRangeCount = 1;
|
|
info.pPushConstantRanges = &layout.push_constant_range;
|
|
}
|
|
|
|
#ifdef VULKAN_DEBUG
|
|
LOGI("Creating pipeline layout.\n");
|
|
#endif
|
|
auto &table = device->get_device_table();
|
|
if (table.vkCreatePipelineLayout(device->get_device(), &info, nullptr, &pipe_layout) != VK_SUCCESS)
|
|
LOGE("Failed to create pipeline layout.\n");
|
|
#ifdef GRANITE_VULKAN_FOSSILIZE
|
|
device->register_pipeline_layout(pipe_layout, get_hash(), info);
|
|
#endif
|
|
|
|
create_update_templates();
|
|
}
|
|
|
|
void PipelineLayout::create_update_templates()
|
|
{
|
|
auto &table = device->get_device_table();
|
|
for (unsigned desc_set = 0; desc_set < VULKAN_NUM_DESCRIPTOR_SETS; desc_set++)
|
|
{
|
|
if ((layout.descriptor_set_mask & (1u << desc_set)) == 0)
|
|
continue;
|
|
if ((layout.bindless_descriptor_set_mask & (1u << desc_set)) != 0)
|
|
continue;
|
|
|
|
VkDescriptorUpdateTemplateEntry update_entries[VULKAN_NUM_BINDINGS];
|
|
uint32_t update_count = 0;
|
|
|
|
auto &set_layout = layout.sets[desc_set];
|
|
|
|
for_each_bit(set_layout.uniform_buffer_mask, [&](uint32_t binding) {
|
|
unsigned array_size = set_layout.array_size[binding];
|
|
VK_ASSERT(update_count < VULKAN_NUM_BINDINGS);
|
|
// Work around a RenderDoc capture bug where descriptorCount > 1 is not handled correctly.
|
|
for (unsigned i = 0; i < array_size; i++)
|
|
{
|
|
auto &entry = update_entries[update_count++];
|
|
entry.descriptorType = desc_set == push_set_index ?
|
|
VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER : VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
|
|
entry.dstBinding = binding;
|
|
entry.dstArrayElement = i;
|
|
entry.descriptorCount = 1;
|
|
if (desc_set == push_set_index)
|
|
entry.offset = offsetof(ResourceBinding, buffer.push) + sizeof(ResourceBinding) * (binding + i);
|
|
else
|
|
entry.offset = offsetof(ResourceBinding, buffer.dynamic) + sizeof(ResourceBinding) * (binding + i);
|
|
entry.stride = sizeof(ResourceBinding);
|
|
}
|
|
});
|
|
|
|
for_each_bit(set_layout.storage_buffer_mask, [&](uint32_t binding) {
|
|
unsigned array_size = set_layout.array_size[binding];
|
|
VK_ASSERT(update_count < VULKAN_NUM_BINDINGS);
|
|
for (unsigned i = 0; i < array_size; i++)
|
|
{
|
|
auto &entry = update_entries[update_count++];
|
|
entry.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
|
|
entry.dstBinding = binding;
|
|
entry.dstArrayElement = i;
|
|
entry.descriptorCount = 1;
|
|
entry.offset = offsetof(ResourceBinding, buffer.dynamic) + sizeof(ResourceBinding) * (binding + i);
|
|
entry.stride = sizeof(ResourceBinding);
|
|
}
|
|
});
|
|
|
|
for_each_bit(set_layout.sampled_texel_buffer_mask, [&](uint32_t binding) {
|
|
unsigned array_size = set_layout.array_size[binding];
|
|
VK_ASSERT(update_count < VULKAN_NUM_BINDINGS);
|
|
for (unsigned i = 0; i < array_size; i++)
|
|
{
|
|
auto &entry = update_entries[update_count++];
|
|
entry.descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER;
|
|
entry.dstBinding = binding;
|
|
entry.dstArrayElement = i;
|
|
entry.descriptorCount = 1;
|
|
entry.offset = offsetof(ResourceBinding, buffer_view) + sizeof(ResourceBinding) * (binding + i);
|
|
entry.stride = sizeof(ResourceBinding);
|
|
}
|
|
});
|
|
|
|
for_each_bit(set_layout.storage_texel_buffer_mask, [&](uint32_t binding) {
|
|
unsigned array_size = set_layout.array_size[binding];
|
|
VK_ASSERT(update_count < VULKAN_NUM_BINDINGS);
|
|
for (unsigned i = 0; i < array_size; i++)
|
|
{
|
|
auto &entry = update_entries[update_count++];
|
|
entry.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_TEXEL_BUFFER;
|
|
entry.dstBinding = binding;
|
|
entry.dstArrayElement = i;
|
|
entry.descriptorCount = 1;
|
|
entry.offset = offsetof(ResourceBinding, buffer_view) + sizeof(ResourceBinding) * (binding + i);
|
|
entry.stride = sizeof(ResourceBinding);
|
|
}
|
|
});
|
|
|
|
for_each_bit(set_layout.sampled_image_mask, [&](uint32_t binding) {
|
|
unsigned array_size = set_layout.array_size[binding];
|
|
VK_ASSERT(update_count < VULKAN_NUM_BINDINGS);
|
|
for (unsigned i = 0; i < array_size; i++)
|
|
{
|
|
auto &entry = update_entries[update_count++];
|
|
entry.descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER;
|
|
entry.dstBinding = binding;
|
|
entry.dstArrayElement = i;
|
|
entry.descriptorCount = 1;
|
|
if (set_layout.fp_mask & (1u << binding))
|
|
entry.offset = offsetof(ResourceBinding, image.fp) + sizeof(ResourceBinding) * (binding + i);
|
|
else
|
|
entry.offset = offsetof(ResourceBinding, image.integer) + sizeof(ResourceBinding) * (binding + i);
|
|
entry.stride = sizeof(ResourceBinding);
|
|
}
|
|
});
|
|
|
|
for_each_bit(set_layout.separate_image_mask, [&](uint32_t binding) {
|
|
unsigned array_size = set_layout.array_size[binding];
|
|
VK_ASSERT(update_count < VULKAN_NUM_BINDINGS);
|
|
for (unsigned i = 0; i < array_size; i++)
|
|
{
|
|
auto &entry = update_entries[update_count++];
|
|
entry.descriptorType = VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE;
|
|
entry.dstBinding = binding;
|
|
entry.dstArrayElement = i;
|
|
entry.descriptorCount = 1;
|
|
if (set_layout.fp_mask & (1u << binding))
|
|
entry.offset = offsetof(ResourceBinding, image.fp) + sizeof(ResourceBinding) * (binding + i);
|
|
else
|
|
entry.offset = offsetof(ResourceBinding, image.integer) + sizeof(ResourceBinding) * (binding + i);
|
|
entry.stride = sizeof(ResourceBinding);
|
|
}
|
|
});
|
|
|
|
for_each_bit(set_layout.sampler_mask & ~set_layout.immutable_sampler_mask, [&](uint32_t binding) {
|
|
unsigned array_size = set_layout.array_size[binding];
|
|
VK_ASSERT(update_count < VULKAN_NUM_BINDINGS);
|
|
for (unsigned i = 0; i < array_size; i++)
|
|
{
|
|
auto &entry = update_entries[update_count++];
|
|
entry.descriptorType = VK_DESCRIPTOR_TYPE_SAMPLER;
|
|
entry.dstBinding = binding;
|
|
entry.dstArrayElement = i;
|
|
entry.descriptorCount = 1;
|
|
entry.offset = offsetof(ResourceBinding, image.fp) + sizeof(ResourceBinding) * (binding + i);
|
|
entry.stride = sizeof(ResourceBinding);
|
|
}
|
|
});
|
|
|
|
for_each_bit(set_layout.storage_image_mask, [&](uint32_t binding) {
|
|
unsigned array_size = set_layout.array_size[binding];
|
|
VK_ASSERT(update_count < VULKAN_NUM_BINDINGS);
|
|
for (unsigned i = 0; i < array_size; i++)
|
|
{
|
|
auto &entry = update_entries[update_count++];
|
|
entry.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
|
|
entry.dstBinding = binding;
|
|
entry.dstArrayElement = i;
|
|
entry.descriptorCount = 1;
|
|
if (set_layout.fp_mask & (1u << binding))
|
|
entry.offset = offsetof(ResourceBinding, image.fp) + sizeof(ResourceBinding) * (binding + i);
|
|
else
|
|
entry.offset = offsetof(ResourceBinding, image.integer) + sizeof(ResourceBinding) * (binding + i);
|
|
entry.stride = sizeof(ResourceBinding);
|
|
}
|
|
});
|
|
|
|
for_each_bit(set_layout.input_attachment_mask, [&](uint32_t binding) {
|
|
unsigned array_size = set_layout.array_size[binding];
|
|
VK_ASSERT(update_count < VULKAN_NUM_BINDINGS);
|
|
for (unsigned i = 0; i < array_size; i++)
|
|
{
|
|
auto &entry = update_entries[update_count++];
|
|
entry.descriptorType = VK_DESCRIPTOR_TYPE_INPUT_ATTACHMENT;
|
|
entry.dstBinding = binding;
|
|
entry.dstArrayElement = i;
|
|
entry.descriptorCount = 1;
|
|
if (set_layout.fp_mask & (1u << binding))
|
|
entry.offset = offsetof(ResourceBinding, image.fp) + sizeof(ResourceBinding) * (binding + i);
|
|
else
|
|
entry.offset = offsetof(ResourceBinding, image.integer) + sizeof(ResourceBinding) * (binding + i);
|
|
entry.stride = sizeof(ResourceBinding);
|
|
}
|
|
});
|
|
|
|
VkDescriptorUpdateTemplateCreateInfo info = { VK_STRUCTURE_TYPE_DESCRIPTOR_UPDATE_TEMPLATE_CREATE_INFO };
|
|
info.pipelineLayout = pipe_layout;
|
|
|
|
if (desc_set == push_set_index)
|
|
{
|
|
info.descriptorSetLayout = set_allocators[desc_set]->get_layout_for_push();
|
|
info.templateType = VK_DESCRIPTOR_UPDATE_TEMPLATE_TYPE_PUSH_DESCRIPTORS_KHR;
|
|
}
|
|
else
|
|
{
|
|
info.descriptorSetLayout = set_allocators[desc_set]->get_layout_for_pool();
|
|
info.templateType = VK_DESCRIPTOR_UPDATE_TEMPLATE_TYPE_DESCRIPTOR_SET;
|
|
}
|
|
|
|
info.set = desc_set;
|
|
info.descriptorUpdateEntryCount = update_count;
|
|
info.pDescriptorUpdateEntries = update_entries;
|
|
info.pipelineBindPoint = (layout.stages_for_sets[desc_set] & VK_SHADER_STAGE_COMPUTE_BIT) ?
|
|
VK_PIPELINE_BIND_POINT_COMPUTE : VK_PIPELINE_BIND_POINT_GRAPHICS;
|
|
|
|
if (table.vkCreateDescriptorUpdateTemplate(device->get_device(), &info, nullptr,
|
|
&update_template[desc_set]) != VK_SUCCESS)
|
|
{
|
|
LOGE("Failed to create descriptor update template.\n");
|
|
}
|
|
}
|
|
}
|
|
|
|
PipelineLayout::~PipelineLayout()
|
|
{
|
|
auto &table = device->get_device_table();
|
|
if (pipe_layout != VK_NULL_HANDLE)
|
|
table.vkDestroyPipelineLayout(device->get_device(), pipe_layout, nullptr);
|
|
|
|
for (auto &update : update_template)
|
|
if (update != VK_NULL_HANDLE)
|
|
table.vkDestroyDescriptorUpdateTemplate(device->get_device(), update, nullptr);
|
|
}
|
|
|
|
const char *Shader::stage_to_name(ShaderStage stage)
|
|
{
|
|
switch (stage)
|
|
{
|
|
case ShaderStage::Compute:
|
|
return "compute";
|
|
case ShaderStage::Vertex:
|
|
return "vertex";
|
|
case ShaderStage::Fragment:
|
|
return "fragment";
|
|
case ShaderStage::Geometry:
|
|
return "geometry";
|
|
case ShaderStage::TessControl:
|
|
return "tess_control";
|
|
case ShaderStage::TessEvaluation:
|
|
return "tess_evaluation";
|
|
default:
|
|
return "unknown";
|
|
}
|
|
}
|
|
|
|
// Implicitly also checks for endian issues.
|
|
static const uint16_t reflection_magic[] = { 'G', 'R', 'A', ResourceLayout::Version };
|
|
|
|
size_t ResourceLayout::serialization_size()
|
|
{
|
|
return sizeof(ResourceLayout) + sizeof(reflection_magic);
|
|
}
|
|
|
|
bool ResourceLayout::serialize(uint8_t *data, size_t size) const
|
|
{
|
|
if (size != serialization_size())
|
|
return false;
|
|
|
|
// Cannot serialize externally defined immutable samplers.
|
|
for (auto &set : sets)
|
|
if (set.immutable_sampler_mask != 0)
|
|
return false;
|
|
|
|
memcpy(data, reflection_magic, sizeof(reflection_magic));
|
|
memcpy(data + sizeof(reflection_magic), this, sizeof(*this));
|
|
return true;
|
|
}
|
|
|
|
bool ResourceLayout::unserialize(const uint8_t *data, size_t size)
|
|
{
|
|
if (size != sizeof(*this) + sizeof(reflection_magic))
|
|
{
|
|
LOGE("Reflection size mismatch.\n");
|
|
return false;
|
|
}
|
|
|
|
if (memcmp(data, reflection_magic, sizeof(reflection_magic)) != 0)
|
|
{
|
|
LOGE("Magic mismatch.\n");
|
|
return false;
|
|
}
|
|
|
|
memcpy(this, data + sizeof(reflection_magic), sizeof(*this));
|
|
return true;
|
|
}
|
|
|
|
Util::Hash Shader::hash(const uint32_t *data, size_t size)
|
|
{
|
|
Util::Hasher hasher;
|
|
hasher.data(data, size);
|
|
return hasher.get();
|
|
}
|
|
|
|
#ifdef GRANITE_VULKAN_SPIRV_CROSS
|
|
static void update_array_info(ResourceLayout &layout, const SPIRType &type, unsigned set, unsigned binding)
|
|
{
|
|
auto &size = layout.sets[set].array_size[binding];
|
|
if (!type.array.empty())
|
|
{
|
|
if (type.array.size() != 1)
|
|
LOGE("Array dimension must be 1.\n");
|
|
else if (!type.array_size_literal.front())
|
|
LOGE("Array dimension must be a literal.\n");
|
|
else
|
|
{
|
|
if (type.array.front() == 0)
|
|
{
|
|
if (binding != 0)
|
|
LOGE("Bindless textures can only be used with binding = 0 in a set.\n");
|
|
|
|
if (type.basetype != SPIRType::Image || type.image.dim == spv::DimBuffer)
|
|
{
|
|
LOGE("Can only use bindless for sampled images.\n");
|
|
}
|
|
else
|
|
{
|
|
layout.bindless_set_mask |= 1u << set;
|
|
// Ignore fp_mask for bindless since we can mix and match.
|
|
layout.sets[set].fp_mask = 0;
|
|
}
|
|
|
|
size = DescriptorSetLayout::UNSIZED_ARRAY;
|
|
}
|
|
else if (size && size != type.array.front())
|
|
LOGE("Array dimension for (%u, %u) is inconsistent.\n", set, binding);
|
|
else if (type.array.front() + binding > VULKAN_NUM_BINDINGS)
|
|
LOGE("Binding array will go out of bounds.\n");
|
|
else
|
|
size = uint8_t(type.array.front());
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (size && size != 1)
|
|
LOGE("Array dimension for (%u, %u) is inconsistent.\n", set, binding);
|
|
size = 1;
|
|
}
|
|
}
|
|
|
|
bool Shader::reflect_resource_layout(ResourceLayout &layout, const uint32_t *data, size_t size)
|
|
{
|
|
Compiler compiler(data, size / sizeof(uint32_t));
|
|
|
|
#ifdef VULKAN_DEBUG
|
|
LOGI("Reflecting shader layout.\n");
|
|
#endif
|
|
|
|
auto resources = compiler.get_shader_resources();
|
|
for (auto &image : resources.sampled_images)
|
|
{
|
|
auto set = compiler.get_decoration(image.id, spv::DecorationDescriptorSet);
|
|
auto binding = compiler.get_decoration(image.id, spv::DecorationBinding);
|
|
VK_ASSERT(set < VULKAN_NUM_DESCRIPTOR_SETS);
|
|
VK_ASSERT(binding < VULKAN_NUM_BINDINGS);
|
|
|
|
auto &type = compiler.get_type(image.type_id);
|
|
if (type.image.dim == spv::DimBuffer)
|
|
layout.sets[set].sampled_texel_buffer_mask |= 1u << binding;
|
|
else
|
|
layout.sets[set].sampled_image_mask |= 1u << binding;
|
|
|
|
if (compiler.get_type(type.image.type).basetype == SPIRType::BaseType::Float)
|
|
layout.sets[set].fp_mask |= 1u << binding;
|
|
|
|
update_array_info(layout, type, set, binding);
|
|
}
|
|
|
|
for (auto &image : resources.subpass_inputs)
|
|
{
|
|
auto set = compiler.get_decoration(image.id, spv::DecorationDescriptorSet);
|
|
auto binding = compiler.get_decoration(image.id, spv::DecorationBinding);
|
|
VK_ASSERT(set < VULKAN_NUM_DESCRIPTOR_SETS);
|
|
VK_ASSERT(binding < VULKAN_NUM_BINDINGS);
|
|
|
|
layout.sets[set].input_attachment_mask |= 1u << binding;
|
|
|
|
auto &type = compiler.get_type(image.type_id);
|
|
if (compiler.get_type(type.image.type).basetype == SPIRType::BaseType::Float)
|
|
layout.sets[set].fp_mask |= 1u << binding;
|
|
update_array_info(layout, type, set, binding);
|
|
}
|
|
|
|
for (auto &image : resources.separate_images)
|
|
{
|
|
auto set = compiler.get_decoration(image.id, spv::DecorationDescriptorSet);
|
|
auto binding = compiler.get_decoration(image.id, spv::DecorationBinding);
|
|
VK_ASSERT(set < VULKAN_NUM_DESCRIPTOR_SETS);
|
|
VK_ASSERT(binding < VULKAN_NUM_BINDINGS);
|
|
|
|
auto &type = compiler.get_type(image.type_id);
|
|
if (compiler.get_type(type.image.type).basetype == SPIRType::BaseType::Float)
|
|
layout.sets[set].fp_mask |= 1u << binding;
|
|
|
|
if (type.image.dim == spv::DimBuffer)
|
|
layout.sets[set].sampled_texel_buffer_mask |= 1u << binding;
|
|
else
|
|
layout.sets[set].separate_image_mask |= 1u << binding;
|
|
|
|
update_array_info(layout, type, set, binding);
|
|
}
|
|
|
|
for (auto &image : resources.separate_samplers)
|
|
{
|
|
auto set = compiler.get_decoration(image.id, spv::DecorationDescriptorSet);
|
|
auto binding = compiler.get_decoration(image.id, spv::DecorationBinding);
|
|
VK_ASSERT(set < VULKAN_NUM_DESCRIPTOR_SETS);
|
|
VK_ASSERT(binding < VULKAN_NUM_BINDINGS);
|
|
|
|
layout.sets[set].sampler_mask |= 1u << binding;
|
|
update_array_info(layout, compiler.get_type(image.type_id), set, binding);
|
|
}
|
|
|
|
for (auto &image : resources.storage_images)
|
|
{
|
|
auto set = compiler.get_decoration(image.id, spv::DecorationDescriptorSet);
|
|
auto binding = compiler.get_decoration(image.id, spv::DecorationBinding);
|
|
VK_ASSERT(set < VULKAN_NUM_DESCRIPTOR_SETS);
|
|
VK_ASSERT(binding < VULKAN_NUM_BINDINGS);
|
|
|
|
auto &type = compiler.get_type(image.type_id);
|
|
if (type.image.dim == spv::DimBuffer)
|
|
layout.sets[set].storage_texel_buffer_mask |= 1u << binding;
|
|
else
|
|
layout.sets[set].storage_image_mask |= 1u << binding;
|
|
|
|
if (compiler.get_type(type.image.type).basetype == SPIRType::BaseType::Float)
|
|
layout.sets[set].fp_mask |= 1u << binding;
|
|
|
|
update_array_info(layout, type, set, binding);
|
|
}
|
|
|
|
for (auto &buffer : resources.uniform_buffers)
|
|
{
|
|
auto set = compiler.get_decoration(buffer.id, spv::DecorationDescriptorSet);
|
|
auto binding = compiler.get_decoration(buffer.id, spv::DecorationBinding);
|
|
VK_ASSERT(set < VULKAN_NUM_DESCRIPTOR_SETS);
|
|
VK_ASSERT(binding < VULKAN_NUM_BINDINGS);
|
|
|
|
layout.sets[set].uniform_buffer_mask |= 1u << binding;
|
|
update_array_info(layout, compiler.get_type(buffer.type_id), set, binding);
|
|
}
|
|
|
|
for (auto &buffer : resources.storage_buffers)
|
|
{
|
|
auto set = compiler.get_decoration(buffer.id, spv::DecorationDescriptorSet);
|
|
auto binding = compiler.get_decoration(buffer.id, spv::DecorationBinding);
|
|
VK_ASSERT(set < VULKAN_NUM_DESCRIPTOR_SETS);
|
|
VK_ASSERT(binding < VULKAN_NUM_BINDINGS);
|
|
|
|
layout.sets[set].storage_buffer_mask |= 1u << binding;
|
|
update_array_info(layout, compiler.get_type(buffer.type_id), set, binding);
|
|
}
|
|
|
|
for (auto &attrib : resources.stage_inputs)
|
|
{
|
|
auto location = compiler.get_decoration(attrib.id, spv::DecorationLocation);
|
|
layout.input_mask |= 1u << location;
|
|
}
|
|
|
|
for (auto &attrib : resources.stage_outputs)
|
|
{
|
|
auto location = compiler.get_decoration(attrib.id, spv::DecorationLocation);
|
|
layout.output_mask |= 1u << location;
|
|
}
|
|
|
|
if (!resources.push_constant_buffers.empty())
|
|
{
|
|
// Don't bother trying to extract which part of a push constant block we're using.
|
|
// Just assume we're accessing everything. At least on older validation layers,
|
|
// it did not do a static analysis to determine similar information, so we got a lot
|
|
// of false positives.
|
|
layout.push_constant_size =
|
|
compiler.get_declared_struct_size(compiler.get_type(resources.push_constant_buffers.front().base_type_id));
|
|
}
|
|
|
|
auto spec_constants = compiler.get_specialization_constants();
|
|
for (auto &c : spec_constants)
|
|
{
|
|
if (c.constant_id >= VULKAN_NUM_TOTAL_SPEC_CONSTANTS)
|
|
{
|
|
LOGE("Spec constant ID: %u is out of range, will be ignored.\n", c.constant_id);
|
|
continue;
|
|
}
|
|
|
|
layout.spec_constant_mask |= 1u << c.constant_id;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
#else
|
|
bool Shader::reflect_resource_layout(ResourceLayout &, const uint32_t *, size_t)
|
|
{
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
Shader::Shader(Hash hash, Device *device_, const uint32_t *data, size_t size,
|
|
const ResourceLayout *resource_layout)
|
|
: IntrusiveHashMapEnabled<Shader>(hash)
|
|
, device(device_)
|
|
{
|
|
VkShaderModuleCreateInfo info = { VK_STRUCTURE_TYPE_SHADER_MODULE_CREATE_INFO };
|
|
info.codeSize = size;
|
|
info.pCode = data;
|
|
|
|
#ifdef VULKAN_DEBUG
|
|
LOGI("Creating shader module.\n");
|
|
#endif
|
|
auto &table = device->get_device_table();
|
|
if (table.vkCreateShaderModule(device->get_device(), &info, nullptr, &module) != VK_SUCCESS)
|
|
LOGE("Failed to create shader module.\n");
|
|
|
|
#ifdef GRANITE_VULKAN_FOSSILIZE
|
|
device->register_shader_module(module, get_hash(), info);
|
|
#endif
|
|
|
|
if (resource_layout)
|
|
layout = *resource_layout;
|
|
#ifdef GRANITE_VULKAN_SPIRV_CROSS
|
|
else if (!reflect_resource_layout(layout, data, size))
|
|
LOGE("Failed to reflect resource layout.\n");
|
|
#endif
|
|
|
|
if (layout.bindless_set_mask != 0 && !device->get_device_features().vk12_features.descriptorIndexing)
|
|
LOGE("Sufficient features for descriptor indexing is not supported on this device.\n");
|
|
}
|
|
|
|
Shader::~Shader()
|
|
{
|
|
auto &table = device->get_device_table();
|
|
if (module)
|
|
table.vkDestroyShaderModule(device->get_device(), module, nullptr);
|
|
}
|
|
|
|
void Program::set_shader(ShaderStage stage, Shader *handle)
|
|
{
|
|
shaders[Util::ecast(stage)] = handle;
|
|
}
|
|
|
|
Program::Program(Device *device_, Shader *vertex, Shader *fragment, const ImmutableSamplerBank *sampler_bank)
|
|
: device(device_)
|
|
{
|
|
set_shader(ShaderStage::Vertex, vertex);
|
|
set_shader(ShaderStage::Fragment, fragment);
|
|
device->bake_program(*this, sampler_bank);
|
|
}
|
|
|
|
Program::Program(Device *device_, Shader *task, Shader *mesh, Shader *fragment, const ImmutableSamplerBank *sampler_bank)
|
|
: device(device_)
|
|
{
|
|
if (task)
|
|
set_shader(ShaderStage::Task, task);
|
|
set_shader(ShaderStage::Mesh, mesh);
|
|
set_shader(ShaderStage::Fragment, fragment);
|
|
device->bake_program(*this, sampler_bank);
|
|
}
|
|
|
|
Program::Program(Device *device_, Shader *compute_shader, const ImmutableSamplerBank *sampler_bank)
|
|
: device(device_)
|
|
{
|
|
set_shader(ShaderStage::Compute, compute_shader);
|
|
device->bake_program(*this, sampler_bank);
|
|
}
|
|
|
|
Pipeline Program::get_pipeline(Hash hash) const
|
|
{
|
|
auto *ret = pipelines.find(hash);
|
|
return ret ? ret->get() : Pipeline{};
|
|
}
|
|
|
|
Pipeline Program::add_pipeline(Hash hash, const Pipeline &pipeline)
|
|
{
|
|
return pipelines.emplace_yield(hash, pipeline)->get();
|
|
}
|
|
|
|
void Program::destroy_pipeline(const Pipeline &pipeline)
|
|
{
|
|
device->get_device_table().vkDestroyPipeline(device->get_device(), pipeline.pipeline, nullptr);
|
|
}
|
|
|
|
void Program::promote_read_write_to_read_only()
|
|
{
|
|
pipelines.move_to_read_only();
|
|
}
|
|
|
|
Program::~Program()
|
|
{
|
|
for (auto &pipe : pipelines.get_read_only())
|
|
destroy_pipeline(pipe.get());
|
|
for (auto &pipe : pipelines.get_read_write())
|
|
destroy_pipeline(pipe.get());
|
|
}
|
|
}
|