mirror of
https://github.com/skyline-emu/skyline.git
synced 2024-12-25 16:41:50 +01:00
1969 lines
106 KiB
C++
1969 lines
106 KiB
C++
|
// Copyright (c) 2015-2019 The Khronos Group Inc.
|
||
|
//
|
||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||
|
// you may not use this file except in compliance with the License.
|
||
|
// You may obtain a copy of the License at
|
||
|
//
|
||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||
|
//
|
||
|
// Unless required by applicable law or agreed to in writing, software
|
||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
|
// See the License for the specific language governing permissions and
|
||
|
// limitations under the License.
|
||
|
//
|
||
|
// ---- Exceptions to the Apache 2.0 License: ----
|
||
|
//
|
||
|
// As an exception, if you use this Software to generate code and portions of
|
||
|
// this Software are embedded into the generated code as a result, you may
|
||
|
// redistribute such product without providing attribution as would otherwise
|
||
|
// be required by Sections 4(a), 4(b) and 4(d) of the License.
|
||
|
//
|
||
|
// In addition, if you combine or link code generated by this Software with
|
||
|
// software that is licensed under the GPLv2 or the LGPL v2.0 or 2.1
|
||
|
// ("`Combined Software`") and if a court of competent jurisdiction determines
|
||
|
// that the patent provision (Section 3), the indemnity provision (Section 9)
|
||
|
// or other Section of the License conflicts with the conditions of the
|
||
|
// applicable GPL or LGPL license, you may retroactively and prospectively
|
||
|
// choose to deem waived or otherwise exclude such Section(s) of the License,
|
||
|
// but only in their entirety and only with respect to the Combined Software.
|
||
|
//
|
||
|
|
||
|
// This header is generated from the Khronos Vulkan XML API Registry.
|
||
|
|
||
|
#pragma once
|
||
|
|
||
|
#include "base.hpp"
|
||
|
|
||
|
namespace VULKAN_HPP_NAMESPACE
|
||
|
{
|
||
|
class DispatchLoaderDynamic;
|
||
|
#if !defined(VULKAN_HPP_DISPATCH_LOADER_DYNAMIC)
|
||
|
# if defined(VK_NO_PROTOTYPES)
|
||
|
# define VULKAN_HPP_DISPATCH_LOADER_DYNAMIC 1
|
||
|
# else
|
||
|
# define VULKAN_HPP_DISPATCH_LOADER_DYNAMIC 0
|
||
|
# endif
|
||
|
#endif
|
||
|
|
||
|
#if !defined(VULKAN_HPP_DEFAULT_DISPATCHER)
|
||
|
# if VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1
|
||
|
# define VULKAN_HPP_DEFAULT_DISPATCHER ::VULKAN_HPP_NAMESPACE::defaultDispatchLoaderDynamic
|
||
|
# define VULKAN_HPP_DEFAULT_DISPATCH_LOADER_DYNAMIC_STORAGE namespace VULKAN_HPP_NAMESPACE { DispatchLoaderDynamic defaultDispatchLoaderDynamic; }
|
||
|
extern DispatchLoaderDynamic defaultDispatchLoaderDynamic;
|
||
|
# else
|
||
|
# define VULKAN_HPP_DEFAULT_DISPATCHER ::VULKAN_HPP_NAMESPACE::DispatchLoaderStatic()
|
||
|
# define VULKAN_HPP_DEFAULT_DISPATCH_LOADER_DYNAMIC_STORAGE
|
||
|
# endif
|
||
|
#endif
|
||
|
|
||
|
#if !defined(VULKAN_HPP_DEFAULT_DISPATCHER_TYPE)
|
||
|
# if VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1
|
||
|
#define VULKAN_HPP_DEFAULT_DISPATCHER_TYPE ::VULKAN_HPP_NAMESPACE::DispatchLoaderDynamic
|
||
|
# else
|
||
|
# define VULKAN_HPP_DEFAULT_DISPATCHER_TYPE ::VULKAN_HPP_NAMESPACE::DispatchLoaderStatic
|
||
|
# endif
|
||
|
#endif
|
||
|
|
||
|
#if !defined(VK_NO_PROTOTYPES)
|
||
|
class DispatchLoaderStatic
|
||
|
{
|
||
|
public:
|
||
|
VkResult vkCreateInstance( const VkInstanceCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkInstance* pInstance ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateInstance( pCreateInfo, pAllocator, pInstance );
|
||
|
}
|
||
|
|
||
|
VkResult vkEnumerateInstanceExtensionProperties( const char* pLayerName, uint32_t* pPropertyCount, VkExtensionProperties* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkEnumerateInstanceExtensionProperties( pLayerName, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkEnumerateInstanceLayerProperties( uint32_t* pPropertyCount, VkLayerProperties* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkEnumerateInstanceLayerProperties( pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkEnumerateInstanceVersion( uint32_t* pApiVersion ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkEnumerateInstanceVersion( pApiVersion );
|
||
|
}
|
||
|
|
||
|
VkResult vkBeginCommandBuffer( VkCommandBuffer commandBuffer, const VkCommandBufferBeginInfo* pBeginInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkBeginCommandBuffer( commandBuffer, pBeginInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdBeginConditionalRenderingEXT( VkCommandBuffer commandBuffer, const VkConditionalRenderingBeginInfoEXT* pConditionalRenderingBegin ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBeginConditionalRenderingEXT( commandBuffer, pConditionalRenderingBegin );
|
||
|
}
|
||
|
|
||
|
void vkCmdBeginDebugUtilsLabelEXT( VkCommandBuffer commandBuffer, const VkDebugUtilsLabelEXT* pLabelInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBeginDebugUtilsLabelEXT( commandBuffer, pLabelInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdBeginQuery( VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t query, VkQueryControlFlags flags ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBeginQuery( commandBuffer, queryPool, query, flags );
|
||
|
}
|
||
|
|
||
|
void vkCmdBeginQueryIndexedEXT( VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t query, VkQueryControlFlags flags, uint32_t index ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBeginQueryIndexedEXT( commandBuffer, queryPool, query, flags, index );
|
||
|
}
|
||
|
|
||
|
void vkCmdBeginRenderPass( VkCommandBuffer commandBuffer, const VkRenderPassBeginInfo* pRenderPassBegin, VkSubpassContents contents ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBeginRenderPass( commandBuffer, pRenderPassBegin, contents );
|
||
|
}
|
||
|
|
||
|
void vkCmdBeginRenderPass2KHR( VkCommandBuffer commandBuffer, const VkRenderPassBeginInfo* pRenderPassBegin, const VkSubpassBeginInfoKHR* pSubpassBeginInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBeginRenderPass2KHR( commandBuffer, pRenderPassBegin, pSubpassBeginInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdBeginTransformFeedbackEXT( VkCommandBuffer commandBuffer, uint32_t firstCounterBuffer, uint32_t counterBufferCount, const VkBuffer* pCounterBuffers, const VkDeviceSize* pCounterBufferOffsets ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBeginTransformFeedbackEXT( commandBuffer, firstCounterBuffer, counterBufferCount, pCounterBuffers, pCounterBufferOffsets );
|
||
|
}
|
||
|
|
||
|
void vkCmdBindDescriptorSets( VkCommandBuffer commandBuffer, VkPipelineBindPoint pipelineBindPoint, VkPipelineLayout layout, uint32_t firstSet, uint32_t descriptorSetCount, const VkDescriptorSet* pDescriptorSets, uint32_t dynamicOffsetCount, const uint32_t* pDynamicOffsets ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBindDescriptorSets( commandBuffer, pipelineBindPoint, layout, firstSet, descriptorSetCount, pDescriptorSets, dynamicOffsetCount, pDynamicOffsets );
|
||
|
}
|
||
|
|
||
|
void vkCmdBindIndexBuffer( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, VkIndexType indexType ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBindIndexBuffer( commandBuffer, buffer, offset, indexType );
|
||
|
}
|
||
|
|
||
|
void vkCmdBindPipeline( VkCommandBuffer commandBuffer, VkPipelineBindPoint pipelineBindPoint, VkPipeline pipeline ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBindPipeline( commandBuffer, pipelineBindPoint, pipeline );
|
||
|
}
|
||
|
|
||
|
void vkCmdBindShadingRateImageNV( VkCommandBuffer commandBuffer, VkImageView imageView, VkImageLayout imageLayout ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBindShadingRateImageNV( commandBuffer, imageView, imageLayout );
|
||
|
}
|
||
|
|
||
|
void vkCmdBindTransformFeedbackBuffersEXT( VkCommandBuffer commandBuffer, uint32_t firstBinding, uint32_t bindingCount, const VkBuffer* pBuffers, const VkDeviceSize* pOffsets, const VkDeviceSize* pSizes ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBindTransformFeedbackBuffersEXT( commandBuffer, firstBinding, bindingCount, pBuffers, pOffsets, pSizes );
|
||
|
}
|
||
|
|
||
|
void vkCmdBindVertexBuffers( VkCommandBuffer commandBuffer, uint32_t firstBinding, uint32_t bindingCount, const VkBuffer* pBuffers, const VkDeviceSize* pOffsets ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBindVertexBuffers( commandBuffer, firstBinding, bindingCount, pBuffers, pOffsets );
|
||
|
}
|
||
|
|
||
|
void vkCmdBlitImage( VkCommandBuffer commandBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage dstImage, VkImageLayout dstImageLayout, uint32_t regionCount, const VkImageBlit* pRegions, VkFilter filter ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBlitImage( commandBuffer, srcImage, srcImageLayout, dstImage, dstImageLayout, regionCount, pRegions, filter );
|
||
|
}
|
||
|
|
||
|
void vkCmdBuildAccelerationStructureNV( VkCommandBuffer commandBuffer, const VkAccelerationStructureInfoNV* pInfo, VkBuffer instanceData, VkDeviceSize instanceOffset, VkBool32 update, VkAccelerationStructureNV dst, VkAccelerationStructureNV src, VkBuffer scratch, VkDeviceSize scratchOffset ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdBuildAccelerationStructureNV( commandBuffer, pInfo, instanceData, instanceOffset, update, dst, src, scratch, scratchOffset );
|
||
|
}
|
||
|
|
||
|
void vkCmdClearAttachments( VkCommandBuffer commandBuffer, uint32_t attachmentCount, const VkClearAttachment* pAttachments, uint32_t rectCount, const VkClearRect* pRects ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdClearAttachments( commandBuffer, attachmentCount, pAttachments, rectCount, pRects );
|
||
|
}
|
||
|
|
||
|
void vkCmdClearColorImage( VkCommandBuffer commandBuffer, VkImage image, VkImageLayout imageLayout, const VkClearColorValue* pColor, uint32_t rangeCount, const VkImageSubresourceRange* pRanges ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdClearColorImage( commandBuffer, image, imageLayout, pColor, rangeCount, pRanges );
|
||
|
}
|
||
|
|
||
|
void vkCmdClearDepthStencilImage( VkCommandBuffer commandBuffer, VkImage image, VkImageLayout imageLayout, const VkClearDepthStencilValue* pDepthStencil, uint32_t rangeCount, const VkImageSubresourceRange* pRanges ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdClearDepthStencilImage( commandBuffer, image, imageLayout, pDepthStencil, rangeCount, pRanges );
|
||
|
}
|
||
|
|
||
|
void vkCmdCopyAccelerationStructureNV( VkCommandBuffer commandBuffer, VkAccelerationStructureNV dst, VkAccelerationStructureNV src, VkCopyAccelerationStructureModeNV mode ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdCopyAccelerationStructureNV( commandBuffer, dst, src, mode );
|
||
|
}
|
||
|
|
||
|
void vkCmdCopyBuffer( VkCommandBuffer commandBuffer, VkBuffer srcBuffer, VkBuffer dstBuffer, uint32_t regionCount, const VkBufferCopy* pRegions ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdCopyBuffer( commandBuffer, srcBuffer, dstBuffer, regionCount, pRegions );
|
||
|
}
|
||
|
|
||
|
void vkCmdCopyBufferToImage( VkCommandBuffer commandBuffer, VkBuffer srcBuffer, VkImage dstImage, VkImageLayout dstImageLayout, uint32_t regionCount, const VkBufferImageCopy* pRegions ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdCopyBufferToImage( commandBuffer, srcBuffer, dstImage, dstImageLayout, regionCount, pRegions );
|
||
|
}
|
||
|
|
||
|
void vkCmdCopyImage( VkCommandBuffer commandBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage dstImage, VkImageLayout dstImageLayout, uint32_t regionCount, const VkImageCopy* pRegions ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdCopyImage( commandBuffer, srcImage, srcImageLayout, dstImage, dstImageLayout, regionCount, pRegions );
|
||
|
}
|
||
|
|
||
|
void vkCmdCopyImageToBuffer( VkCommandBuffer commandBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkBuffer dstBuffer, uint32_t regionCount, const VkBufferImageCopy* pRegions ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdCopyImageToBuffer( commandBuffer, srcImage, srcImageLayout, dstBuffer, regionCount, pRegions );
|
||
|
}
|
||
|
|
||
|
void vkCmdCopyQueryPoolResults( VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t firstQuery, uint32_t queryCount, VkBuffer dstBuffer, VkDeviceSize dstOffset, VkDeviceSize stride, VkQueryResultFlags flags ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdCopyQueryPoolResults( commandBuffer, queryPool, firstQuery, queryCount, dstBuffer, dstOffset, stride, flags );
|
||
|
}
|
||
|
|
||
|
void vkCmdDebugMarkerBeginEXT( VkCommandBuffer commandBuffer, const VkDebugMarkerMarkerInfoEXT* pMarkerInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDebugMarkerBeginEXT( commandBuffer, pMarkerInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdDebugMarkerEndEXT( VkCommandBuffer commandBuffer ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDebugMarkerEndEXT( commandBuffer );
|
||
|
}
|
||
|
|
||
|
void vkCmdDebugMarkerInsertEXT( VkCommandBuffer commandBuffer, const VkDebugMarkerMarkerInfoEXT* pMarkerInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDebugMarkerInsertEXT( commandBuffer, pMarkerInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdDispatch( VkCommandBuffer commandBuffer, uint32_t groupCountX, uint32_t groupCountY, uint32_t groupCountZ ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDispatch( commandBuffer, groupCountX, groupCountY, groupCountZ );
|
||
|
}
|
||
|
|
||
|
void vkCmdDispatchBase( VkCommandBuffer commandBuffer, uint32_t baseGroupX, uint32_t baseGroupY, uint32_t baseGroupZ, uint32_t groupCountX, uint32_t groupCountY, uint32_t groupCountZ ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDispatchBase( commandBuffer, baseGroupX, baseGroupY, baseGroupZ, groupCountX, groupCountY, groupCountZ );
|
||
|
}
|
||
|
|
||
|
void vkCmdDispatchBaseKHR( VkCommandBuffer commandBuffer, uint32_t baseGroupX, uint32_t baseGroupY, uint32_t baseGroupZ, uint32_t groupCountX, uint32_t groupCountY, uint32_t groupCountZ ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDispatchBaseKHR( commandBuffer, baseGroupX, baseGroupY, baseGroupZ, groupCountX, groupCountY, groupCountZ );
|
||
|
}
|
||
|
|
||
|
void vkCmdDispatchIndirect( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDispatchIndirect( commandBuffer, buffer, offset );
|
||
|
}
|
||
|
|
||
|
void vkCmdDraw( VkCommandBuffer commandBuffer, uint32_t vertexCount, uint32_t instanceCount, uint32_t firstVertex, uint32_t firstInstance ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDraw( commandBuffer, vertexCount, instanceCount, firstVertex, firstInstance );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawIndexed( VkCommandBuffer commandBuffer, uint32_t indexCount, uint32_t instanceCount, uint32_t firstIndex, int32_t vertexOffset, uint32_t firstInstance ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawIndexed( commandBuffer, indexCount, instanceCount, firstIndex, vertexOffset, firstInstance );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawIndexedIndirect( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t drawCount, uint32_t stride ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawIndexedIndirect( commandBuffer, buffer, offset, drawCount, stride );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawIndexedIndirectCountKHR( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, VkBuffer countBuffer, VkDeviceSize countBufferOffset, uint32_t maxDrawCount, uint32_t stride ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawIndexedIndirectCountKHR( commandBuffer, buffer, offset, countBuffer, countBufferOffset, maxDrawCount, stride );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawIndexedIndirectCountAMD( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, VkBuffer countBuffer, VkDeviceSize countBufferOffset, uint32_t maxDrawCount, uint32_t stride ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawIndexedIndirectCountAMD( commandBuffer, buffer, offset, countBuffer, countBufferOffset, maxDrawCount, stride );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawIndirect( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t drawCount, uint32_t stride ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawIndirect( commandBuffer, buffer, offset, drawCount, stride );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawIndirectByteCountEXT( VkCommandBuffer commandBuffer, uint32_t instanceCount, uint32_t firstInstance, VkBuffer counterBuffer, VkDeviceSize counterBufferOffset, uint32_t counterOffset, uint32_t vertexStride ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawIndirectByteCountEXT( commandBuffer, instanceCount, firstInstance, counterBuffer, counterBufferOffset, counterOffset, vertexStride );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawIndirectCountKHR( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, VkBuffer countBuffer, VkDeviceSize countBufferOffset, uint32_t maxDrawCount, uint32_t stride ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawIndirectCountKHR( commandBuffer, buffer, offset, countBuffer, countBufferOffset, maxDrawCount, stride );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawIndirectCountAMD( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, VkBuffer countBuffer, VkDeviceSize countBufferOffset, uint32_t maxDrawCount, uint32_t stride ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawIndirectCountAMD( commandBuffer, buffer, offset, countBuffer, countBufferOffset, maxDrawCount, stride );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawMeshTasksIndirectCountNV( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, VkBuffer countBuffer, VkDeviceSize countBufferOffset, uint32_t maxDrawCount, uint32_t stride ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawMeshTasksIndirectCountNV( commandBuffer, buffer, offset, countBuffer, countBufferOffset, maxDrawCount, stride );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawMeshTasksIndirectNV( VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t drawCount, uint32_t stride ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawMeshTasksIndirectNV( commandBuffer, buffer, offset, drawCount, stride );
|
||
|
}
|
||
|
|
||
|
void vkCmdDrawMeshTasksNV( VkCommandBuffer commandBuffer, uint32_t taskCount, uint32_t firstTask ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdDrawMeshTasksNV( commandBuffer, taskCount, firstTask );
|
||
|
}
|
||
|
|
||
|
void vkCmdEndConditionalRenderingEXT( VkCommandBuffer commandBuffer ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdEndConditionalRenderingEXT( commandBuffer );
|
||
|
}
|
||
|
|
||
|
void vkCmdEndDebugUtilsLabelEXT( VkCommandBuffer commandBuffer ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdEndDebugUtilsLabelEXT( commandBuffer );
|
||
|
}
|
||
|
|
||
|
void vkCmdEndQuery( VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t query ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdEndQuery( commandBuffer, queryPool, query );
|
||
|
}
|
||
|
|
||
|
void vkCmdEndQueryIndexedEXT( VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t query, uint32_t index ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdEndQueryIndexedEXT( commandBuffer, queryPool, query, index );
|
||
|
}
|
||
|
|
||
|
void vkCmdEndRenderPass( VkCommandBuffer commandBuffer ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdEndRenderPass( commandBuffer );
|
||
|
}
|
||
|
|
||
|
void vkCmdEndRenderPass2KHR( VkCommandBuffer commandBuffer, const VkSubpassEndInfoKHR* pSubpassEndInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdEndRenderPass2KHR( commandBuffer, pSubpassEndInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdEndTransformFeedbackEXT( VkCommandBuffer commandBuffer, uint32_t firstCounterBuffer, uint32_t counterBufferCount, const VkBuffer* pCounterBuffers, const VkDeviceSize* pCounterBufferOffsets ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdEndTransformFeedbackEXT( commandBuffer, firstCounterBuffer, counterBufferCount, pCounterBuffers, pCounterBufferOffsets );
|
||
|
}
|
||
|
|
||
|
void vkCmdExecuteCommands( VkCommandBuffer commandBuffer, uint32_t commandBufferCount, const VkCommandBuffer* pCommandBuffers ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdExecuteCommands( commandBuffer, commandBufferCount, pCommandBuffers );
|
||
|
}
|
||
|
|
||
|
void vkCmdFillBuffer( VkCommandBuffer commandBuffer, VkBuffer dstBuffer, VkDeviceSize dstOffset, VkDeviceSize size, uint32_t data ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdFillBuffer( commandBuffer, dstBuffer, dstOffset, size, data );
|
||
|
}
|
||
|
|
||
|
void vkCmdInsertDebugUtilsLabelEXT( VkCommandBuffer commandBuffer, const VkDebugUtilsLabelEXT* pLabelInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdInsertDebugUtilsLabelEXT( commandBuffer, pLabelInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdNextSubpass( VkCommandBuffer commandBuffer, VkSubpassContents contents ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdNextSubpass( commandBuffer, contents );
|
||
|
}
|
||
|
|
||
|
void vkCmdNextSubpass2KHR( VkCommandBuffer commandBuffer, const VkSubpassBeginInfoKHR* pSubpassBeginInfo, const VkSubpassEndInfoKHR* pSubpassEndInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdNextSubpass2KHR( commandBuffer, pSubpassBeginInfo, pSubpassEndInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdPipelineBarrier( VkCommandBuffer commandBuffer, VkPipelineStageFlags srcStageMask, VkPipelineStageFlags dstStageMask, VkDependencyFlags dependencyFlags, uint32_t memoryBarrierCount, const VkMemoryBarrier* pMemoryBarriers, uint32_t bufferMemoryBarrierCount, const VkBufferMemoryBarrier* pBufferMemoryBarriers, uint32_t imageMemoryBarrierCount, const VkImageMemoryBarrier* pImageMemoryBarriers ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdPipelineBarrier( commandBuffer, srcStageMask, dstStageMask, dependencyFlags, memoryBarrierCount, pMemoryBarriers, bufferMemoryBarrierCount, pBufferMemoryBarriers, imageMemoryBarrierCount, pImageMemoryBarriers );
|
||
|
}
|
||
|
|
||
|
void vkCmdProcessCommandsNVX( VkCommandBuffer commandBuffer, const VkCmdProcessCommandsInfoNVX* pProcessCommandsInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdProcessCommandsNVX( commandBuffer, pProcessCommandsInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdPushConstants( VkCommandBuffer commandBuffer, VkPipelineLayout layout, VkShaderStageFlags stageFlags, uint32_t offset, uint32_t size, const void* pValues ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdPushConstants( commandBuffer, layout, stageFlags, offset, size, pValues );
|
||
|
}
|
||
|
|
||
|
void vkCmdPushDescriptorSetKHR( VkCommandBuffer commandBuffer, VkPipelineBindPoint pipelineBindPoint, VkPipelineLayout layout, uint32_t set, uint32_t descriptorWriteCount, const VkWriteDescriptorSet* pDescriptorWrites ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdPushDescriptorSetKHR( commandBuffer, pipelineBindPoint, layout, set, descriptorWriteCount, pDescriptorWrites );
|
||
|
}
|
||
|
|
||
|
void vkCmdPushDescriptorSetWithTemplateKHR( VkCommandBuffer commandBuffer, VkDescriptorUpdateTemplate descriptorUpdateTemplate, VkPipelineLayout layout, uint32_t set, const void* pData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdPushDescriptorSetWithTemplateKHR( commandBuffer, descriptorUpdateTemplate, layout, set, pData );
|
||
|
}
|
||
|
|
||
|
void vkCmdReserveSpaceForCommandsNVX( VkCommandBuffer commandBuffer, const VkCmdReserveSpaceForCommandsInfoNVX* pReserveSpaceInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdReserveSpaceForCommandsNVX( commandBuffer, pReserveSpaceInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdResetEvent( VkCommandBuffer commandBuffer, VkEvent event, VkPipelineStageFlags stageMask ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdResetEvent( commandBuffer, event, stageMask );
|
||
|
}
|
||
|
|
||
|
void vkCmdResetQueryPool( VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t firstQuery, uint32_t queryCount ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdResetQueryPool( commandBuffer, queryPool, firstQuery, queryCount );
|
||
|
}
|
||
|
|
||
|
void vkCmdResolveImage( VkCommandBuffer commandBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage dstImage, VkImageLayout dstImageLayout, uint32_t regionCount, const VkImageResolve* pRegions ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdResolveImage( commandBuffer, srcImage, srcImageLayout, dstImage, dstImageLayout, regionCount, pRegions );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetBlendConstants( VkCommandBuffer commandBuffer, const float blendConstants[4] ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetBlendConstants( commandBuffer, blendConstants );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetCheckpointNV( VkCommandBuffer commandBuffer, const void* pCheckpointMarker ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetCheckpointNV( commandBuffer, pCheckpointMarker );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetCoarseSampleOrderNV( VkCommandBuffer commandBuffer, VkCoarseSampleOrderTypeNV sampleOrderType, uint32_t customSampleOrderCount, const VkCoarseSampleOrderCustomNV* pCustomSampleOrders ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetCoarseSampleOrderNV( commandBuffer, sampleOrderType, customSampleOrderCount, pCustomSampleOrders );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetDepthBias( VkCommandBuffer commandBuffer, float depthBiasConstantFactor, float depthBiasClamp, float depthBiasSlopeFactor ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetDepthBias( commandBuffer, depthBiasConstantFactor, depthBiasClamp, depthBiasSlopeFactor );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetDepthBounds( VkCommandBuffer commandBuffer, float minDepthBounds, float maxDepthBounds ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetDepthBounds( commandBuffer, minDepthBounds, maxDepthBounds );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetDeviceMask( VkCommandBuffer commandBuffer, uint32_t deviceMask ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetDeviceMask( commandBuffer, deviceMask );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetDeviceMaskKHR( VkCommandBuffer commandBuffer, uint32_t deviceMask ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetDeviceMaskKHR( commandBuffer, deviceMask );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetDiscardRectangleEXT( VkCommandBuffer commandBuffer, uint32_t firstDiscardRectangle, uint32_t discardRectangleCount, const VkRect2D* pDiscardRectangles ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetDiscardRectangleEXT( commandBuffer, firstDiscardRectangle, discardRectangleCount, pDiscardRectangles );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetEvent( VkCommandBuffer commandBuffer, VkEvent event, VkPipelineStageFlags stageMask ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetEvent( commandBuffer, event, stageMask );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetExclusiveScissorNV( VkCommandBuffer commandBuffer, uint32_t firstExclusiveScissor, uint32_t exclusiveScissorCount, const VkRect2D* pExclusiveScissors ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetExclusiveScissorNV( commandBuffer, firstExclusiveScissor, exclusiveScissorCount, pExclusiveScissors );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetLineStippleEXT( VkCommandBuffer commandBuffer, uint32_t lineStippleFactor, uint16_t lineStipplePattern ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetLineStippleEXT( commandBuffer, lineStippleFactor, lineStipplePattern );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetLineWidth( VkCommandBuffer commandBuffer, float lineWidth ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetLineWidth( commandBuffer, lineWidth );
|
||
|
}
|
||
|
|
||
|
VkResult vkCmdSetPerformanceMarkerINTEL( VkCommandBuffer commandBuffer, const VkPerformanceMarkerInfoINTEL* pMarkerInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetPerformanceMarkerINTEL( commandBuffer, pMarkerInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkCmdSetPerformanceOverrideINTEL( VkCommandBuffer commandBuffer, const VkPerformanceOverrideInfoINTEL* pOverrideInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetPerformanceOverrideINTEL( commandBuffer, pOverrideInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkCmdSetPerformanceStreamMarkerINTEL( VkCommandBuffer commandBuffer, const VkPerformanceStreamMarkerInfoINTEL* pMarkerInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetPerformanceStreamMarkerINTEL( commandBuffer, pMarkerInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetSampleLocationsEXT( VkCommandBuffer commandBuffer, const VkSampleLocationsInfoEXT* pSampleLocationsInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetSampleLocationsEXT( commandBuffer, pSampleLocationsInfo );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetScissor( VkCommandBuffer commandBuffer, uint32_t firstScissor, uint32_t scissorCount, const VkRect2D* pScissors ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetScissor( commandBuffer, firstScissor, scissorCount, pScissors );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetStencilCompareMask( VkCommandBuffer commandBuffer, VkStencilFaceFlags faceMask, uint32_t compareMask ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetStencilCompareMask( commandBuffer, faceMask, compareMask );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetStencilReference( VkCommandBuffer commandBuffer, VkStencilFaceFlags faceMask, uint32_t reference ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetStencilReference( commandBuffer, faceMask, reference );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetStencilWriteMask( VkCommandBuffer commandBuffer, VkStencilFaceFlags faceMask, uint32_t writeMask ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetStencilWriteMask( commandBuffer, faceMask, writeMask );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetViewport( VkCommandBuffer commandBuffer, uint32_t firstViewport, uint32_t viewportCount, const VkViewport* pViewports ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetViewport( commandBuffer, firstViewport, viewportCount, pViewports );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetViewportShadingRatePaletteNV( VkCommandBuffer commandBuffer, uint32_t firstViewport, uint32_t viewportCount, const VkShadingRatePaletteNV* pShadingRatePalettes ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetViewportShadingRatePaletteNV( commandBuffer, firstViewport, viewportCount, pShadingRatePalettes );
|
||
|
}
|
||
|
|
||
|
void vkCmdSetViewportWScalingNV( VkCommandBuffer commandBuffer, uint32_t firstViewport, uint32_t viewportCount, const VkViewportWScalingNV* pViewportWScalings ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdSetViewportWScalingNV( commandBuffer, firstViewport, viewportCount, pViewportWScalings );
|
||
|
}
|
||
|
|
||
|
void vkCmdTraceRaysNV( VkCommandBuffer commandBuffer, VkBuffer raygenShaderBindingTableBuffer, VkDeviceSize raygenShaderBindingOffset, VkBuffer missShaderBindingTableBuffer, VkDeviceSize missShaderBindingOffset, VkDeviceSize missShaderBindingStride, VkBuffer hitShaderBindingTableBuffer, VkDeviceSize hitShaderBindingOffset, VkDeviceSize hitShaderBindingStride, VkBuffer callableShaderBindingTableBuffer, VkDeviceSize callableShaderBindingOffset, VkDeviceSize callableShaderBindingStride, uint32_t width, uint32_t height, uint32_t depth ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdTraceRaysNV( commandBuffer, raygenShaderBindingTableBuffer, raygenShaderBindingOffset, missShaderBindingTableBuffer, missShaderBindingOffset, missShaderBindingStride, hitShaderBindingTableBuffer, hitShaderBindingOffset, hitShaderBindingStride, callableShaderBindingTableBuffer, callableShaderBindingOffset, callableShaderBindingStride, width, height, depth );
|
||
|
}
|
||
|
|
||
|
void vkCmdUpdateBuffer( VkCommandBuffer commandBuffer, VkBuffer dstBuffer, VkDeviceSize dstOffset, VkDeviceSize dataSize, const void* pData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdUpdateBuffer( commandBuffer, dstBuffer, dstOffset, dataSize, pData );
|
||
|
}
|
||
|
|
||
|
void vkCmdWaitEvents( VkCommandBuffer commandBuffer, uint32_t eventCount, const VkEvent* pEvents, VkPipelineStageFlags srcStageMask, VkPipelineStageFlags dstStageMask, uint32_t memoryBarrierCount, const VkMemoryBarrier* pMemoryBarriers, uint32_t bufferMemoryBarrierCount, const VkBufferMemoryBarrier* pBufferMemoryBarriers, uint32_t imageMemoryBarrierCount, const VkImageMemoryBarrier* pImageMemoryBarriers ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdWaitEvents( commandBuffer, eventCount, pEvents, srcStageMask, dstStageMask, memoryBarrierCount, pMemoryBarriers, bufferMemoryBarrierCount, pBufferMemoryBarriers, imageMemoryBarrierCount, pImageMemoryBarriers );
|
||
|
}
|
||
|
|
||
|
void vkCmdWriteAccelerationStructuresPropertiesNV( VkCommandBuffer commandBuffer, uint32_t accelerationStructureCount, const VkAccelerationStructureNV* pAccelerationStructures, VkQueryType queryType, VkQueryPool queryPool, uint32_t firstQuery ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdWriteAccelerationStructuresPropertiesNV( commandBuffer, accelerationStructureCount, pAccelerationStructures, queryType, queryPool, firstQuery );
|
||
|
}
|
||
|
|
||
|
void vkCmdWriteBufferMarkerAMD( VkCommandBuffer commandBuffer, VkPipelineStageFlagBits pipelineStage, VkBuffer dstBuffer, VkDeviceSize dstOffset, uint32_t marker ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdWriteBufferMarkerAMD( commandBuffer, pipelineStage, dstBuffer, dstOffset, marker );
|
||
|
}
|
||
|
|
||
|
void vkCmdWriteTimestamp( VkCommandBuffer commandBuffer, VkPipelineStageFlagBits pipelineStage, VkQueryPool queryPool, uint32_t query ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCmdWriteTimestamp( commandBuffer, pipelineStage, queryPool, query );
|
||
|
}
|
||
|
|
||
|
VkResult vkEndCommandBuffer( VkCommandBuffer commandBuffer ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkEndCommandBuffer( commandBuffer );
|
||
|
}
|
||
|
|
||
|
VkResult vkResetCommandBuffer( VkCommandBuffer commandBuffer, VkCommandBufferResetFlags flags ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkResetCommandBuffer( commandBuffer, flags );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkAcquireFullScreenExclusiveModeEXT( VkDevice device, VkSwapchainKHR swapchain ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkAcquireFullScreenExclusiveModeEXT( device, swapchain );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
VkResult vkAcquireNextImage2KHR( VkDevice device, const VkAcquireNextImageInfoKHR* pAcquireInfo, uint32_t* pImageIndex ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkAcquireNextImage2KHR( device, pAcquireInfo, pImageIndex );
|
||
|
}
|
||
|
|
||
|
VkResult vkAcquireNextImageKHR( VkDevice device, VkSwapchainKHR swapchain, uint64_t timeout, VkSemaphore semaphore, VkFence fence, uint32_t* pImageIndex ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkAcquireNextImageKHR( device, swapchain, timeout, semaphore, fence, pImageIndex );
|
||
|
}
|
||
|
|
||
|
VkResult vkAcquirePerformanceConfigurationINTEL( VkDevice device, const VkPerformanceConfigurationAcquireInfoINTEL* pAcquireInfo, VkPerformanceConfigurationINTEL* pConfiguration ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkAcquirePerformanceConfigurationINTEL( device, pAcquireInfo, pConfiguration );
|
||
|
}
|
||
|
|
||
|
VkResult vkAllocateCommandBuffers( VkDevice device, const VkCommandBufferAllocateInfo* pAllocateInfo, VkCommandBuffer* pCommandBuffers ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkAllocateCommandBuffers( device, pAllocateInfo, pCommandBuffers );
|
||
|
}
|
||
|
|
||
|
VkResult vkAllocateDescriptorSets( VkDevice device, const VkDescriptorSetAllocateInfo* pAllocateInfo, VkDescriptorSet* pDescriptorSets ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkAllocateDescriptorSets( device, pAllocateInfo, pDescriptorSets );
|
||
|
}
|
||
|
|
||
|
VkResult vkAllocateMemory( VkDevice device, const VkMemoryAllocateInfo* pAllocateInfo, const VkAllocationCallbacks* pAllocator, VkDeviceMemory* pMemory ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkAllocateMemory( device, pAllocateInfo, pAllocator, pMemory );
|
||
|
}
|
||
|
|
||
|
VkResult vkBindAccelerationStructureMemoryNV( VkDevice device, uint32_t bindInfoCount, const VkBindAccelerationStructureMemoryInfoNV* pBindInfos ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkBindAccelerationStructureMemoryNV( device, bindInfoCount, pBindInfos );
|
||
|
}
|
||
|
|
||
|
VkResult vkBindBufferMemory( VkDevice device, VkBuffer buffer, VkDeviceMemory memory, VkDeviceSize memoryOffset ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkBindBufferMemory( device, buffer, memory, memoryOffset );
|
||
|
}
|
||
|
|
||
|
VkResult vkBindBufferMemory2( VkDevice device, uint32_t bindInfoCount, const VkBindBufferMemoryInfo* pBindInfos ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkBindBufferMemory2( device, bindInfoCount, pBindInfos );
|
||
|
}
|
||
|
|
||
|
VkResult vkBindBufferMemory2KHR( VkDevice device, uint32_t bindInfoCount, const VkBindBufferMemoryInfo* pBindInfos ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkBindBufferMemory2KHR( device, bindInfoCount, pBindInfos );
|
||
|
}
|
||
|
|
||
|
VkResult vkBindImageMemory( VkDevice device, VkImage image, VkDeviceMemory memory, VkDeviceSize memoryOffset ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkBindImageMemory( device, image, memory, memoryOffset );
|
||
|
}
|
||
|
|
||
|
VkResult vkBindImageMemory2( VkDevice device, uint32_t bindInfoCount, const VkBindImageMemoryInfo* pBindInfos ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkBindImageMemory2( device, bindInfoCount, pBindInfos );
|
||
|
}
|
||
|
|
||
|
VkResult vkBindImageMemory2KHR( VkDevice device, uint32_t bindInfoCount, const VkBindImageMemoryInfo* pBindInfos ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkBindImageMemory2KHR( device, bindInfoCount, pBindInfos );
|
||
|
}
|
||
|
|
||
|
VkResult vkCompileDeferredNV( VkDevice device, VkPipeline pipeline, uint32_t shader ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCompileDeferredNV( device, pipeline, shader );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateAccelerationStructureNV( VkDevice device, const VkAccelerationStructureCreateInfoNV* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkAccelerationStructureNV* pAccelerationStructure ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateAccelerationStructureNV( device, pCreateInfo, pAllocator, pAccelerationStructure );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateBuffer( VkDevice device, const VkBufferCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkBuffer* pBuffer ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateBuffer( device, pCreateInfo, pAllocator, pBuffer );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateBufferView( VkDevice device, const VkBufferViewCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkBufferView* pView ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateBufferView( device, pCreateInfo, pAllocator, pView );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateCommandPool( VkDevice device, const VkCommandPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkCommandPool* pCommandPool ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateCommandPool( device, pCreateInfo, pAllocator, pCommandPool );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateComputePipelines( VkDevice device, VkPipelineCache pipelineCache, uint32_t createInfoCount, const VkComputePipelineCreateInfo* pCreateInfos, const VkAllocationCallbacks* pAllocator, VkPipeline* pPipelines ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateComputePipelines( device, pipelineCache, createInfoCount, pCreateInfos, pAllocator, pPipelines );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateDescriptorPool( VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateDescriptorPool( device, pCreateInfo, pAllocator, pDescriptorPool );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateDescriptorSetLayout( VkDevice device, const VkDescriptorSetLayoutCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorSetLayout* pSetLayout ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateDescriptorSetLayout( device, pCreateInfo, pAllocator, pSetLayout );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateDescriptorUpdateTemplate( VkDevice device, const VkDescriptorUpdateTemplateCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorUpdateTemplate* pDescriptorUpdateTemplate ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateDescriptorUpdateTemplate( device, pCreateInfo, pAllocator, pDescriptorUpdateTemplate );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateDescriptorUpdateTemplateKHR( VkDevice device, const VkDescriptorUpdateTemplateCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorUpdateTemplate* pDescriptorUpdateTemplate ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateDescriptorUpdateTemplateKHR( device, pCreateInfo, pAllocator, pDescriptorUpdateTemplate );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateEvent( VkDevice device, const VkEventCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkEvent* pEvent ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateEvent( device, pCreateInfo, pAllocator, pEvent );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateFence( VkDevice device, const VkFenceCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkFence* pFence ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateFence( device, pCreateInfo, pAllocator, pFence );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateFramebuffer( VkDevice device, const VkFramebufferCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkFramebuffer* pFramebuffer ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateFramebuffer( device, pCreateInfo, pAllocator, pFramebuffer );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateGraphicsPipelines( VkDevice device, VkPipelineCache pipelineCache, uint32_t createInfoCount, const VkGraphicsPipelineCreateInfo* pCreateInfos, const VkAllocationCallbacks* pAllocator, VkPipeline* pPipelines ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateGraphicsPipelines( device, pipelineCache, createInfoCount, pCreateInfos, pAllocator, pPipelines );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateImage( VkDevice device, const VkImageCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkImage* pImage ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateImage( device, pCreateInfo, pAllocator, pImage );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateImageView( VkDevice device, const VkImageViewCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkImageView* pView ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateImageView( device, pCreateInfo, pAllocator, pView );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateIndirectCommandsLayoutNVX( VkDevice device, const VkIndirectCommandsLayoutCreateInfoNVX* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkIndirectCommandsLayoutNVX* pIndirectCommandsLayout ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateIndirectCommandsLayoutNVX( device, pCreateInfo, pAllocator, pIndirectCommandsLayout );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateObjectTableNVX( VkDevice device, const VkObjectTableCreateInfoNVX* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkObjectTableNVX* pObjectTable ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateObjectTableNVX( device, pCreateInfo, pAllocator, pObjectTable );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreatePipelineCache( VkDevice device, const VkPipelineCacheCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkPipelineCache* pPipelineCache ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreatePipelineCache( device, pCreateInfo, pAllocator, pPipelineCache );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreatePipelineLayout( VkDevice device, const VkPipelineLayoutCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkPipelineLayout* pPipelineLayout ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreatePipelineLayout( device, pCreateInfo, pAllocator, pPipelineLayout );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateQueryPool( VkDevice device, const VkQueryPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkQueryPool* pQueryPool ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateQueryPool( device, pCreateInfo, pAllocator, pQueryPool );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateRayTracingPipelinesNV( VkDevice device, VkPipelineCache pipelineCache, uint32_t createInfoCount, const VkRayTracingPipelineCreateInfoNV* pCreateInfos, const VkAllocationCallbacks* pAllocator, VkPipeline* pPipelines ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateRayTracingPipelinesNV( device, pipelineCache, createInfoCount, pCreateInfos, pAllocator, pPipelines );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateRenderPass( VkDevice device, const VkRenderPassCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkRenderPass* pRenderPass ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateRenderPass( device, pCreateInfo, pAllocator, pRenderPass );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateRenderPass2KHR( VkDevice device, const VkRenderPassCreateInfo2KHR* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkRenderPass* pRenderPass ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateRenderPass2KHR( device, pCreateInfo, pAllocator, pRenderPass );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateSampler( VkDevice device, const VkSamplerCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSampler* pSampler ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateSampler( device, pCreateInfo, pAllocator, pSampler );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateSamplerYcbcrConversion( VkDevice device, const VkSamplerYcbcrConversionCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSamplerYcbcrConversion* pYcbcrConversion ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateSamplerYcbcrConversion( device, pCreateInfo, pAllocator, pYcbcrConversion );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateSamplerYcbcrConversionKHR( VkDevice device, const VkSamplerYcbcrConversionCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSamplerYcbcrConversion* pYcbcrConversion ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateSamplerYcbcrConversionKHR( device, pCreateInfo, pAllocator, pYcbcrConversion );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateSemaphore( VkDevice device, const VkSemaphoreCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSemaphore* pSemaphore ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateSemaphore( device, pCreateInfo, pAllocator, pSemaphore );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateShaderModule( VkDevice device, const VkShaderModuleCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkShaderModule* pShaderModule ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateShaderModule( device, pCreateInfo, pAllocator, pShaderModule );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateSharedSwapchainsKHR( VkDevice device, uint32_t swapchainCount, const VkSwapchainCreateInfoKHR* pCreateInfos, const VkAllocationCallbacks* pAllocator, VkSwapchainKHR* pSwapchains ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateSharedSwapchainsKHR( device, swapchainCount, pCreateInfos, pAllocator, pSwapchains );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateSwapchainKHR( VkDevice device, const VkSwapchainCreateInfoKHR* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSwapchainKHR* pSwapchain ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateSwapchainKHR( device, pCreateInfo, pAllocator, pSwapchain );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateValidationCacheEXT( VkDevice device, const VkValidationCacheCreateInfoEXT* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkValidationCacheEXT* pValidationCache ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateValidationCacheEXT( device, pCreateInfo, pAllocator, pValidationCache );
|
||
|
}
|
||
|
|
||
|
VkResult vkDebugMarkerSetObjectNameEXT( VkDevice device, const VkDebugMarkerObjectNameInfoEXT* pNameInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDebugMarkerSetObjectNameEXT( device, pNameInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkDebugMarkerSetObjectTagEXT( VkDevice device, const VkDebugMarkerObjectTagInfoEXT* pTagInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDebugMarkerSetObjectTagEXT( device, pTagInfo );
|
||
|
}
|
||
|
|
||
|
void vkDestroyAccelerationStructureNV( VkDevice device, VkAccelerationStructureNV accelerationStructure, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyAccelerationStructureNV( device, accelerationStructure, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyBuffer( VkDevice device, VkBuffer buffer, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyBuffer( device, buffer, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyBufferView( VkDevice device, VkBufferView bufferView, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyBufferView( device, bufferView, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyCommandPool( VkDevice device, VkCommandPool commandPool, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyCommandPool( device, commandPool, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyDescriptorPool( VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyDescriptorPool( device, descriptorPool, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyDescriptorSetLayout( VkDevice device, VkDescriptorSetLayout descriptorSetLayout, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyDescriptorSetLayout( device, descriptorSetLayout, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyDescriptorUpdateTemplate( VkDevice device, VkDescriptorUpdateTemplate descriptorUpdateTemplate, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyDescriptorUpdateTemplate( device, descriptorUpdateTemplate, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyDescriptorUpdateTemplateKHR( VkDevice device, VkDescriptorUpdateTemplate descriptorUpdateTemplate, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyDescriptorUpdateTemplateKHR( device, descriptorUpdateTemplate, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyDevice( VkDevice device, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyDevice( device, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyEvent( VkDevice device, VkEvent event, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyEvent( device, event, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyFence( VkDevice device, VkFence fence, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyFence( device, fence, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyFramebuffer( VkDevice device, VkFramebuffer framebuffer, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyFramebuffer( device, framebuffer, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyImage( VkDevice device, VkImage image, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyImage( device, image, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyImageView( VkDevice device, VkImageView imageView, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyImageView( device, imageView, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyIndirectCommandsLayoutNVX( VkDevice device, VkIndirectCommandsLayoutNVX indirectCommandsLayout, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyIndirectCommandsLayoutNVX( device, indirectCommandsLayout, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyObjectTableNVX( VkDevice device, VkObjectTableNVX objectTable, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyObjectTableNVX( device, objectTable, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyPipeline( VkDevice device, VkPipeline pipeline, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyPipeline( device, pipeline, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyPipelineCache( VkDevice device, VkPipelineCache pipelineCache, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyPipelineCache( device, pipelineCache, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyPipelineLayout( VkDevice device, VkPipelineLayout pipelineLayout, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyPipelineLayout( device, pipelineLayout, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyQueryPool( VkDevice device, VkQueryPool queryPool, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyQueryPool( device, queryPool, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyRenderPass( VkDevice device, VkRenderPass renderPass, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyRenderPass( device, renderPass, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroySampler( VkDevice device, VkSampler sampler, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroySampler( device, sampler, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroySamplerYcbcrConversion( VkDevice device, VkSamplerYcbcrConversion ycbcrConversion, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroySamplerYcbcrConversion( device, ycbcrConversion, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroySamplerYcbcrConversionKHR( VkDevice device, VkSamplerYcbcrConversion ycbcrConversion, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroySamplerYcbcrConversionKHR( device, ycbcrConversion, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroySemaphore( VkDevice device, VkSemaphore semaphore, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroySemaphore( device, semaphore, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyShaderModule( VkDevice device, VkShaderModule shaderModule, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyShaderModule( device, shaderModule, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroySwapchainKHR( VkDevice device, VkSwapchainKHR swapchain, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroySwapchainKHR( device, swapchain, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyValidationCacheEXT( VkDevice device, VkValidationCacheEXT validationCache, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyValidationCacheEXT( device, validationCache, pAllocator );
|
||
|
}
|
||
|
|
||
|
VkResult vkDeviceWaitIdle( VkDevice device ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDeviceWaitIdle( device );
|
||
|
}
|
||
|
|
||
|
VkResult vkDisplayPowerControlEXT( VkDevice device, VkDisplayKHR display, const VkDisplayPowerInfoEXT* pDisplayPowerInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDisplayPowerControlEXT( device, display, pDisplayPowerInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkFlushMappedMemoryRanges( VkDevice device, uint32_t memoryRangeCount, const VkMappedMemoryRange* pMemoryRanges ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkFlushMappedMemoryRanges( device, memoryRangeCount, pMemoryRanges );
|
||
|
}
|
||
|
|
||
|
void vkFreeCommandBuffers( VkDevice device, VkCommandPool commandPool, uint32_t commandBufferCount, const VkCommandBuffer* pCommandBuffers ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkFreeCommandBuffers( device, commandPool, commandBufferCount, pCommandBuffers );
|
||
|
}
|
||
|
|
||
|
VkResult vkFreeDescriptorSets( VkDevice device, VkDescriptorPool descriptorPool, uint32_t descriptorSetCount, const VkDescriptorSet* pDescriptorSets ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkFreeDescriptorSets( device, descriptorPool, descriptorSetCount, pDescriptorSets );
|
||
|
}
|
||
|
|
||
|
void vkFreeMemory( VkDevice device, VkDeviceMemory memory, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkFreeMemory( device, memory, pAllocator );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetAccelerationStructureHandleNV( VkDevice device, VkAccelerationStructureNV accelerationStructure, size_t dataSize, void* pData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetAccelerationStructureHandleNV( device, accelerationStructure, dataSize, pData );
|
||
|
}
|
||
|
|
||
|
void vkGetAccelerationStructureMemoryRequirementsNV( VkDevice device, const VkAccelerationStructureMemoryRequirementsInfoNV* pInfo, VkMemoryRequirements2KHR* pMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetAccelerationStructureMemoryRequirementsNV( device, pInfo, pMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_ANDROID_KHR
|
||
|
VkResult vkGetAndroidHardwareBufferPropertiesANDROID( VkDevice device, const struct AHardwareBuffer* buffer, VkAndroidHardwareBufferPropertiesANDROID* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetAndroidHardwareBufferPropertiesANDROID( device, buffer, pProperties );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_ANDROID_KHR*/
|
||
|
|
||
|
VkDeviceAddress vkGetBufferDeviceAddressEXT( VkDevice device, const VkBufferDeviceAddressInfoEXT* pInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetBufferDeviceAddressEXT( device, pInfo );
|
||
|
}
|
||
|
|
||
|
void vkGetBufferMemoryRequirements( VkDevice device, VkBuffer buffer, VkMemoryRequirements* pMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetBufferMemoryRequirements( device, buffer, pMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
void vkGetBufferMemoryRequirements2( VkDevice device, const VkBufferMemoryRequirementsInfo2* pInfo, VkMemoryRequirements2* pMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetBufferMemoryRequirements2( device, pInfo, pMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
void vkGetBufferMemoryRequirements2KHR( VkDevice device, const VkBufferMemoryRequirementsInfo2* pInfo, VkMemoryRequirements2* pMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetBufferMemoryRequirements2KHR( device, pInfo, pMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetCalibratedTimestampsEXT( VkDevice device, uint32_t timestampCount, const VkCalibratedTimestampInfoEXT* pTimestampInfos, uint64_t* pTimestamps, uint64_t* pMaxDeviation ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetCalibratedTimestampsEXT( device, timestampCount, pTimestampInfos, pTimestamps, pMaxDeviation );
|
||
|
}
|
||
|
|
||
|
void vkGetDescriptorSetLayoutSupport( VkDevice device, const VkDescriptorSetLayoutCreateInfo* pCreateInfo, VkDescriptorSetLayoutSupport* pSupport ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDescriptorSetLayoutSupport( device, pCreateInfo, pSupport );
|
||
|
}
|
||
|
|
||
|
void vkGetDescriptorSetLayoutSupportKHR( VkDevice device, const VkDescriptorSetLayoutCreateInfo* pCreateInfo, VkDescriptorSetLayoutSupport* pSupport ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDescriptorSetLayoutSupportKHR( device, pCreateInfo, pSupport );
|
||
|
}
|
||
|
|
||
|
void vkGetDeviceGroupPeerMemoryFeatures( VkDevice device, uint32_t heapIndex, uint32_t localDeviceIndex, uint32_t remoteDeviceIndex, VkPeerMemoryFeatureFlags* pPeerMemoryFeatures ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDeviceGroupPeerMemoryFeatures( device, heapIndex, localDeviceIndex, remoteDeviceIndex, pPeerMemoryFeatures );
|
||
|
}
|
||
|
|
||
|
void vkGetDeviceGroupPeerMemoryFeaturesKHR( VkDevice device, uint32_t heapIndex, uint32_t localDeviceIndex, uint32_t remoteDeviceIndex, VkPeerMemoryFeatureFlags* pPeerMemoryFeatures ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDeviceGroupPeerMemoryFeaturesKHR( device, heapIndex, localDeviceIndex, remoteDeviceIndex, pPeerMemoryFeatures );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetDeviceGroupPresentCapabilitiesKHR( VkDevice device, VkDeviceGroupPresentCapabilitiesKHR* pDeviceGroupPresentCapabilities ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDeviceGroupPresentCapabilitiesKHR( device, pDeviceGroupPresentCapabilities );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkGetDeviceGroupSurfacePresentModes2EXT( VkDevice device, const VkPhysicalDeviceSurfaceInfo2KHR* pSurfaceInfo, VkDeviceGroupPresentModeFlagsKHR* pModes ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDeviceGroupSurfacePresentModes2EXT( device, pSurfaceInfo, pModes );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
VkResult vkGetDeviceGroupSurfacePresentModesKHR( VkDevice device, VkSurfaceKHR surface, VkDeviceGroupPresentModeFlagsKHR* pModes ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDeviceGroupSurfacePresentModesKHR( device, surface, pModes );
|
||
|
}
|
||
|
|
||
|
void vkGetDeviceMemoryCommitment( VkDevice device, VkDeviceMemory memory, VkDeviceSize* pCommittedMemoryInBytes ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDeviceMemoryCommitment( device, memory, pCommittedMemoryInBytes );
|
||
|
}
|
||
|
|
||
|
PFN_vkVoidFunction vkGetDeviceProcAddr( VkDevice device, const char* pName ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDeviceProcAddr( device, pName );
|
||
|
}
|
||
|
|
||
|
void vkGetDeviceQueue( VkDevice device, uint32_t queueFamilyIndex, uint32_t queueIndex, VkQueue* pQueue ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDeviceQueue( device, queueFamilyIndex, queueIndex, pQueue );
|
||
|
}
|
||
|
|
||
|
void vkGetDeviceQueue2( VkDevice device, const VkDeviceQueueInfo2* pQueueInfo, VkQueue* pQueue ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDeviceQueue2( device, pQueueInfo, pQueue );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetEventStatus( VkDevice device, VkEvent event ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetEventStatus( device, event );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetFenceFdKHR( VkDevice device, const VkFenceGetFdInfoKHR* pGetFdInfo, int* pFd ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetFenceFdKHR( device, pGetFdInfo, pFd );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetFenceStatus( VkDevice device, VkFence fence ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetFenceStatus( device, fence );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkGetFenceWin32HandleKHR( VkDevice device, const VkFenceGetWin32HandleInfoKHR* pGetWin32HandleInfo, HANDLE* pHandle ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetFenceWin32HandleKHR( device, pGetWin32HandleInfo, pHandle );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
VkResult vkGetImageDrmFormatModifierPropertiesEXT( VkDevice device, VkImage image, VkImageDrmFormatModifierPropertiesEXT* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetImageDrmFormatModifierPropertiesEXT( device, image, pProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetImageMemoryRequirements( VkDevice device, VkImage image, VkMemoryRequirements* pMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetImageMemoryRequirements( device, image, pMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
void vkGetImageMemoryRequirements2( VkDevice device, const VkImageMemoryRequirementsInfo2* pInfo, VkMemoryRequirements2* pMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetImageMemoryRequirements2( device, pInfo, pMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
void vkGetImageMemoryRequirements2KHR( VkDevice device, const VkImageMemoryRequirementsInfo2* pInfo, VkMemoryRequirements2* pMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetImageMemoryRequirements2KHR( device, pInfo, pMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
void vkGetImageSparseMemoryRequirements( VkDevice device, VkImage image, uint32_t* pSparseMemoryRequirementCount, VkSparseImageMemoryRequirements* pSparseMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetImageSparseMemoryRequirements( device, image, pSparseMemoryRequirementCount, pSparseMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
void vkGetImageSparseMemoryRequirements2( VkDevice device, const VkImageSparseMemoryRequirementsInfo2* pInfo, uint32_t* pSparseMemoryRequirementCount, VkSparseImageMemoryRequirements2* pSparseMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetImageSparseMemoryRequirements2( device, pInfo, pSparseMemoryRequirementCount, pSparseMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
void vkGetImageSparseMemoryRequirements2KHR( VkDevice device, const VkImageSparseMemoryRequirementsInfo2* pInfo, uint32_t* pSparseMemoryRequirementCount, VkSparseImageMemoryRequirements2* pSparseMemoryRequirements ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetImageSparseMemoryRequirements2KHR( device, pInfo, pSparseMemoryRequirementCount, pSparseMemoryRequirements );
|
||
|
}
|
||
|
|
||
|
void vkGetImageSubresourceLayout( VkDevice device, VkImage image, const VkImageSubresource* pSubresource, VkSubresourceLayout* pLayout ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetImageSubresourceLayout( device, image, pSubresource, pLayout );
|
||
|
}
|
||
|
|
||
|
uint32_t vkGetImageViewHandleNVX( VkDevice device, const VkImageViewHandleInfoNVX* pInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetImageViewHandleNVX( device, pInfo );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_ANDROID_KHR
|
||
|
VkResult vkGetMemoryAndroidHardwareBufferANDROID( VkDevice device, const VkMemoryGetAndroidHardwareBufferInfoANDROID* pInfo, struct AHardwareBuffer** pBuffer ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetMemoryAndroidHardwareBufferANDROID( device, pInfo, pBuffer );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_ANDROID_KHR*/
|
||
|
|
||
|
VkResult vkGetMemoryFdKHR( VkDevice device, const VkMemoryGetFdInfoKHR* pGetFdInfo, int* pFd ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetMemoryFdKHR( device, pGetFdInfo, pFd );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetMemoryFdPropertiesKHR( VkDevice device, VkExternalMemoryHandleTypeFlagBits handleType, int fd, VkMemoryFdPropertiesKHR* pMemoryFdProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetMemoryFdPropertiesKHR( device, handleType, fd, pMemoryFdProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetMemoryHostPointerPropertiesEXT( VkDevice device, VkExternalMemoryHandleTypeFlagBits handleType, const void* pHostPointer, VkMemoryHostPointerPropertiesEXT* pMemoryHostPointerProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetMemoryHostPointerPropertiesEXT( device, handleType, pHostPointer, pMemoryHostPointerProperties );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkGetMemoryWin32HandleKHR( VkDevice device, const VkMemoryGetWin32HandleInfoKHR* pGetWin32HandleInfo, HANDLE* pHandle ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetMemoryWin32HandleKHR( device, pGetWin32HandleInfo, pHandle );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkGetMemoryWin32HandleNV( VkDevice device, VkDeviceMemory memory, VkExternalMemoryHandleTypeFlagsNV handleType, HANDLE* pHandle ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetMemoryWin32HandleNV( device, memory, handleType, pHandle );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkGetMemoryWin32HandlePropertiesKHR( VkDevice device, VkExternalMemoryHandleTypeFlagBits handleType, HANDLE handle, VkMemoryWin32HandlePropertiesKHR* pMemoryWin32HandleProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetMemoryWin32HandlePropertiesKHR( device, handleType, handle, pMemoryWin32HandleProperties );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
VkResult vkGetPastPresentationTimingGOOGLE( VkDevice device, VkSwapchainKHR swapchain, uint32_t* pPresentationTimingCount, VkPastPresentationTimingGOOGLE* pPresentationTimings ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPastPresentationTimingGOOGLE( device, swapchain, pPresentationTimingCount, pPresentationTimings );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPerformanceParameterINTEL( VkDevice device, VkPerformanceParameterTypeINTEL parameter, VkPerformanceValueINTEL* pValue ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPerformanceParameterINTEL( device, parameter, pValue );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPipelineCacheData( VkDevice device, VkPipelineCache pipelineCache, size_t* pDataSize, void* pData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPipelineCacheData( device, pipelineCache, pDataSize, pData );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPipelineExecutableInternalRepresentationsKHR( VkDevice device, const VkPipelineExecutableInfoKHR* pExecutableInfo, uint32_t* pInternalRepresentationCount, VkPipelineExecutableInternalRepresentationKHR* pInternalRepresentations ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPipelineExecutableInternalRepresentationsKHR( device, pExecutableInfo, pInternalRepresentationCount, pInternalRepresentations );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPipelineExecutablePropertiesKHR( VkDevice device, const VkPipelineInfoKHR* pPipelineInfo, uint32_t* pExecutableCount, VkPipelineExecutablePropertiesKHR* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPipelineExecutablePropertiesKHR( device, pPipelineInfo, pExecutableCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPipelineExecutableStatisticsKHR( VkDevice device, const VkPipelineExecutableInfoKHR* pExecutableInfo, uint32_t* pStatisticCount, VkPipelineExecutableStatisticKHR* pStatistics ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPipelineExecutableStatisticsKHR( device, pExecutableInfo, pStatisticCount, pStatistics );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetQueryPoolResults( VkDevice device, VkQueryPool queryPool, uint32_t firstQuery, uint32_t queryCount, size_t dataSize, void* pData, VkDeviceSize stride, VkQueryResultFlags flags ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetQueryPoolResults( device, queryPool, firstQuery, queryCount, dataSize, pData, stride, flags );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetRayTracingShaderGroupHandlesNV( VkDevice device, VkPipeline pipeline, uint32_t firstGroup, uint32_t groupCount, size_t dataSize, void* pData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetRayTracingShaderGroupHandlesNV( device, pipeline, firstGroup, groupCount, dataSize, pData );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetRefreshCycleDurationGOOGLE( VkDevice device, VkSwapchainKHR swapchain, VkRefreshCycleDurationGOOGLE* pDisplayTimingProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetRefreshCycleDurationGOOGLE( device, swapchain, pDisplayTimingProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetRenderAreaGranularity( VkDevice device, VkRenderPass renderPass, VkExtent2D* pGranularity ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetRenderAreaGranularity( device, renderPass, pGranularity );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetSemaphoreFdKHR( VkDevice device, const VkSemaphoreGetFdInfoKHR* pGetFdInfo, int* pFd ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetSemaphoreFdKHR( device, pGetFdInfo, pFd );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkGetSemaphoreWin32HandleKHR( VkDevice device, const VkSemaphoreGetWin32HandleInfoKHR* pGetWin32HandleInfo, HANDLE* pHandle ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetSemaphoreWin32HandleKHR( device, pGetWin32HandleInfo, pHandle );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
VkResult vkGetShaderInfoAMD( VkDevice device, VkPipeline pipeline, VkShaderStageFlagBits shaderStage, VkShaderInfoTypeAMD infoType, size_t* pInfoSize, void* pInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetShaderInfoAMD( device, pipeline, shaderStage, infoType, pInfoSize, pInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetSwapchainCounterEXT( VkDevice device, VkSwapchainKHR swapchain, VkSurfaceCounterFlagBitsEXT counter, uint64_t* pCounterValue ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetSwapchainCounterEXT( device, swapchain, counter, pCounterValue );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetSwapchainImagesKHR( VkDevice device, VkSwapchainKHR swapchain, uint32_t* pSwapchainImageCount, VkImage* pSwapchainImages ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetSwapchainImagesKHR( device, swapchain, pSwapchainImageCount, pSwapchainImages );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetSwapchainStatusKHR( VkDevice device, VkSwapchainKHR swapchain ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetSwapchainStatusKHR( device, swapchain );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetValidationCacheDataEXT( VkDevice device, VkValidationCacheEXT validationCache, size_t* pDataSize, void* pData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetValidationCacheDataEXT( device, validationCache, pDataSize, pData );
|
||
|
}
|
||
|
|
||
|
VkResult vkImportFenceFdKHR( VkDevice device, const VkImportFenceFdInfoKHR* pImportFenceFdInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkImportFenceFdKHR( device, pImportFenceFdInfo );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkImportFenceWin32HandleKHR( VkDevice device, const VkImportFenceWin32HandleInfoKHR* pImportFenceWin32HandleInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkImportFenceWin32HandleKHR( device, pImportFenceWin32HandleInfo );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
VkResult vkImportSemaphoreFdKHR( VkDevice device, const VkImportSemaphoreFdInfoKHR* pImportSemaphoreFdInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkImportSemaphoreFdKHR( device, pImportSemaphoreFdInfo );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkImportSemaphoreWin32HandleKHR( VkDevice device, const VkImportSemaphoreWin32HandleInfoKHR* pImportSemaphoreWin32HandleInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkImportSemaphoreWin32HandleKHR( device, pImportSemaphoreWin32HandleInfo );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
VkResult vkInitializePerformanceApiINTEL( VkDevice device, const VkInitializePerformanceApiInfoINTEL* pInitializeInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkInitializePerformanceApiINTEL( device, pInitializeInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkInvalidateMappedMemoryRanges( VkDevice device, uint32_t memoryRangeCount, const VkMappedMemoryRange* pMemoryRanges ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkInvalidateMappedMemoryRanges( device, memoryRangeCount, pMemoryRanges );
|
||
|
}
|
||
|
|
||
|
VkResult vkMapMemory( VkDevice device, VkDeviceMemory memory, VkDeviceSize offset, VkDeviceSize size, VkMemoryMapFlags flags, void** ppData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkMapMemory( device, memory, offset, size, flags, ppData );
|
||
|
}
|
||
|
|
||
|
VkResult vkMergePipelineCaches( VkDevice device, VkPipelineCache dstCache, uint32_t srcCacheCount, const VkPipelineCache* pSrcCaches ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkMergePipelineCaches( device, dstCache, srcCacheCount, pSrcCaches );
|
||
|
}
|
||
|
|
||
|
VkResult vkMergeValidationCachesEXT( VkDevice device, VkValidationCacheEXT dstCache, uint32_t srcCacheCount, const VkValidationCacheEXT* pSrcCaches ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkMergeValidationCachesEXT( device, dstCache, srcCacheCount, pSrcCaches );
|
||
|
}
|
||
|
|
||
|
VkResult vkRegisterDeviceEventEXT( VkDevice device, const VkDeviceEventInfoEXT* pDeviceEventInfo, const VkAllocationCallbacks* pAllocator, VkFence* pFence ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkRegisterDeviceEventEXT( device, pDeviceEventInfo, pAllocator, pFence );
|
||
|
}
|
||
|
|
||
|
VkResult vkRegisterDisplayEventEXT( VkDevice device, VkDisplayKHR display, const VkDisplayEventInfoEXT* pDisplayEventInfo, const VkAllocationCallbacks* pAllocator, VkFence* pFence ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkRegisterDisplayEventEXT( device, display, pDisplayEventInfo, pAllocator, pFence );
|
||
|
}
|
||
|
|
||
|
VkResult vkRegisterObjectsNVX( VkDevice device, VkObjectTableNVX objectTable, uint32_t objectCount, const VkObjectTableEntryNVX* const* ppObjectTableEntries, const uint32_t* pObjectIndices ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkRegisterObjectsNVX( device, objectTable, objectCount, ppObjectTableEntries, pObjectIndices );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkReleaseFullScreenExclusiveModeEXT( VkDevice device, VkSwapchainKHR swapchain ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkReleaseFullScreenExclusiveModeEXT( device, swapchain );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
VkResult vkReleasePerformanceConfigurationINTEL( VkDevice device, VkPerformanceConfigurationINTEL configuration ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkReleasePerformanceConfigurationINTEL( device, configuration );
|
||
|
}
|
||
|
|
||
|
VkResult vkResetCommandPool( VkDevice device, VkCommandPool commandPool, VkCommandPoolResetFlags flags ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkResetCommandPool( device, commandPool, flags );
|
||
|
}
|
||
|
|
||
|
VkResult vkResetDescriptorPool( VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkResetDescriptorPool( device, descriptorPool, flags );
|
||
|
}
|
||
|
|
||
|
VkResult vkResetEvent( VkDevice device, VkEvent event ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkResetEvent( device, event );
|
||
|
}
|
||
|
|
||
|
VkResult vkResetFences( VkDevice device, uint32_t fenceCount, const VkFence* pFences ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkResetFences( device, fenceCount, pFences );
|
||
|
}
|
||
|
|
||
|
void vkResetQueryPoolEXT( VkDevice device, VkQueryPool queryPool, uint32_t firstQuery, uint32_t queryCount ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkResetQueryPoolEXT( device, queryPool, firstQuery, queryCount );
|
||
|
}
|
||
|
|
||
|
VkResult vkSetDebugUtilsObjectNameEXT( VkDevice device, const VkDebugUtilsObjectNameInfoEXT* pNameInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkSetDebugUtilsObjectNameEXT( device, pNameInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkSetDebugUtilsObjectTagEXT( VkDevice device, const VkDebugUtilsObjectTagInfoEXT* pTagInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkSetDebugUtilsObjectTagEXT( device, pTagInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkSetEvent( VkDevice device, VkEvent event ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkSetEvent( device, event );
|
||
|
}
|
||
|
|
||
|
void vkSetHdrMetadataEXT( VkDevice device, uint32_t swapchainCount, const VkSwapchainKHR* pSwapchains, const VkHdrMetadataEXT* pMetadata ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkSetHdrMetadataEXT( device, swapchainCount, pSwapchains, pMetadata );
|
||
|
}
|
||
|
|
||
|
void vkSetLocalDimmingAMD( VkDevice device, VkSwapchainKHR swapChain, VkBool32 localDimmingEnable ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkSetLocalDimmingAMD( device, swapChain, localDimmingEnable );
|
||
|
}
|
||
|
|
||
|
void vkTrimCommandPool( VkDevice device, VkCommandPool commandPool, VkCommandPoolTrimFlags flags ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkTrimCommandPool( device, commandPool, flags );
|
||
|
}
|
||
|
|
||
|
void vkTrimCommandPoolKHR( VkDevice device, VkCommandPool commandPool, VkCommandPoolTrimFlags flags ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkTrimCommandPoolKHR( device, commandPool, flags );
|
||
|
}
|
||
|
|
||
|
void vkUninitializePerformanceApiINTEL( VkDevice device ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkUninitializePerformanceApiINTEL( device );
|
||
|
}
|
||
|
|
||
|
void vkUnmapMemory( VkDevice device, VkDeviceMemory memory ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkUnmapMemory( device, memory );
|
||
|
}
|
||
|
|
||
|
VkResult vkUnregisterObjectsNVX( VkDevice device, VkObjectTableNVX objectTable, uint32_t objectCount, const VkObjectEntryTypeNVX* pObjectEntryTypes, const uint32_t* pObjectIndices ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkUnregisterObjectsNVX( device, objectTable, objectCount, pObjectEntryTypes, pObjectIndices );
|
||
|
}
|
||
|
|
||
|
void vkUpdateDescriptorSetWithTemplate( VkDevice device, VkDescriptorSet descriptorSet, VkDescriptorUpdateTemplate descriptorUpdateTemplate, const void* pData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkUpdateDescriptorSetWithTemplate( device, descriptorSet, descriptorUpdateTemplate, pData );
|
||
|
}
|
||
|
|
||
|
void vkUpdateDescriptorSetWithTemplateKHR( VkDevice device, VkDescriptorSet descriptorSet, VkDescriptorUpdateTemplate descriptorUpdateTemplate, const void* pData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkUpdateDescriptorSetWithTemplateKHR( device, descriptorSet, descriptorUpdateTemplate, pData );
|
||
|
}
|
||
|
|
||
|
void vkUpdateDescriptorSets( VkDevice device, uint32_t descriptorWriteCount, const VkWriteDescriptorSet* pDescriptorWrites, uint32_t descriptorCopyCount, const VkCopyDescriptorSet* pDescriptorCopies ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkUpdateDescriptorSets( device, descriptorWriteCount, pDescriptorWrites, descriptorCopyCount, pDescriptorCopies );
|
||
|
}
|
||
|
|
||
|
VkResult vkWaitForFences( VkDevice device, uint32_t fenceCount, const VkFence* pFences, VkBool32 waitAll, uint64_t timeout ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkWaitForFences( device, fenceCount, pFences, waitAll, timeout );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_ANDROID_KHR
|
||
|
VkResult vkCreateAndroidSurfaceKHR( VkInstance instance, const VkAndroidSurfaceCreateInfoKHR* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateAndroidSurfaceKHR( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_ANDROID_KHR*/
|
||
|
|
||
|
VkResult vkCreateDebugReportCallbackEXT( VkInstance instance, const VkDebugReportCallbackCreateInfoEXT* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDebugReportCallbackEXT* pCallback ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateDebugReportCallbackEXT( instance, pCreateInfo, pAllocator, pCallback );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateDebugUtilsMessengerEXT( VkInstance instance, const VkDebugUtilsMessengerCreateInfoEXT* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDebugUtilsMessengerEXT* pMessenger ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateDebugUtilsMessengerEXT( instance, pCreateInfo, pAllocator, pMessenger );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateDisplayPlaneSurfaceKHR( VkInstance instance, const VkDisplaySurfaceCreateInfoKHR* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateDisplayPlaneSurfaceKHR( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateHeadlessSurfaceEXT( VkInstance instance, const VkHeadlessSurfaceCreateInfoEXT* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateHeadlessSurfaceEXT( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_IOS_MVK
|
||
|
VkResult vkCreateIOSSurfaceMVK( VkInstance instance, const VkIOSSurfaceCreateInfoMVK* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateIOSSurfaceMVK( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_IOS_MVK*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_FUCHSIA
|
||
|
VkResult vkCreateImagePipeSurfaceFUCHSIA( VkInstance instance, const VkImagePipeSurfaceCreateInfoFUCHSIA* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateImagePipeSurfaceFUCHSIA( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_FUCHSIA*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_MACOS_MVK
|
||
|
VkResult vkCreateMacOSSurfaceMVK( VkInstance instance, const VkMacOSSurfaceCreateInfoMVK* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateMacOSSurfaceMVK( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_MACOS_MVK*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_METAL_EXT
|
||
|
VkResult vkCreateMetalSurfaceEXT( VkInstance instance, const VkMetalSurfaceCreateInfoEXT* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateMetalSurfaceEXT( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_METAL_EXT*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_GGP
|
||
|
VkResult vkCreateStreamDescriptorSurfaceGGP( VkInstance instance, const VkStreamDescriptorSurfaceCreateInfoGGP* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateStreamDescriptorSurfaceGGP( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_GGP*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_VI_NN
|
||
|
VkResult vkCreateViSurfaceNN( VkInstance instance, const VkViSurfaceCreateInfoNN* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateViSurfaceNN( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_VI_NN*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WAYLAND_KHR
|
||
|
VkResult vkCreateWaylandSurfaceKHR( VkInstance instance, const VkWaylandSurfaceCreateInfoKHR* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateWaylandSurfaceKHR( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WAYLAND_KHR*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkCreateWin32SurfaceKHR( VkInstance instance, const VkWin32SurfaceCreateInfoKHR* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateWin32SurfaceKHR( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_XCB_KHR
|
||
|
VkResult vkCreateXcbSurfaceKHR( VkInstance instance, const VkXcbSurfaceCreateInfoKHR* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateXcbSurfaceKHR( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_XCB_KHR*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_XLIB_KHR
|
||
|
VkResult vkCreateXlibSurfaceKHR( VkInstance instance, const VkXlibSurfaceCreateInfoKHR* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSurfaceKHR* pSurface ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateXlibSurfaceKHR( instance, pCreateInfo, pAllocator, pSurface );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_XLIB_KHR*/
|
||
|
|
||
|
void vkDebugReportMessageEXT( VkInstance instance, VkDebugReportFlagsEXT flags, VkDebugReportObjectTypeEXT objectType, uint64_t object, size_t location, int32_t messageCode, const char* pLayerPrefix, const char* pMessage ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDebugReportMessageEXT( instance, flags, objectType, object, location, messageCode, pLayerPrefix, pMessage );
|
||
|
}
|
||
|
|
||
|
void vkDestroyDebugReportCallbackEXT( VkInstance instance, VkDebugReportCallbackEXT callback, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyDebugReportCallbackEXT( instance, callback, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyDebugUtilsMessengerEXT( VkInstance instance, VkDebugUtilsMessengerEXT messenger, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyDebugUtilsMessengerEXT( instance, messenger, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroyInstance( VkInstance instance, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroyInstance( instance, pAllocator );
|
||
|
}
|
||
|
|
||
|
void vkDestroySurfaceKHR( VkInstance instance, VkSurfaceKHR surface, const VkAllocationCallbacks* pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkDestroySurfaceKHR( instance, surface, pAllocator );
|
||
|
}
|
||
|
|
||
|
VkResult vkEnumeratePhysicalDeviceGroups( VkInstance instance, uint32_t* pPhysicalDeviceGroupCount, VkPhysicalDeviceGroupProperties* pPhysicalDeviceGroupProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkEnumeratePhysicalDeviceGroups( instance, pPhysicalDeviceGroupCount, pPhysicalDeviceGroupProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkEnumeratePhysicalDeviceGroupsKHR( VkInstance instance, uint32_t* pPhysicalDeviceGroupCount, VkPhysicalDeviceGroupProperties* pPhysicalDeviceGroupProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkEnumeratePhysicalDeviceGroupsKHR( instance, pPhysicalDeviceGroupCount, pPhysicalDeviceGroupProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkEnumeratePhysicalDevices( VkInstance instance, uint32_t* pPhysicalDeviceCount, VkPhysicalDevice* pPhysicalDevices ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkEnumeratePhysicalDevices( instance, pPhysicalDeviceCount, pPhysicalDevices );
|
||
|
}
|
||
|
|
||
|
PFN_vkVoidFunction vkGetInstanceProcAddr( VkInstance instance, const char* pName ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetInstanceProcAddr( instance, pName );
|
||
|
}
|
||
|
|
||
|
void vkSubmitDebugUtilsMessageEXT( VkInstance instance, VkDebugUtilsMessageSeverityFlagBitsEXT messageSeverity, VkDebugUtilsMessageTypeFlagsEXT messageTypes, const VkDebugUtilsMessengerCallbackDataEXT* pCallbackData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkSubmitDebugUtilsMessageEXT( instance, messageSeverity, messageTypes, pCallbackData );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_XLIB_XRANDR_EXT
|
||
|
VkResult vkAcquireXlibDisplayEXT( VkPhysicalDevice physicalDevice, Display* dpy, VkDisplayKHR display ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkAcquireXlibDisplayEXT( physicalDevice, dpy, display );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_XLIB_XRANDR_EXT*/
|
||
|
|
||
|
VkResult vkCreateDevice( VkPhysicalDevice physicalDevice, const VkDeviceCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDevice* pDevice ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateDevice( physicalDevice, pCreateInfo, pAllocator, pDevice );
|
||
|
}
|
||
|
|
||
|
VkResult vkCreateDisplayModeKHR( VkPhysicalDevice physicalDevice, VkDisplayKHR display, const VkDisplayModeCreateInfoKHR* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDisplayModeKHR* pMode ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkCreateDisplayModeKHR( physicalDevice, display, pCreateInfo, pAllocator, pMode );
|
||
|
}
|
||
|
|
||
|
VkResult vkEnumerateDeviceExtensionProperties( VkPhysicalDevice physicalDevice, const char* pLayerName, uint32_t* pPropertyCount, VkExtensionProperties* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkEnumerateDeviceExtensionProperties( physicalDevice, pLayerName, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkEnumerateDeviceLayerProperties( VkPhysicalDevice physicalDevice, uint32_t* pPropertyCount, VkLayerProperties* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkEnumerateDeviceLayerProperties( physicalDevice, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetDisplayModeProperties2KHR( VkPhysicalDevice physicalDevice, VkDisplayKHR display, uint32_t* pPropertyCount, VkDisplayModeProperties2KHR* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDisplayModeProperties2KHR( physicalDevice, display, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetDisplayModePropertiesKHR( VkPhysicalDevice physicalDevice, VkDisplayKHR display, uint32_t* pPropertyCount, VkDisplayModePropertiesKHR* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDisplayModePropertiesKHR( physicalDevice, display, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetDisplayPlaneCapabilities2KHR( VkPhysicalDevice physicalDevice, const VkDisplayPlaneInfo2KHR* pDisplayPlaneInfo, VkDisplayPlaneCapabilities2KHR* pCapabilities ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDisplayPlaneCapabilities2KHR( physicalDevice, pDisplayPlaneInfo, pCapabilities );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetDisplayPlaneCapabilitiesKHR( VkPhysicalDevice physicalDevice, VkDisplayModeKHR mode, uint32_t planeIndex, VkDisplayPlaneCapabilitiesKHR* pCapabilities ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDisplayPlaneCapabilitiesKHR( physicalDevice, mode, planeIndex, pCapabilities );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetDisplayPlaneSupportedDisplaysKHR( VkPhysicalDevice physicalDevice, uint32_t planeIndex, uint32_t* pDisplayCount, VkDisplayKHR* pDisplays ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetDisplayPlaneSupportedDisplaysKHR( physicalDevice, planeIndex, pDisplayCount, pDisplays );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceCalibrateableTimeDomainsEXT( VkPhysicalDevice physicalDevice, uint32_t* pTimeDomainCount, VkTimeDomainEXT* pTimeDomains ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceCalibrateableTimeDomainsEXT( physicalDevice, pTimeDomainCount, pTimeDomains );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceCooperativeMatrixPropertiesNV( VkPhysicalDevice physicalDevice, uint32_t* pPropertyCount, VkCooperativeMatrixPropertiesNV* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceCooperativeMatrixPropertiesNV( physicalDevice, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceDisplayPlaneProperties2KHR( VkPhysicalDevice physicalDevice, uint32_t* pPropertyCount, VkDisplayPlaneProperties2KHR* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceDisplayPlaneProperties2KHR( physicalDevice, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceDisplayPlanePropertiesKHR( VkPhysicalDevice physicalDevice, uint32_t* pPropertyCount, VkDisplayPlanePropertiesKHR* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceDisplayPlanePropertiesKHR( physicalDevice, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceDisplayProperties2KHR( VkPhysicalDevice physicalDevice, uint32_t* pPropertyCount, VkDisplayProperties2KHR* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceDisplayProperties2KHR( physicalDevice, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceDisplayPropertiesKHR( VkPhysicalDevice physicalDevice, uint32_t* pPropertyCount, VkDisplayPropertiesKHR* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceDisplayPropertiesKHR( physicalDevice, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceExternalBufferProperties( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceExternalBufferInfo* pExternalBufferInfo, VkExternalBufferProperties* pExternalBufferProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceExternalBufferProperties( physicalDevice, pExternalBufferInfo, pExternalBufferProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceExternalBufferPropertiesKHR( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceExternalBufferInfo* pExternalBufferInfo, VkExternalBufferProperties* pExternalBufferProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceExternalBufferPropertiesKHR( physicalDevice, pExternalBufferInfo, pExternalBufferProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceExternalFenceProperties( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceExternalFenceInfo* pExternalFenceInfo, VkExternalFenceProperties* pExternalFenceProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceExternalFenceProperties( physicalDevice, pExternalFenceInfo, pExternalFenceProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceExternalFencePropertiesKHR( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceExternalFenceInfo* pExternalFenceInfo, VkExternalFenceProperties* pExternalFenceProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceExternalFencePropertiesKHR( physicalDevice, pExternalFenceInfo, pExternalFenceProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceExternalImageFormatPropertiesNV( VkPhysicalDevice physicalDevice, VkFormat format, VkImageType type, VkImageTiling tiling, VkImageUsageFlags usage, VkImageCreateFlags flags, VkExternalMemoryHandleTypeFlagsNV externalHandleType, VkExternalImageFormatPropertiesNV* pExternalImageFormatProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceExternalImageFormatPropertiesNV( physicalDevice, format, type, tiling, usage, flags, externalHandleType, pExternalImageFormatProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceExternalSemaphoreProperties( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceExternalSemaphoreInfo* pExternalSemaphoreInfo, VkExternalSemaphoreProperties* pExternalSemaphoreProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceExternalSemaphoreProperties( physicalDevice, pExternalSemaphoreInfo, pExternalSemaphoreProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceExternalSemaphorePropertiesKHR( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceExternalSemaphoreInfo* pExternalSemaphoreInfo, VkExternalSemaphoreProperties* pExternalSemaphoreProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceExternalSemaphorePropertiesKHR( physicalDevice, pExternalSemaphoreInfo, pExternalSemaphoreProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceFeatures( VkPhysicalDevice physicalDevice, VkPhysicalDeviceFeatures* pFeatures ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceFeatures( physicalDevice, pFeatures );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceFeatures2( VkPhysicalDevice physicalDevice, VkPhysicalDeviceFeatures2* pFeatures ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceFeatures2( physicalDevice, pFeatures );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceFeatures2KHR( VkPhysicalDevice physicalDevice, VkPhysicalDeviceFeatures2* pFeatures ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceFeatures2KHR( physicalDevice, pFeatures );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceFormatProperties( VkPhysicalDevice physicalDevice, VkFormat format, VkFormatProperties* pFormatProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceFormatProperties( physicalDevice, format, pFormatProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceFormatProperties2( VkPhysicalDevice physicalDevice, VkFormat format, VkFormatProperties2* pFormatProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceFormatProperties2( physicalDevice, format, pFormatProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceFormatProperties2KHR( VkPhysicalDevice physicalDevice, VkFormat format, VkFormatProperties2* pFormatProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceFormatProperties2KHR( physicalDevice, format, pFormatProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceGeneratedCommandsPropertiesNVX( VkPhysicalDevice physicalDevice, VkDeviceGeneratedCommandsFeaturesNVX* pFeatures, VkDeviceGeneratedCommandsLimitsNVX* pLimits ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceGeneratedCommandsPropertiesNVX( physicalDevice, pFeatures, pLimits );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceImageFormatProperties( VkPhysicalDevice physicalDevice, VkFormat format, VkImageType type, VkImageTiling tiling, VkImageUsageFlags usage, VkImageCreateFlags flags, VkImageFormatProperties* pImageFormatProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceImageFormatProperties( physicalDevice, format, type, tiling, usage, flags, pImageFormatProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceImageFormatProperties2( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceImageFormatInfo2* pImageFormatInfo, VkImageFormatProperties2* pImageFormatProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceImageFormatProperties2( physicalDevice, pImageFormatInfo, pImageFormatProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceImageFormatProperties2KHR( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceImageFormatInfo2* pImageFormatInfo, VkImageFormatProperties2* pImageFormatProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceImageFormatProperties2KHR( physicalDevice, pImageFormatInfo, pImageFormatProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceMemoryProperties( VkPhysicalDevice physicalDevice, VkPhysicalDeviceMemoryProperties* pMemoryProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceMemoryProperties( physicalDevice, pMemoryProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceMemoryProperties2( VkPhysicalDevice physicalDevice, VkPhysicalDeviceMemoryProperties2* pMemoryProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceMemoryProperties2( physicalDevice, pMemoryProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceMemoryProperties2KHR( VkPhysicalDevice physicalDevice, VkPhysicalDeviceMemoryProperties2* pMemoryProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceMemoryProperties2KHR( physicalDevice, pMemoryProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceMultisamplePropertiesEXT( VkPhysicalDevice physicalDevice, VkSampleCountFlagBits samples, VkMultisamplePropertiesEXT* pMultisampleProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceMultisamplePropertiesEXT( physicalDevice, samples, pMultisampleProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDevicePresentRectanglesKHR( VkPhysicalDevice physicalDevice, VkSurfaceKHR surface, uint32_t* pRectCount, VkRect2D* pRects ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDevicePresentRectanglesKHR( physicalDevice, surface, pRectCount, pRects );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceProperties( VkPhysicalDevice physicalDevice, VkPhysicalDeviceProperties* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceProperties( physicalDevice, pProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceProperties2( VkPhysicalDevice physicalDevice, VkPhysicalDeviceProperties2* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceProperties2( physicalDevice, pProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceProperties2KHR( VkPhysicalDevice physicalDevice, VkPhysicalDeviceProperties2* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceProperties2KHR( physicalDevice, pProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceQueueFamilyProperties( VkPhysicalDevice physicalDevice, uint32_t* pQueueFamilyPropertyCount, VkQueueFamilyProperties* pQueueFamilyProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceQueueFamilyProperties( physicalDevice, pQueueFamilyPropertyCount, pQueueFamilyProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceQueueFamilyProperties2( VkPhysicalDevice physicalDevice, uint32_t* pQueueFamilyPropertyCount, VkQueueFamilyProperties2* pQueueFamilyProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceQueueFamilyProperties2( physicalDevice, pQueueFamilyPropertyCount, pQueueFamilyProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceQueueFamilyProperties2KHR( VkPhysicalDevice physicalDevice, uint32_t* pQueueFamilyPropertyCount, VkQueueFamilyProperties2* pQueueFamilyProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceQueueFamilyProperties2KHR( physicalDevice, pQueueFamilyPropertyCount, pQueueFamilyProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceSparseImageFormatProperties( VkPhysicalDevice physicalDevice, VkFormat format, VkImageType type, VkSampleCountFlagBits samples, VkImageUsageFlags usage, VkImageTiling tiling, uint32_t* pPropertyCount, VkSparseImageFormatProperties* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSparseImageFormatProperties( physicalDevice, format, type, samples, usage, tiling, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceSparseImageFormatProperties2( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceSparseImageFormatInfo2* pFormatInfo, uint32_t* pPropertyCount, VkSparseImageFormatProperties2* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSparseImageFormatProperties2( physicalDevice, pFormatInfo, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
void vkGetPhysicalDeviceSparseImageFormatProperties2KHR( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceSparseImageFormatInfo2* pFormatInfo, uint32_t* pPropertyCount, VkSparseImageFormatProperties2* pProperties ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSparseImageFormatProperties2KHR( physicalDevice, pFormatInfo, pPropertyCount, pProperties );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV( VkPhysicalDevice physicalDevice, uint32_t* pCombinationCount, VkFramebufferMixedSamplesCombinationNV* pCombinations ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV( physicalDevice, pCombinationCount, pCombinations );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceSurfaceCapabilities2EXT( VkPhysicalDevice physicalDevice, VkSurfaceKHR surface, VkSurfaceCapabilities2EXT* pSurfaceCapabilities ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSurfaceCapabilities2EXT( physicalDevice, surface, pSurfaceCapabilities );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceSurfaceCapabilities2KHR( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceSurfaceInfo2KHR* pSurfaceInfo, VkSurfaceCapabilities2KHR* pSurfaceCapabilities ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSurfaceCapabilities2KHR( physicalDevice, pSurfaceInfo, pSurfaceCapabilities );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceSurfaceCapabilitiesKHR( VkPhysicalDevice physicalDevice, VkSurfaceKHR surface, VkSurfaceCapabilitiesKHR* pSurfaceCapabilities ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSurfaceCapabilitiesKHR( physicalDevice, surface, pSurfaceCapabilities );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceSurfaceFormats2KHR( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceSurfaceInfo2KHR* pSurfaceInfo, uint32_t* pSurfaceFormatCount, VkSurfaceFormat2KHR* pSurfaceFormats ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSurfaceFormats2KHR( physicalDevice, pSurfaceInfo, pSurfaceFormatCount, pSurfaceFormats );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceSurfaceFormatsKHR( VkPhysicalDevice physicalDevice, VkSurfaceKHR surface, uint32_t* pSurfaceFormatCount, VkSurfaceFormatKHR* pSurfaceFormats ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSurfaceFormatsKHR( physicalDevice, surface, pSurfaceFormatCount, pSurfaceFormats );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkResult vkGetPhysicalDeviceSurfacePresentModes2EXT( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceSurfaceInfo2KHR* pSurfaceInfo, uint32_t* pPresentModeCount, VkPresentModeKHR* pPresentModes ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSurfacePresentModes2EXT( physicalDevice, pSurfaceInfo, pPresentModeCount, pPresentModes );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceSurfacePresentModesKHR( VkPhysicalDevice physicalDevice, VkSurfaceKHR surface, uint32_t* pPresentModeCount, VkPresentModeKHR* pPresentModes ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSurfacePresentModesKHR( physicalDevice, surface, pPresentModeCount, pPresentModes );
|
||
|
}
|
||
|
|
||
|
VkResult vkGetPhysicalDeviceSurfaceSupportKHR( VkPhysicalDevice physicalDevice, uint32_t queueFamilyIndex, VkSurfaceKHR surface, VkBool32* pSupported ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceSurfaceSupportKHR( physicalDevice, queueFamilyIndex, surface, pSupported );
|
||
|
}
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WAYLAND_KHR
|
||
|
VkBool32 vkGetPhysicalDeviceWaylandPresentationSupportKHR( VkPhysicalDevice physicalDevice, uint32_t queueFamilyIndex, struct wl_display* display ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceWaylandPresentationSupportKHR( physicalDevice, queueFamilyIndex, display );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WAYLAND_KHR*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_WIN32_KHR
|
||
|
VkBool32 vkGetPhysicalDeviceWin32PresentationSupportKHR( VkPhysicalDevice physicalDevice, uint32_t queueFamilyIndex ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceWin32PresentationSupportKHR( physicalDevice, queueFamilyIndex );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_WIN32_KHR*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_XCB_KHR
|
||
|
VkBool32 vkGetPhysicalDeviceXcbPresentationSupportKHR( VkPhysicalDevice physicalDevice, uint32_t queueFamilyIndex, xcb_connection_t* connection, xcb_visualid_t visual_id ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceXcbPresentationSupportKHR( physicalDevice, queueFamilyIndex, connection, visual_id );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_XCB_KHR*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_XLIB_KHR
|
||
|
VkBool32 vkGetPhysicalDeviceXlibPresentationSupportKHR( VkPhysicalDevice physicalDevice, uint32_t queueFamilyIndex, Display* dpy, VisualID visualID ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetPhysicalDeviceXlibPresentationSupportKHR( physicalDevice, queueFamilyIndex, dpy, visualID );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_XLIB_KHR*/
|
||
|
|
||
|
#ifdef VK_USE_PLATFORM_XLIB_XRANDR_EXT
|
||
|
VkResult vkGetRandROutputDisplayEXT( VkPhysicalDevice physicalDevice, Display* dpy, RROutput rrOutput, VkDisplayKHR* pDisplay ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetRandROutputDisplayEXT( physicalDevice, dpy, rrOutput, pDisplay );
|
||
|
}
|
||
|
#endif /*VK_USE_PLATFORM_XLIB_XRANDR_EXT*/
|
||
|
|
||
|
VkResult vkReleaseDisplayEXT( VkPhysicalDevice physicalDevice, VkDisplayKHR display ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkReleaseDisplayEXT( physicalDevice, display );
|
||
|
}
|
||
|
|
||
|
void vkGetQueueCheckpointDataNV( VkQueue queue, uint32_t* pCheckpointDataCount, VkCheckpointDataNV* pCheckpointData ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkGetQueueCheckpointDataNV( queue, pCheckpointDataCount, pCheckpointData );
|
||
|
}
|
||
|
|
||
|
void vkQueueBeginDebugUtilsLabelEXT( VkQueue queue, const VkDebugUtilsLabelEXT* pLabelInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkQueueBeginDebugUtilsLabelEXT( queue, pLabelInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkQueueBindSparse( VkQueue queue, uint32_t bindInfoCount, const VkBindSparseInfo* pBindInfo, VkFence fence ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkQueueBindSparse( queue, bindInfoCount, pBindInfo, fence );
|
||
|
}
|
||
|
|
||
|
void vkQueueEndDebugUtilsLabelEXT( VkQueue queue ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkQueueEndDebugUtilsLabelEXT( queue );
|
||
|
}
|
||
|
|
||
|
void vkQueueInsertDebugUtilsLabelEXT( VkQueue queue, const VkDebugUtilsLabelEXT* pLabelInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkQueueInsertDebugUtilsLabelEXT( queue, pLabelInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkQueuePresentKHR( VkQueue queue, const VkPresentInfoKHR* pPresentInfo ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkQueuePresentKHR( queue, pPresentInfo );
|
||
|
}
|
||
|
|
||
|
VkResult vkQueueSetPerformanceConfigurationINTEL( VkQueue queue, VkPerformanceConfigurationINTEL configuration ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkQueueSetPerformanceConfigurationINTEL( queue, configuration );
|
||
|
}
|
||
|
|
||
|
VkResult vkQueueSubmit( VkQueue queue, uint32_t submitCount, const VkSubmitInfo* pSubmits, VkFence fence ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkQueueSubmit( queue, submitCount, pSubmits, fence );
|
||
|
}
|
||
|
|
||
|
VkResult vkQueueWaitIdle( VkQueue queue ) const VULKAN_HPP_NOEXCEPT
|
||
|
{
|
||
|
return ::vkQueueWaitIdle( queue );
|
||
|
}
|
||
|
};
|
||
|
#endif
|
||
|
} // namespace VULKAN_HPP_NAMESPACE
|