8 #ifndef VULKAN_FUNCS_HPP
9 #define VULKAN_FUNCS_HPP
11 namespace VULKAN_HPP_NAMESPACE
20 template <
typename Dispatch>
22 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
23 VULKAN_HPP_NAMESPACE::Instance * pInstance,
29 reinterpret_cast<VkInstance *
>( pInstance ) ) );
32 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
33 template <
typename Dispatch>
38 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
39 VULKAN_HPP_ASSERT( d.vkCreateInstance &&
"Function <vkCreateInstance> requires <VK_VERSION_1_0>" );
42 VULKAN_HPP_NAMESPACE::Instance instance;
45 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
46 reinterpret_cast<VkInstance *
>( &instance ) ) );
52 # ifndef VULKAN_HPP_NO_SMART_HANDLE
53 template <
typename Dispatch>
58 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
59 VULKAN_HPP_ASSERT( d.vkCreateInstance &&
"Function <vkCreateInstance> requires <VK_VERSION_1_0>" );
62 VULKAN_HPP_NAMESPACE::Instance instance;
65 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
66 reinterpret_cast<VkInstance *
>( &instance ) ) );
75 template <
typename Dispatch>
82 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
83 template <
typename Dispatch>
87 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
88 VULKAN_HPP_ASSERT( d.vkDestroyInstance &&
"Function <vkDestroyInstance> requires <VK_VERSION_1_0>" );
91 d.vkDestroyInstance( m_instance,
92 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
96 template <
typename Dispatch>
98 VULKAN_HPP_NAMESPACE::PhysicalDevice * pPhysicalDevices,
102 return static_cast<Result>( d.vkEnumeratePhysicalDevices( m_instance, pPhysicalDeviceCount,
reinterpret_cast<VkPhysicalDevice *
>( pPhysicalDevices ) ) );
105 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
106 template <
typename PhysicalDeviceAllocator,
typename Dispatch>
108 Instance::enumeratePhysicalDevices( Dispatch
const & d )
const
111 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
112 VULKAN_HPP_ASSERT( d.vkEnumeratePhysicalDevices &&
"Function <vkEnumeratePhysicalDevices> requires <VK_VERSION_1_0>" );
115 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDevice, PhysicalDeviceAllocator> physicalDevices;
116 uint32_t physicalDeviceCount;
121 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && physicalDeviceCount )
123 physicalDevices.resize( physicalDeviceCount );
125 d.vkEnumeratePhysicalDevices( m_instance, &physicalDeviceCount,
reinterpret_cast<VkPhysicalDevice *
>( physicalDevices.data() ) ) );
127 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
130 if ( physicalDeviceCount < physicalDevices.size() )
132 physicalDevices.resize( physicalDeviceCount );
137 template <
typename PhysicalDeviceAllocator,
139 typename std::enable_if<std::is_same<typename PhysicalDeviceAllocator::value_type, VULKAN_HPP_NAMESPACE::PhysicalDevice>::value,
int>::type>
141 Instance::enumeratePhysicalDevices( PhysicalDeviceAllocator & physicalDeviceAllocator, Dispatch
const & d )
const
144 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
145 VULKAN_HPP_ASSERT( d.vkEnumeratePhysicalDevices &&
"Function <vkEnumeratePhysicalDevices> requires <VK_VERSION_1_0>" );
148 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDevice, PhysicalDeviceAllocator> physicalDevices( physicalDeviceAllocator );
149 uint32_t physicalDeviceCount;
154 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && physicalDeviceCount )
156 physicalDevices.resize( physicalDeviceCount );
158 d.vkEnumeratePhysicalDevices( m_instance, &physicalDeviceCount,
reinterpret_cast<VkPhysicalDevice *
>( physicalDevices.data() ) ) );
160 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
163 if ( physicalDeviceCount < physicalDevices.size() )
165 physicalDevices.resize( physicalDeviceCount );
171 template <
typename Dispatch>
178 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
179 template <
typename Dispatch>
184 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
185 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceFeatures &&
"Function <vkGetPhysicalDeviceFeatures> requires <VK_VERSION_1_0>" );
188 VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures features;
195 template <
typename Dispatch>
197 VULKAN_HPP_NAMESPACE::FormatProperties * pFormatProperties,
201 d.vkGetPhysicalDeviceFormatProperties( m_physicalDevice,
static_cast<VkFormat>( format ),
reinterpret_cast<VkFormatProperties *
>( pFormatProperties ) );
204 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
205 template <
typename Dispatch>
210 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
211 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceFormatProperties &&
"Function <vkGetPhysicalDeviceFormatProperties> requires <VK_VERSION_1_0>" );
214 VULKAN_HPP_NAMESPACE::FormatProperties formatProperties;
215 d.vkGetPhysicalDeviceFormatProperties( m_physicalDevice,
static_cast<VkFormat>( format ),
reinterpret_cast<VkFormatProperties *
>( &formatProperties ) );
217 return formatProperties;
221 template <
typename Dispatch>
227 VULKAN_HPP_NAMESPACE::ImageFormatProperties * pImageFormatProperties,
231 return static_cast<Result>( d.vkGetPhysicalDeviceImageFormatProperties( m_physicalDevice,
240 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
241 template <
typename Dispatch>
248 Dispatch
const & d )
const
251 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
252 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceImageFormatProperties &&
"Function <vkGetPhysicalDeviceImageFormatProperties> requires <VK_VERSION_1_0>" );
255 VULKAN_HPP_NAMESPACE::ImageFormatProperties imageFormatProperties;
257 d.vkGetPhysicalDeviceImageFormatProperties( m_physicalDevice,
270 template <
typename Dispatch>
271 VULKAN_HPP_INLINE void PhysicalDevice::getProperties( VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties * pProperties,
278 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
279 template <
typename Dispatch>
284 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
285 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceProperties &&
"Function <vkGetPhysicalDeviceProperties> requires <VK_VERSION_1_0>" );
288 VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties properties;
295 template <
typename Dispatch>
296 VULKAN_HPP_INLINE void PhysicalDevice::getQueueFamilyProperties( uint32_t * pQueueFamilyPropertyCount,
297 VULKAN_HPP_NAMESPACE::QueueFamilyProperties * pQueueFamilyProperties,
301 d.vkGetPhysicalDeviceQueueFamilyProperties(
302 m_physicalDevice, pQueueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties *
>( pQueueFamilyProperties ) );
305 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
306 template <
typename QueueFamilyPropertiesAllocator,
typename Dispatch>
308 PhysicalDevice::getQueueFamilyProperties( Dispatch
const & d )
const
311 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
312 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceQueueFamilyProperties &&
"Function <vkGetPhysicalDeviceQueueFamilyProperties> requires <VK_VERSION_1_0>" );
315 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties, QueueFamilyPropertiesAllocator> queueFamilyProperties;
316 uint32_t queueFamilyPropertyCount;
317 d.vkGetPhysicalDeviceQueueFamilyProperties( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
318 queueFamilyProperties.resize( queueFamilyPropertyCount );
319 d.vkGetPhysicalDeviceQueueFamilyProperties(
320 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties *
>( queueFamilyProperties.data() ) );
323 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
325 queueFamilyProperties.resize( queueFamilyPropertyCount );
327 return queueFamilyProperties;
331 typename QueueFamilyPropertiesAllocator,
333 typename std::enable_if<std::is_same<typename QueueFamilyPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::QueueFamilyProperties>::value,
int>::type>
335 PhysicalDevice::getQueueFamilyProperties( QueueFamilyPropertiesAllocator & queueFamilyPropertiesAllocator, Dispatch
const & d )
const
338 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
339 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceQueueFamilyProperties &&
"Function <vkGetPhysicalDeviceQueueFamilyProperties> requires <VK_VERSION_1_0>" );
342 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties, QueueFamilyPropertiesAllocator> queueFamilyProperties( queueFamilyPropertiesAllocator );
343 uint32_t queueFamilyPropertyCount;
344 d.vkGetPhysicalDeviceQueueFamilyProperties( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
345 queueFamilyProperties.resize( queueFamilyPropertyCount );
346 d.vkGetPhysicalDeviceQueueFamilyProperties(
347 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties *
>( queueFamilyProperties.data() ) );
350 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
352 queueFamilyProperties.resize( queueFamilyPropertyCount );
354 return queueFamilyProperties;
358 template <
typename Dispatch>
359 VULKAN_HPP_INLINE void PhysicalDevice::getMemoryProperties( VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties * pMemoryProperties,
366 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
367 template <
typename Dispatch>
372 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
373 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceMemoryProperties &&
"Function <vkGetPhysicalDeviceMemoryProperties> requires <VK_VERSION_1_0>" );
376 VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties memoryProperties;
379 return memoryProperties;
383 template <
typename Dispatch>
387 return d.vkGetInstanceProcAddr( m_instance, pName );
390 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
391 template <
typename Dispatch>
395 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
396 VULKAN_HPP_ASSERT( d.vkGetInstanceProcAddr &&
"Function <vkGetInstanceProcAddr> requires <VK_VERSION_1_0>" );
405 template <
typename Dispatch>
409 return d.vkGetDeviceProcAddr( m_device, pName );
412 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
413 template <
typename Dispatch>
417 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
418 VULKAN_HPP_ASSERT( d.vkGetDeviceProcAddr &&
"Function <vkGetDeviceProcAddr> requires <VK_VERSION_1_0>" );
427 template <
typename Dispatch>
429 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
430 VULKAN_HPP_NAMESPACE::Device * pDevice,
434 return static_cast<Result>( d.vkCreateDevice( m_physicalDevice,
437 reinterpret_cast<VkDevice *
>( pDevice ) ) );
440 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
441 template <
typename Dispatch>
446 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
447 VULKAN_HPP_ASSERT( d.vkCreateDevice &&
"Function <vkCreateDevice> requires <VK_VERSION_1_0>" );
450 VULKAN_HPP_NAMESPACE::Device device;
452 d.vkCreateDevice( m_physicalDevice,
454 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
455 reinterpret_cast<VkDevice *
>( &device ) ) );
461 # ifndef VULKAN_HPP_NO_SMART_HANDLE
462 template <
typename Dispatch>
464 PhysicalDevice::createDeviceUnique(
const VULKAN_HPP_NAMESPACE::DeviceCreateInfo & createInfo,
466 Dispatch
const & d )
const
469 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
470 VULKAN_HPP_ASSERT( d.vkCreateDevice &&
"Function <vkCreateDevice> requires <VK_VERSION_1_0>" );
473 VULKAN_HPP_NAMESPACE::Device device;
475 d.vkCreateDevice( m_physicalDevice,
477 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
478 reinterpret_cast<VkDevice *
>( &device ) ) );
486 template <
typename Dispatch>
493 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
494 template <
typename Dispatch>
498 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
499 VULKAN_HPP_ASSERT( d.vkDestroyDevice &&
"Function <vkDestroyDevice> requires <VK_VERSION_1_0>" );
502 d.vkDestroyDevice( m_device,
503 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
507 template <
typename Dispatch>
509 uint32_t * pPropertyCount,
510 VULKAN_HPP_NAMESPACE::ExtensionProperties * pProperties,
514 return static_cast<Result>(
515 d.vkEnumerateInstanceExtensionProperties( pLayerName, pPropertyCount,
reinterpret_cast<VkExtensionProperties *
>( pProperties ) ) );
518 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
519 template <
typename ExtensionPropertiesAllocator,
typename Dispatch>
524 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
525 VULKAN_HPP_ASSERT( d.vkEnumerateInstanceExtensionProperties &&
"Function <vkEnumerateInstanceExtensionProperties> requires <VK_VERSION_1_0>" );
528 std::vector<VULKAN_HPP_NAMESPACE::ExtensionProperties, ExtensionPropertiesAllocator> properties;
529 uint32_t propertyCount;
534 d.vkEnumerateInstanceExtensionProperties( layerName ? layerName->c_str() :
nullptr, &propertyCount,
nullptr ) );
535 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
537 properties.resize( propertyCount );
539 layerName ? layerName->c_str() :
nullptr, &propertyCount,
reinterpret_cast<VkExtensionProperties *
>( properties.data() ) ) );
541 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
544 if ( propertyCount < properties.size() )
546 properties.resize( propertyCount );
552 typename ExtensionPropertiesAllocator,
554 typename std::enable_if<std::is_same<typename ExtensionPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::ExtensionProperties>::value,
int>::type>
557 ExtensionPropertiesAllocator & extensionPropertiesAllocator,
561 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
562 VULKAN_HPP_ASSERT( d.vkEnumerateInstanceExtensionProperties &&
"Function <vkEnumerateInstanceExtensionProperties> requires <VK_VERSION_1_0>" );
565 std::vector<VULKAN_HPP_NAMESPACE::ExtensionProperties, ExtensionPropertiesAllocator> properties( extensionPropertiesAllocator );
566 uint32_t propertyCount;
571 d.vkEnumerateInstanceExtensionProperties( layerName ? layerName->c_str() :
nullptr, &propertyCount,
nullptr ) );
572 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
574 properties.resize( propertyCount );
576 layerName ? layerName->c_str() :
nullptr, &propertyCount,
reinterpret_cast<VkExtensionProperties *
>( properties.data() ) ) );
578 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
581 if ( propertyCount < properties.size() )
583 properties.resize( propertyCount );
589 template <
typename Dispatch>
591 uint32_t * pPropertyCount,
592 VULKAN_HPP_NAMESPACE::ExtensionProperties * pProperties,
596 return static_cast<Result>(
597 d.vkEnumerateDeviceExtensionProperties( m_physicalDevice, pLayerName, pPropertyCount,
reinterpret_cast<VkExtensionProperties *
>( pProperties ) ) );
600 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
601 template <
typename ExtensionPropertiesAllocator,
typename Dispatch>
606 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
607 VULKAN_HPP_ASSERT( d.vkEnumerateDeviceExtensionProperties &&
"Function <vkEnumerateDeviceExtensionProperties> requires <VK_VERSION_1_0>" );
610 std::vector<VULKAN_HPP_NAMESPACE::ExtensionProperties, ExtensionPropertiesAllocator> properties;
611 uint32_t propertyCount;
616 d.vkEnumerateDeviceExtensionProperties( m_physicalDevice, layerName ? layerName->c_str() :
nullptr, &propertyCount,
nullptr ) );
617 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
619 properties.resize( propertyCount );
621 m_physicalDevice, layerName ? layerName->c_str() :
nullptr, &propertyCount,
reinterpret_cast<VkExtensionProperties *
>( properties.data() ) ) );
623 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
626 if ( propertyCount < properties.size() )
628 properties.resize( propertyCount );
634 typename ExtensionPropertiesAllocator,
636 typename std::enable_if<std::is_same<typename ExtensionPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::ExtensionProperties>::value,
int>::type>
639 ExtensionPropertiesAllocator & extensionPropertiesAllocator,
640 Dispatch
const & d )
const
643 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
644 VULKAN_HPP_ASSERT( d.vkEnumerateDeviceExtensionProperties &&
"Function <vkEnumerateDeviceExtensionProperties> requires <VK_VERSION_1_0>" );
647 std::vector<VULKAN_HPP_NAMESPACE::ExtensionProperties, ExtensionPropertiesAllocator> properties( extensionPropertiesAllocator );
648 uint32_t propertyCount;
653 d.vkEnumerateDeviceExtensionProperties( m_physicalDevice, layerName ? layerName->c_str() :
nullptr, &propertyCount,
nullptr ) );
654 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
656 properties.resize( propertyCount );
658 m_physicalDevice, layerName ? layerName->c_str() :
nullptr, &propertyCount,
reinterpret_cast<VkExtensionProperties *
>( properties.data() ) ) );
660 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
663 if ( propertyCount < properties.size() )
665 properties.resize( propertyCount );
671 template <
typename Dispatch>
673 VULKAN_HPP_NAMESPACE::LayerProperties * pProperties,
677 return static_cast<Result>( d.vkEnumerateInstanceLayerProperties( pPropertyCount,
reinterpret_cast<VkLayerProperties *
>( pProperties ) ) );
680 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
681 template <
typename LayerPropertiesAllocator,
typename Dispatch>
686 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
687 VULKAN_HPP_ASSERT( d.vkEnumerateInstanceLayerProperties &&
"Function <vkEnumerateInstanceLayerProperties> requires <VK_VERSION_1_0>" );
690 std::vector<VULKAN_HPP_NAMESPACE::LayerProperties, LayerPropertiesAllocator> properties;
691 uint32_t propertyCount;
696 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
698 properties.resize( propertyCount );
700 d.vkEnumerateInstanceLayerProperties( &propertyCount,
reinterpret_cast<VkLayerProperties *
>( properties.data() ) ) );
702 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
705 if ( propertyCount < properties.size() )
707 properties.resize( propertyCount );
712 template <
typename LayerPropertiesAllocator,
714 typename std::enable_if<std::is_same<typename LayerPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::LayerProperties>::value,
int>::type>
719 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
720 VULKAN_HPP_ASSERT( d.vkEnumerateInstanceLayerProperties &&
"Function <vkEnumerateInstanceLayerProperties> requires <VK_VERSION_1_0>" );
723 std::vector<VULKAN_HPP_NAMESPACE::LayerProperties, LayerPropertiesAllocator> properties( layerPropertiesAllocator );
724 uint32_t propertyCount;
729 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
731 properties.resize( propertyCount );
733 d.vkEnumerateInstanceLayerProperties( &propertyCount,
reinterpret_cast<VkLayerProperties *
>( properties.data() ) ) );
735 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
738 if ( propertyCount < properties.size() )
740 properties.resize( propertyCount );
746 template <
typename Dispatch>
748 VULKAN_HPP_NAMESPACE::LayerProperties * pProperties,
752 return static_cast<Result>( d.vkEnumerateDeviceLayerProperties( m_physicalDevice, pPropertyCount,
reinterpret_cast<VkLayerProperties *
>( pProperties ) ) );
755 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
756 template <
typename LayerPropertiesAllocator,
typename Dispatch>
758 PhysicalDevice::enumerateDeviceLayerProperties( Dispatch
const & d )
const
761 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
762 VULKAN_HPP_ASSERT( d.vkEnumerateDeviceLayerProperties &&
"Function <vkEnumerateDeviceLayerProperties> requires <VK_VERSION_1_0>" );
765 std::vector<VULKAN_HPP_NAMESPACE::LayerProperties, LayerPropertiesAllocator> properties;
766 uint32_t propertyCount;
771 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
773 properties.resize( propertyCount );
775 d.vkEnumerateDeviceLayerProperties( m_physicalDevice, &propertyCount,
reinterpret_cast<VkLayerProperties *
>( properties.data() ) ) );
777 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
780 if ( propertyCount < properties.size() )
782 properties.resize( propertyCount );
787 template <
typename LayerPropertiesAllocator,
789 typename std::enable_if<std::is_same<typename LayerPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::LayerProperties>::value,
int>::type>
791 PhysicalDevice::enumerateDeviceLayerProperties( LayerPropertiesAllocator & layerPropertiesAllocator, Dispatch
const & d )
const
794 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
795 VULKAN_HPP_ASSERT( d.vkEnumerateDeviceLayerProperties &&
"Function <vkEnumerateDeviceLayerProperties> requires <VK_VERSION_1_0>" );
798 std::vector<VULKAN_HPP_NAMESPACE::LayerProperties, LayerPropertiesAllocator> properties( layerPropertiesAllocator );
799 uint32_t propertyCount;
804 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
806 properties.resize( propertyCount );
808 d.vkEnumerateDeviceLayerProperties( m_physicalDevice, &propertyCount,
reinterpret_cast<VkLayerProperties *
>( properties.data() ) ) );
810 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
813 if ( propertyCount < properties.size() )
815 properties.resize( propertyCount );
821 template <
typename Dispatch>
823 Device::getQueue( uint32_t queueFamilyIndex, uint32_t queueIndex, VULKAN_HPP_NAMESPACE::Queue * pQueue, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
826 d.vkGetDeviceQueue( m_device, queueFamilyIndex, queueIndex,
reinterpret_cast<VkQueue *
>( pQueue ) );
829 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
830 template <
typename Dispatch>
832 Device::getQueue( uint32_t queueFamilyIndex, uint32_t queueIndex, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
835 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
836 VULKAN_HPP_ASSERT( d.vkGetDeviceQueue &&
"Function <vkGetDeviceQueue> requires <VK_VERSION_1_0>" );
839 VULKAN_HPP_NAMESPACE::Queue queue;
840 d.vkGetDeviceQueue( m_device, queueFamilyIndex, queueIndex,
reinterpret_cast<VkQueue *
>( &queue ) );
846 template <
typename Dispatch>
848 const VULKAN_HPP_NAMESPACE::SubmitInfo * pSubmits,
849 VULKAN_HPP_NAMESPACE::Fence fence,
853 return static_cast<Result>( d.vkQueueSubmit( m_queue, submitCount,
reinterpret_cast<const VkSubmitInfo *
>( pSubmits ),
static_cast<VkFence>( fence ) ) );
856 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
857 template <
typename Dispatch>
859 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SubmitInfo>
const & submits, VULKAN_HPP_NAMESPACE::Fence fence, Dispatch
const & d )
const
862 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
863 VULKAN_HPP_ASSERT( d.vkQueueSubmit &&
"Function <vkQueueSubmit> requires <VK_VERSION_1_0>" );
867 d.vkQueueSubmit( m_queue, submits.size(),
reinterpret_cast<const VkSubmitInfo *
>( submits.data() ),
static_cast<VkFence>( fence ) ) );
874 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
875 template <
typename Dispatch>
879 return static_cast<Result>( d.vkQueueWaitIdle( m_queue ) );
882 template <
typename Dispatch>
886 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
887 VULKAN_HPP_ASSERT( d.vkQueueWaitIdle &&
"Function <vkQueueWaitIdle> requires <VK_VERSION_1_0>" );
897 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
898 template <
typename Dispatch>
902 return static_cast<Result>( d.vkDeviceWaitIdle( m_device ) );
905 template <
typename Dispatch>
909 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
910 VULKAN_HPP_ASSERT( d.vkDeviceWaitIdle &&
"Function <vkDeviceWaitIdle> requires <VK_VERSION_1_0>" );
920 template <
typename Dispatch>
922 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
923 VULKAN_HPP_NAMESPACE::DeviceMemory * pMemory,
927 return static_cast<Result>( d.vkAllocateMemory( m_device,
933 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
934 template <
typename Dispatch>
936 Device::allocateMemory(
const VULKAN_HPP_NAMESPACE::MemoryAllocateInfo & allocateInfo,
938 Dispatch
const & d )
const
941 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
942 VULKAN_HPP_ASSERT( d.vkAllocateMemory &&
"Function <vkAllocateMemory> requires <VK_VERSION_1_0>" );
945 VULKAN_HPP_NAMESPACE::DeviceMemory memory;
947 d.vkAllocateMemory( m_device,
949 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
956 # ifndef VULKAN_HPP_NO_SMART_HANDLE
957 template <
typename Dispatch>
959 Device::allocateMemoryUnique(
const VULKAN_HPP_NAMESPACE::MemoryAllocateInfo & allocateInfo,
961 Dispatch
const & d )
const
964 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
965 VULKAN_HPP_ASSERT( d.vkAllocateMemory &&
"Function <vkAllocateMemory> requires <VK_VERSION_1_0>" );
968 VULKAN_HPP_NAMESPACE::DeviceMemory memory;
970 d.vkAllocateMemory( m_device,
972 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
982 template <
typename Dispatch>
984 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
991 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
992 template <
typename Dispatch>
998 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
999 VULKAN_HPP_ASSERT( d.vkFreeMemory &&
"Function <vkFreeMemory> requires <VK_VERSION_1_0>" );
1002 d.vkFreeMemory( m_device,
1004 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
1008 template <
typename Dispatch>
1009 VULKAN_HPP_INLINE void( Device::free )( VULKAN_HPP_NAMESPACE::DeviceMemory memory,
1010 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1017 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1018 template <
typename Dispatch>
1019 VULKAN_HPP_INLINE void( Device::free )( VULKAN_HPP_NAMESPACE::DeviceMemory memory,
1020 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
1024 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1025 VULKAN_HPP_ASSERT( d.vkFreeMemory &&
"Function <vkFreeMemory> requires <VK_VERSION_1_0>" );
1028 d.vkFreeMemory( m_device,
1030 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
1034 template <
typename Dispatch>
1043 return static_cast<Result>( d.vkMapMemory( m_device,
1051 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1052 template <
typename Dispatch>
1057 Dispatch
const & d )
const
1060 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1061 VULKAN_HPP_ASSERT( d.vkMapMemory &&
"Function <vkMapMemory> requires <VK_VERSION_1_0>" );
1077 template <
typename Dispatch>
1081 d.vkUnmapMemory( m_device,
static_cast<VkDeviceMemory>( memory ) );
1084 template <
typename Dispatch>
1086 const VULKAN_HPP_NAMESPACE::MappedMemoryRange * pMemoryRanges,
1090 return static_cast<Result>( d.vkFlushMappedMemoryRanges( m_device, memoryRangeCount,
reinterpret_cast<const VkMappedMemoryRange *
>( pMemoryRanges ) ) );
1093 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1094 template <
typename Dispatch>
1096 Device::flushMappedMemoryRanges( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::MappedMemoryRange>
const & memoryRanges,
1097 Dispatch
const & d )
const
1100 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1101 VULKAN_HPP_ASSERT( d.vkFlushMappedMemoryRanges &&
"Function <vkFlushMappedMemoryRanges> requires <VK_VERSION_1_0>" );
1105 d.vkFlushMappedMemoryRanges( m_device, memoryRanges.size(),
reinterpret_cast<const VkMappedMemoryRange *
>( memoryRanges.data() ) ) );
1112 template <
typename Dispatch>
1114 const VULKAN_HPP_NAMESPACE::MappedMemoryRange * pMemoryRanges,
1118 return static_cast<Result>(
1119 d.vkInvalidateMappedMemoryRanges( m_device, memoryRangeCount,
reinterpret_cast<const VkMappedMemoryRange *
>( pMemoryRanges ) ) );
1122 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1123 template <
typename Dispatch>
1125 Device::invalidateMappedMemoryRanges( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::MappedMemoryRange>
const & memoryRanges,
1126 Dispatch
const & d )
const
1129 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1130 VULKAN_HPP_ASSERT( d.vkInvalidateMappedMemoryRanges &&
"Function <vkInvalidateMappedMemoryRanges> requires <VK_VERSION_1_0>" );
1134 d.vkInvalidateMappedMemoryRanges( m_device, memoryRanges.size(),
reinterpret_cast<const VkMappedMemoryRange *
>( memoryRanges.data() ) ) );
1141 template <
typename Dispatch>
1147 d.vkGetDeviceMemoryCommitment( m_device,
static_cast<VkDeviceMemory>( memory ),
reinterpret_cast<VkDeviceSize *
>( pCommittedMemoryInBytes ) );
1150 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1151 template <
typename Dispatch>
1156 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1157 VULKAN_HPP_ASSERT( d.vkGetDeviceMemoryCommitment &&
"Function <vkGetDeviceMemoryCommitment> requires <VK_VERSION_1_0>" );
1161 d.vkGetDeviceMemoryCommitment( m_device,
static_cast<VkDeviceMemory>( memory ),
reinterpret_cast<VkDeviceSize *
>( &committedMemoryInBytes ) );
1163 return committedMemoryInBytes;
1167 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
1168 template <
typename Dispatch>
1170 VULKAN_HPP_NAMESPACE::DeviceMemory memory,
1175 return static_cast<Result>(
1179 template <
typename Dispatch>
1184 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1185 VULKAN_HPP_ASSERT( d.vkBindBufferMemory &&
"Function <vkBindBufferMemory> requires <VK_VERSION_1_0>" );
1196 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
1197 template <
typename Dispatch>
1199 VULKAN_HPP_NAMESPACE::DeviceMemory memory,
1204 return static_cast<Result>(
1208 template <
typename Dispatch>
1213 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1214 VULKAN_HPP_ASSERT( d.vkBindImageMemory &&
"Function <vkBindImageMemory> requires <VK_VERSION_1_0>" );
1225 template <
typename Dispatch>
1227 VULKAN_HPP_NAMESPACE::MemoryRequirements * pMemoryRequirements,
1231 d.vkGetBufferMemoryRequirements( m_device,
static_cast<VkBuffer>( buffer ),
reinterpret_cast<VkMemoryRequirements *
>( pMemoryRequirements ) );
1234 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1235 template <
typename Dispatch>
1237 Device::getBufferMemoryRequirements( VULKAN_HPP_NAMESPACE::Buffer buffer, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
1240 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1241 VULKAN_HPP_ASSERT( d.vkGetBufferMemoryRequirements &&
"Function <vkGetBufferMemoryRequirements> requires <VK_VERSION_1_0>" );
1244 VULKAN_HPP_NAMESPACE::MemoryRequirements memoryRequirements;
1245 d.vkGetBufferMemoryRequirements( m_device,
static_cast<VkBuffer>( buffer ),
reinterpret_cast<VkMemoryRequirements *
>( &memoryRequirements ) );
1247 return memoryRequirements;
1251 template <
typename Dispatch>
1253 VULKAN_HPP_NAMESPACE::MemoryRequirements * pMemoryRequirements,
1257 d.vkGetImageMemoryRequirements( m_device,
static_cast<VkImage>( image ),
reinterpret_cast<VkMemoryRequirements *
>( pMemoryRequirements ) );
1260 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1261 template <
typename Dispatch>
1263 Device::getImageMemoryRequirements( VULKAN_HPP_NAMESPACE::Image image, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
1266 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1267 VULKAN_HPP_ASSERT( d.vkGetImageMemoryRequirements &&
"Function <vkGetImageMemoryRequirements> requires <VK_VERSION_1_0>" );
1270 VULKAN_HPP_NAMESPACE::MemoryRequirements memoryRequirements;
1271 d.vkGetImageMemoryRequirements( m_device,
static_cast<VkImage>( image ),
reinterpret_cast<VkMemoryRequirements *
>( &memoryRequirements ) );
1273 return memoryRequirements;
1277 template <
typename Dispatch>
1279 uint32_t * pSparseMemoryRequirementCount,
1280 VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements * pSparseMemoryRequirements,
1284 d.vkGetImageSparseMemoryRequirements( m_device,
1285 static_cast<VkImage>( image ),
1286 pSparseMemoryRequirementCount,
1290 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1291 template <
typename SparseImageMemoryRequirementsAllocator,
typename Dispatch>
1293 Device::getImageSparseMemoryRequirements( VULKAN_HPP_NAMESPACE::Image image, Dispatch
const & d )
const
1296 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1297 VULKAN_HPP_ASSERT( d.vkGetImageSparseMemoryRequirements &&
"Function <vkGetImageSparseMemoryRequirements> requires <VK_VERSION_1_0>" );
1300 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements, SparseImageMemoryRequirementsAllocator> sparseMemoryRequirements;
1301 uint32_t sparseMemoryRequirementCount;
1302 d.vkGetImageSparseMemoryRequirements( m_device,
static_cast<VkImage>( image ), &sparseMemoryRequirementCount,
nullptr );
1303 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
1304 d.vkGetImageSparseMemoryRequirements( m_device,
1305 static_cast<VkImage>( image ),
1306 &sparseMemoryRequirementCount,
1309 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
1310 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
1312 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
1314 return sparseMemoryRequirements;
1317 template <
typename SparseImageMemoryRequirementsAllocator,
1319 typename std::enable_if<
1320 std::is_same<typename SparseImageMemoryRequirementsAllocator::value_type, VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements>::value,
1323 Device::getImageSparseMemoryRequirements( VULKAN_HPP_NAMESPACE::Image image,
1324 SparseImageMemoryRequirementsAllocator & sparseImageMemoryRequirementsAllocator,
1325 Dispatch
const & d )
const
1328 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1329 VULKAN_HPP_ASSERT( d.vkGetImageSparseMemoryRequirements &&
"Function <vkGetImageSparseMemoryRequirements> requires <VK_VERSION_1_0>" );
1332 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements, SparseImageMemoryRequirementsAllocator> sparseMemoryRequirements(
1333 sparseImageMemoryRequirementsAllocator );
1334 uint32_t sparseMemoryRequirementCount;
1335 d.vkGetImageSparseMemoryRequirements( m_device,
static_cast<VkImage>( image ), &sparseMemoryRequirementCount,
nullptr );
1336 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
1337 d.vkGetImageSparseMemoryRequirements( m_device,
1338 static_cast<VkImage>( image ),
1339 &sparseMemoryRequirementCount,
1342 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
1343 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
1345 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
1347 return sparseMemoryRequirements;
1351 template <
typename Dispatch>
1357 uint32_t * pPropertyCount,
1358 VULKAN_HPP_NAMESPACE::SparseImageFormatProperties * pProperties,
1362 d.vkGetPhysicalDeviceSparseImageFormatProperties( m_physicalDevice,
1372 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1373 template <
typename SparseImageFormatPropertiesAllocator,
typename Dispatch>
1380 Dispatch
const & d )
const
1383 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1385 "Function <vkGetPhysicalDeviceSparseImageFormatProperties> requires <VK_VERSION_1_0>" );
1388 std::vector<VULKAN_HPP_NAMESPACE::SparseImageFormatProperties, SparseImageFormatPropertiesAllocator> properties;
1389 uint32_t propertyCount;
1390 d.vkGetPhysicalDeviceSparseImageFormatProperties( m_physicalDevice,
1398 properties.resize( propertyCount );
1399 d.vkGetPhysicalDeviceSparseImageFormatProperties( m_physicalDevice,
1409 if ( propertyCount < properties.size() )
1411 properties.resize( propertyCount );
1417 typename SparseImageFormatPropertiesAllocator,
1419 typename std::enable_if<std::is_same<typename SparseImageFormatPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::SparseImageFormatProperties>::value,
1427 SparseImageFormatPropertiesAllocator & sparseImageFormatPropertiesAllocator,
1428 Dispatch
const & d )
const
1431 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1433 "Function <vkGetPhysicalDeviceSparseImageFormatProperties> requires <VK_VERSION_1_0>" );
1436 std::vector<VULKAN_HPP_NAMESPACE::SparseImageFormatProperties, SparseImageFormatPropertiesAllocator> properties( sparseImageFormatPropertiesAllocator );
1437 uint32_t propertyCount;
1438 d.vkGetPhysicalDeviceSparseImageFormatProperties( m_physicalDevice,
1446 properties.resize( propertyCount );
1447 d.vkGetPhysicalDeviceSparseImageFormatProperties( m_physicalDevice,
1457 if ( propertyCount < properties.size() )
1459 properties.resize( propertyCount );
1465 template <
typename Dispatch>
1467 const VULKAN_HPP_NAMESPACE::BindSparseInfo * pBindInfo,
1468 VULKAN_HPP_NAMESPACE::Fence fence,
1472 return static_cast<Result>(
1473 d.vkQueueBindSparse( m_queue, bindInfoCount,
reinterpret_cast<const VkBindSparseInfo *
>( pBindInfo ),
static_cast<VkFence>( fence ) ) );
1476 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1477 template <
typename Dispatch>
1479 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BindSparseInfo>
const & bindInfo, VULKAN_HPP_NAMESPACE::Fence fence, Dispatch
const & d )
const
1482 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1483 VULKAN_HPP_ASSERT( d.vkQueueBindSparse &&
"Function <vkQueueBindSparse> requires <VK_VERSION_1_0>" );
1487 d.vkQueueBindSparse( m_queue, bindInfo.size(),
reinterpret_cast<const VkBindSparseInfo *
>( bindInfo.data() ),
static_cast<VkFence>( fence ) ) );
1494 template <
typename Dispatch>
1496 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1497 VULKAN_HPP_NAMESPACE::Fence * pFence,
1501 return static_cast<Result>( d.vkCreateFence( m_device,
1504 reinterpret_cast<VkFence *
>( pFence ) ) );
1507 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1508 template <
typename Dispatch>
1513 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1514 VULKAN_HPP_ASSERT( d.vkCreateFence &&
"Function <vkCreateFence> requires <VK_VERSION_1_0>" );
1517 VULKAN_HPP_NAMESPACE::Fence fence;
1519 d.vkCreateFence( m_device,
1521 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
1522 reinterpret_cast<VkFence *
>( &fence ) ) );
1528 # ifndef VULKAN_HPP_NO_SMART_HANDLE
1529 template <
typename Dispatch>
1534 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1535 VULKAN_HPP_ASSERT( d.vkCreateFence &&
"Function <vkCreateFence> requires <VK_VERSION_1_0>" );
1538 VULKAN_HPP_NAMESPACE::Fence fence;
1540 d.vkCreateFence( m_device,
1542 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
1543 reinterpret_cast<VkFence *
>( &fence ) ) );
1552 template <
typename Dispatch>
1554 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1561 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1562 template <
typename Dispatch>
1568 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1569 VULKAN_HPP_ASSERT( d.vkDestroyFence &&
"Function <vkDestroyFence> requires <VK_VERSION_1_0>" );
1572 d.vkDestroyFence( m_device,
1573 static_cast<VkFence>( fence ),
1574 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
1578 template <
typename Dispatch>
1580 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1587 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1588 template <
typename Dispatch>
1594 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1595 VULKAN_HPP_ASSERT( d.vkDestroyFence &&
"Function <vkDestroyFence> requires <VK_VERSION_1_0>" );
1598 d.vkDestroyFence( m_device,
1599 static_cast<VkFence>( fence ),
1600 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
1604 template <
typename Dispatch>
1606 const VULKAN_HPP_NAMESPACE::Fence * pFences,
1610 return static_cast<Result>( d.vkResetFences( m_device, fenceCount,
reinterpret_cast<const VkFence *
>( pFences ) ) );
1613 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1614 template <
typename Dispatch>
1616 Device::resetFences( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Fence>
const & fences, Dispatch
const & d )
const
1619 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1620 VULKAN_HPP_ASSERT( d.vkResetFences &&
"Function <vkResetFences> requires <VK_VERSION_1_0>" );
1631 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
1632 template <
typename Dispatch>
1636 return static_cast<Result>( d.vkGetFenceStatus( m_device,
static_cast<VkFence>( fence ) ) );
1639 template <
typename Dispatch>
1643 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1644 VULKAN_HPP_ASSERT( d.vkGetFenceStatus &&
"Function <vkGetFenceStatus> requires <VK_VERSION_1_0>" );
1649 result,
VULKAN_HPP_NAMESPACE_STRING "::Device::getFenceStatus", { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eNotReady } );
1655 template <
typename Dispatch>
1657 const VULKAN_HPP_NAMESPACE::Fence * pFences,
1663 return static_cast<Result>(
1664 d.vkWaitForFences( m_device, fenceCount,
reinterpret_cast<const VkFence *
>( pFences ),
static_cast<VkBool32>( waitAll ), timeout ) );
1667 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1668 template <
typename Dispatch>
1670 Device::waitForFences( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Fence>
const & fences,
1673 Dispatch
const & d )
const
1676 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1677 VULKAN_HPP_ASSERT( d.vkWaitForFences &&
"Function <vkWaitForFences> requires <VK_VERSION_1_0>" );
1681 d.vkWaitForFences( m_device, fences.size(),
reinterpret_cast<const VkFence *
>( fences.data() ),
static_cast<VkBool32>( waitAll ), timeout ) );
1683 result,
VULKAN_HPP_NAMESPACE_STRING "::Device::waitForFences", { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eTimeout } );
1689 template <
typename Dispatch>
1691 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1692 VULKAN_HPP_NAMESPACE::Semaphore * pSemaphore,
1696 return static_cast<Result>( d.vkCreateSemaphore( m_device,
1699 reinterpret_cast<VkSemaphore *
>( pSemaphore ) ) );
1702 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1703 template <
typename Dispatch>
1705 Device::createSemaphore(
const VULKAN_HPP_NAMESPACE::SemaphoreCreateInfo & createInfo,
1707 Dispatch
const & d )
const
1710 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1711 VULKAN_HPP_ASSERT( d.vkCreateSemaphore &&
"Function <vkCreateSemaphore> requires <VK_VERSION_1_0>" );
1714 VULKAN_HPP_NAMESPACE::Semaphore semaphore;
1716 d.vkCreateSemaphore( m_device,
1718 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
1719 reinterpret_cast<VkSemaphore *
>( &semaphore ) ) );
1725 # ifndef VULKAN_HPP_NO_SMART_HANDLE
1726 template <
typename Dispatch>
1728 Device::createSemaphoreUnique(
const VULKAN_HPP_NAMESPACE::SemaphoreCreateInfo & createInfo,
1730 Dispatch
const & d )
const
1733 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1734 VULKAN_HPP_ASSERT( d.vkCreateSemaphore &&
"Function <vkCreateSemaphore> requires <VK_VERSION_1_0>" );
1737 VULKAN_HPP_NAMESPACE::Semaphore semaphore;
1739 d.vkCreateSemaphore( m_device,
1741 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
1742 reinterpret_cast<VkSemaphore *
>( &semaphore ) ) );
1751 template <
typename Dispatch>
1753 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1760 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1761 template <
typename Dispatch>
1767 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1768 VULKAN_HPP_ASSERT( d.vkDestroySemaphore &&
"Function <vkDestroySemaphore> requires <VK_VERSION_1_0>" );
1771 d.vkDestroySemaphore( m_device,
1773 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
1777 template <
typename Dispatch>
1779 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1786 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1787 template <
typename Dispatch>
1793 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1794 VULKAN_HPP_ASSERT( d.vkDestroySemaphore &&
"Function <vkDestroySemaphore> requires <VK_VERSION_1_0>" );
1797 d.vkDestroySemaphore( m_device,
1799 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
1803 template <
typename Dispatch>
1805 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1806 VULKAN_HPP_NAMESPACE::Event * pEvent,
1810 return static_cast<Result>( d.vkCreateEvent( m_device,
1813 reinterpret_cast<VkEvent *
>( pEvent ) ) );
1816 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1817 template <
typename Dispatch>
1822 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1823 VULKAN_HPP_ASSERT( d.vkCreateEvent &&
"Function <vkCreateEvent> requires <VK_VERSION_1_0>" );
1826 VULKAN_HPP_NAMESPACE::Event event;
1828 d.vkCreateEvent( m_device,
1830 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
1831 reinterpret_cast<VkEvent *
>( &event ) ) );
1837 # ifndef VULKAN_HPP_NO_SMART_HANDLE
1838 template <
typename Dispatch>
1843 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1844 VULKAN_HPP_ASSERT( d.vkCreateEvent &&
"Function <vkCreateEvent> requires <VK_VERSION_1_0>" );
1847 VULKAN_HPP_NAMESPACE::Event event;
1849 d.vkCreateEvent( m_device,
1851 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
1852 reinterpret_cast<VkEvent *
>( &event ) ) );
1861 template <
typename Dispatch>
1863 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1870 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1871 template <
typename Dispatch>
1877 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1878 VULKAN_HPP_ASSERT( d.vkDestroyEvent &&
"Function <vkDestroyEvent> requires <VK_VERSION_1_0>" );
1881 d.vkDestroyEvent( m_device,
1882 static_cast<VkEvent>( event ),
1883 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
1887 template <
typename Dispatch>
1889 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1896 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1897 template <
typename Dispatch>
1903 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1904 VULKAN_HPP_ASSERT( d.vkDestroyEvent &&
"Function <vkDestroyEvent> requires <VK_VERSION_1_0>" );
1907 d.vkDestroyEvent( m_device,
1908 static_cast<VkEvent>( event ),
1909 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
1913 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
1914 template <
typename Dispatch>
1918 return static_cast<Result>( d.vkGetEventStatus( m_device,
static_cast<VkEvent>( event ) ) );
1921 template <
typename Dispatch>
1925 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1926 VULKAN_HPP_ASSERT( d.vkGetEventStatus &&
"Function <vkGetEventStatus> requires <VK_VERSION_1_0>" );
1931 result,
VULKAN_HPP_NAMESPACE_STRING "::Device::getEventStatus", { VULKAN_HPP_NAMESPACE::Result::eEventSet, VULKAN_HPP_NAMESPACE::Result::eEventReset } );
1937 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
1938 template <
typename Dispatch>
1942 return static_cast<Result>( d.vkSetEvent( m_device,
static_cast<VkEvent>( event ) ) );
1945 template <
typename Dispatch>
1947 Dispatch
const & d )
const
1950 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1951 VULKAN_HPP_ASSERT( d.vkSetEvent &&
"Function <vkSetEvent> requires <VK_VERSION_1_0>" );
1961 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
1962 template <
typename Dispatch>
1966 return static_cast<Result>( d.vkResetEvent( m_device,
static_cast<VkEvent>( event ) ) );
1969 template <
typename Dispatch>
1973 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
1974 VULKAN_HPP_ASSERT( d.vkResetEvent &&
"Function <vkResetEvent> requires <VK_VERSION_1_0>" );
1984 template <
typename Dispatch>
1986 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
1987 VULKAN_HPP_NAMESPACE::QueryPool * pQueryPool,
1991 return static_cast<Result>( d.vkCreateQueryPool( m_device,
1994 reinterpret_cast<VkQueryPool *
>( pQueryPool ) ) );
1997 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
1998 template <
typename Dispatch>
2000 Device::createQueryPool(
const VULKAN_HPP_NAMESPACE::QueryPoolCreateInfo & createInfo,
2002 Dispatch
const & d )
const
2005 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2006 VULKAN_HPP_ASSERT( d.vkCreateQueryPool &&
"Function <vkCreateQueryPool> requires <VK_VERSION_1_0>" );
2009 VULKAN_HPP_NAMESPACE::QueryPool queryPool;
2011 d.vkCreateQueryPool( m_device,
2013 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2014 reinterpret_cast<VkQueryPool *
>( &queryPool ) ) );
2020 # ifndef VULKAN_HPP_NO_SMART_HANDLE
2021 template <
typename Dispatch>
2023 Device::createQueryPoolUnique(
const VULKAN_HPP_NAMESPACE::QueryPoolCreateInfo & createInfo,
2025 Dispatch
const & d )
const
2028 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2029 VULKAN_HPP_ASSERT( d.vkCreateQueryPool &&
"Function <vkCreateQueryPool> requires <VK_VERSION_1_0>" );
2032 VULKAN_HPP_NAMESPACE::QueryPool queryPool;
2034 d.vkCreateQueryPool( m_device,
2036 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2037 reinterpret_cast<VkQueryPool *
>( &queryPool ) ) );
2046 template <
typename Dispatch>
2048 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2055 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2056 template <
typename Dispatch>
2062 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2063 VULKAN_HPP_ASSERT( d.vkDestroyQueryPool &&
"Function <vkDestroyQueryPool> requires <VK_VERSION_1_0>" );
2066 d.vkDestroyQueryPool( m_device,
2068 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2072 template <
typename Dispatch>
2074 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2081 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2082 template <
typename Dispatch>
2088 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2089 VULKAN_HPP_ASSERT( d.vkDestroyQueryPool &&
"Function <vkDestroyQueryPool> requires <VK_VERSION_1_0>" );
2092 d.vkDestroyQueryPool( m_device,
2094 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2098 template <
typename Dispatch>
2100 uint32_t firstQuery,
2101 uint32_t queryCount,
2109 return static_cast<Result>( d.vkGetQueryPoolResults( m_device,
2119 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2120 template <
typename DataType,
typename DataTypeAllocator,
typename Dispatch>
2122 Device::getQueryPoolResults( VULKAN_HPP_NAMESPACE::QueryPool queryPool,
2123 uint32_t firstQuery,
2124 uint32_t queryCount,
2128 Dispatch
const & d )
const
2131 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2132 VULKAN_HPP_ASSERT( d.vkGetQueryPoolResults &&
"Function <vkGetQueryPoolResults> requires <VK_VERSION_1_0>" );
2136 std::vector<DataType, DataTypeAllocator> data( dataSize /
sizeof( DataType ) );
2141 data.size() *
sizeof( DataType ),
2142 reinterpret_cast<void *
>( data.data() ),
2147 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eNotReady } );
2152 template <
typename DataType,
typename Dispatch>
2154 uint32_t firstQuery,
2155 uint32_t queryCount,
2158 Dispatch
const & d )
const
2161 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2162 VULKAN_HPP_ASSERT( d.vkGetQueryPoolResults &&
"Function <vkGetQueryPoolResults> requires <VK_VERSION_1_0>" );
2171 reinterpret_cast<void *
>( &data ),
2175 result,
VULKAN_HPP_NAMESPACE_STRING "::Device::getQueryPoolResult", { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eNotReady } );
2181 template <
typename Dispatch>
2183 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2184 VULKAN_HPP_NAMESPACE::Buffer * pBuffer,
2188 return static_cast<Result>( d.vkCreateBuffer( m_device,
2191 reinterpret_cast<VkBuffer *
>( pBuffer ) ) );
2194 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2195 template <
typename Dispatch>
2200 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2201 VULKAN_HPP_ASSERT( d.vkCreateBuffer &&
"Function <vkCreateBuffer> requires <VK_VERSION_1_0>" );
2204 VULKAN_HPP_NAMESPACE::Buffer buffer;
2206 d.vkCreateBuffer( m_device,
2208 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2209 reinterpret_cast<VkBuffer *
>( &buffer ) ) );
2215 # ifndef VULKAN_HPP_NO_SMART_HANDLE
2216 template <
typename Dispatch>
2221 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2222 VULKAN_HPP_ASSERT( d.vkCreateBuffer &&
"Function <vkCreateBuffer> requires <VK_VERSION_1_0>" );
2225 VULKAN_HPP_NAMESPACE::Buffer buffer;
2227 d.vkCreateBuffer( m_device,
2229 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2230 reinterpret_cast<VkBuffer *
>( &buffer ) ) );
2239 template <
typename Dispatch>
2241 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2248 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2249 template <
typename Dispatch>
2255 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2256 VULKAN_HPP_ASSERT( d.vkDestroyBuffer &&
"Function <vkDestroyBuffer> requires <VK_VERSION_1_0>" );
2259 d.vkDestroyBuffer( m_device,
2261 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2265 template <
typename Dispatch>
2267 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2274 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2275 template <
typename Dispatch>
2281 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2282 VULKAN_HPP_ASSERT( d.vkDestroyBuffer &&
"Function <vkDestroyBuffer> requires <VK_VERSION_1_0>" );
2285 d.vkDestroyBuffer( m_device,
2287 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2291 template <
typename Dispatch>
2293 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2294 VULKAN_HPP_NAMESPACE::BufferView * pView,
2298 return static_cast<Result>( d.vkCreateBufferView( m_device,
2304 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2305 template <
typename Dispatch>
2307 Device::createBufferView(
const VULKAN_HPP_NAMESPACE::BufferViewCreateInfo & createInfo,
2309 Dispatch
const & d )
const
2312 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2313 VULKAN_HPP_ASSERT( d.vkCreateBufferView &&
"Function <vkCreateBufferView> requires <VK_VERSION_1_0>" );
2316 VULKAN_HPP_NAMESPACE::BufferView view;
2318 d.vkCreateBufferView( m_device,
2320 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2327 # ifndef VULKAN_HPP_NO_SMART_HANDLE
2328 template <
typename Dispatch>
2330 Device::createBufferViewUnique(
const VULKAN_HPP_NAMESPACE::BufferViewCreateInfo & createInfo,
2332 Dispatch
const & d )
const
2335 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2336 VULKAN_HPP_ASSERT( d.vkCreateBufferView &&
"Function <vkCreateBufferView> requires <VK_VERSION_1_0>" );
2339 VULKAN_HPP_NAMESPACE::BufferView view;
2341 d.vkCreateBufferView( m_device,
2343 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2353 template <
typename Dispatch>
2355 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2362 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2363 template <
typename Dispatch>
2369 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2370 VULKAN_HPP_ASSERT( d.vkDestroyBufferView &&
"Function <vkDestroyBufferView> requires <VK_VERSION_1_0>" );
2373 d.vkDestroyBufferView( m_device,
2375 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2379 template <
typename Dispatch>
2381 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2388 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2389 template <
typename Dispatch>
2395 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2396 VULKAN_HPP_ASSERT( d.vkDestroyBufferView &&
"Function <vkDestroyBufferView> requires <VK_VERSION_1_0>" );
2399 d.vkDestroyBufferView( m_device,
2401 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2405 template <
typename Dispatch>
2407 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2408 VULKAN_HPP_NAMESPACE::Image * pImage,
2412 return static_cast<Result>( d.vkCreateImage( m_device,
2415 reinterpret_cast<VkImage *
>( pImage ) ) );
2418 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2419 template <
typename Dispatch>
2424 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2425 VULKAN_HPP_ASSERT( d.vkCreateImage &&
"Function <vkCreateImage> requires <VK_VERSION_1_0>" );
2428 VULKAN_HPP_NAMESPACE::Image image;
2430 d.vkCreateImage( m_device,
2432 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2433 reinterpret_cast<VkImage *
>( &image ) ) );
2439 # ifndef VULKAN_HPP_NO_SMART_HANDLE
2440 template <
typename Dispatch>
2445 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2446 VULKAN_HPP_ASSERT( d.vkCreateImage &&
"Function <vkCreateImage> requires <VK_VERSION_1_0>" );
2449 VULKAN_HPP_NAMESPACE::Image image;
2451 d.vkCreateImage( m_device,
2453 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2454 reinterpret_cast<VkImage *
>( &image ) ) );
2463 template <
typename Dispatch>
2465 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2472 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2473 template <
typename Dispatch>
2479 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2480 VULKAN_HPP_ASSERT( d.vkDestroyImage &&
"Function <vkDestroyImage> requires <VK_VERSION_1_0>" );
2483 d.vkDestroyImage( m_device,
2484 static_cast<VkImage>( image ),
2485 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2489 template <
typename Dispatch>
2491 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2498 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2499 template <
typename Dispatch>
2505 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2506 VULKAN_HPP_ASSERT( d.vkDestroyImage &&
"Function <vkDestroyImage> requires <VK_VERSION_1_0>" );
2509 d.vkDestroyImage( m_device,
2510 static_cast<VkImage>( image ),
2511 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2515 template <
typename Dispatch>
2517 const VULKAN_HPP_NAMESPACE::ImageSubresource * pSubresource,
2518 VULKAN_HPP_NAMESPACE::SubresourceLayout * pLayout,
2522 d.vkGetImageSubresourceLayout( m_device,
2523 static_cast<VkImage>( image ),
2528 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2529 template <
typename Dispatch>
2531 VULKAN_HPP_NAMESPACE::Image image,
const VULKAN_HPP_NAMESPACE::ImageSubresource & subresource, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
2534 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2535 VULKAN_HPP_ASSERT( d.vkGetImageSubresourceLayout &&
"Function <vkGetImageSubresourceLayout> requires <VK_VERSION_1_0>" );
2538 VULKAN_HPP_NAMESPACE::SubresourceLayout layout;
2539 d.vkGetImageSubresourceLayout( m_device,
2540 static_cast<VkImage>( image ),
2548 template <
typename Dispatch>
2550 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2551 VULKAN_HPP_NAMESPACE::ImageView * pView,
2555 return static_cast<Result>( d.vkCreateImageView( m_device,
2561 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2562 template <
typename Dispatch>
2564 Device::createImageView(
const VULKAN_HPP_NAMESPACE::ImageViewCreateInfo & createInfo,
2566 Dispatch
const & d )
const
2569 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2570 VULKAN_HPP_ASSERT( d.vkCreateImageView &&
"Function <vkCreateImageView> requires <VK_VERSION_1_0>" );
2573 VULKAN_HPP_NAMESPACE::ImageView view;
2575 d.vkCreateImageView( m_device,
2577 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2584 # ifndef VULKAN_HPP_NO_SMART_HANDLE
2585 template <
typename Dispatch>
2587 Device::createImageViewUnique(
const VULKAN_HPP_NAMESPACE::ImageViewCreateInfo & createInfo,
2589 Dispatch
const & d )
const
2592 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2593 VULKAN_HPP_ASSERT( d.vkCreateImageView &&
"Function <vkCreateImageView> requires <VK_VERSION_1_0>" );
2596 VULKAN_HPP_NAMESPACE::ImageView view;
2598 d.vkCreateImageView( m_device,
2600 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2610 template <
typename Dispatch>
2612 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2619 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2620 template <
typename Dispatch>
2626 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2627 VULKAN_HPP_ASSERT( d.vkDestroyImageView &&
"Function <vkDestroyImageView> requires <VK_VERSION_1_0>" );
2630 d.vkDestroyImageView( m_device,
2632 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2636 template <
typename Dispatch>
2638 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2645 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2646 template <
typename Dispatch>
2652 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2653 VULKAN_HPP_ASSERT( d.vkDestroyImageView &&
"Function <vkDestroyImageView> requires <VK_VERSION_1_0>" );
2656 d.vkDestroyImageView( m_device,
2658 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2662 template <
typename Dispatch>
2664 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2665 VULKAN_HPP_NAMESPACE::ShaderModule * pShaderModule,
2669 return static_cast<Result>( d.vkCreateShaderModule( m_device,
2675 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2676 template <
typename Dispatch>
2680 Dispatch
const & d )
const
2683 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2684 VULKAN_HPP_ASSERT( d.vkCreateShaderModule &&
"Function <vkCreateShaderModule> requires <VK_VERSION_1_0>" );
2687 VULKAN_HPP_NAMESPACE::ShaderModule shaderModule;
2689 d.vkCreateShaderModule( m_device,
2691 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2698 # ifndef VULKAN_HPP_NO_SMART_HANDLE
2699 template <
typename Dispatch>
2701 Device::createShaderModuleUnique(
const VULKAN_HPP_NAMESPACE::ShaderModuleCreateInfo & createInfo,
2703 Dispatch
const & d )
const
2706 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2707 VULKAN_HPP_ASSERT( d.vkCreateShaderModule &&
"Function <vkCreateShaderModule> requires <VK_VERSION_1_0>" );
2710 VULKAN_HPP_NAMESPACE::ShaderModule shaderModule;
2712 d.vkCreateShaderModule( m_device,
2714 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2724 template <
typename Dispatch>
2725 VULKAN_HPP_INLINE void Device::destroyShaderModule( VULKAN_HPP_NAMESPACE::ShaderModule shaderModule,
2726 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2733 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2734 template <
typename Dispatch>
2735 VULKAN_HPP_INLINE void Device::destroyShaderModule( VULKAN_HPP_NAMESPACE::ShaderModule shaderModule,
2740 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2741 VULKAN_HPP_ASSERT( d.vkDestroyShaderModule &&
"Function <vkDestroyShaderModule> requires <VK_VERSION_1_0>" );
2744 d.vkDestroyShaderModule( m_device,
2746 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2750 template <
typename Dispatch>
2752 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2759 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2760 template <
typename Dispatch>
2766 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2767 VULKAN_HPP_ASSERT( d.vkDestroyShaderModule &&
"Function <vkDestroyShaderModule> requires <VK_VERSION_1_0>" );
2770 d.vkDestroyShaderModule( m_device,
2772 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2776 template <
typename Dispatch>
2778 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2779 VULKAN_HPP_NAMESPACE::PipelineCache * pPipelineCache,
2783 return static_cast<Result>( d.vkCreatePipelineCache( m_device,
2789 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2790 template <
typename Dispatch>
2792 Device::createPipelineCache(
const VULKAN_HPP_NAMESPACE::PipelineCacheCreateInfo & createInfo,
2794 Dispatch
const & d )
const
2797 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2798 VULKAN_HPP_ASSERT( d.vkCreatePipelineCache &&
"Function <vkCreatePipelineCache> requires <VK_VERSION_1_0>" );
2801 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache;
2803 d.vkCreatePipelineCache( m_device,
2805 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2812 # ifndef VULKAN_HPP_NO_SMART_HANDLE
2813 template <
typename Dispatch>
2815 Device::createPipelineCacheUnique(
const VULKAN_HPP_NAMESPACE::PipelineCacheCreateInfo & createInfo,
2817 Dispatch
const & d )
const
2820 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2821 VULKAN_HPP_ASSERT( d.vkCreatePipelineCache &&
"Function <vkCreatePipelineCache> requires <VK_VERSION_1_0>" );
2824 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache;
2826 d.vkCreatePipelineCache( m_device,
2828 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
2838 template <
typename Dispatch>
2839 VULKAN_HPP_INLINE void Device::destroyPipelineCache( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
2840 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2847 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2848 template <
typename Dispatch>
2849 VULKAN_HPP_INLINE void Device::destroyPipelineCache( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
2854 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2855 VULKAN_HPP_ASSERT( d.vkDestroyPipelineCache &&
"Function <vkDestroyPipelineCache> requires <VK_VERSION_1_0>" );
2858 d.vkDestroyPipelineCache( m_device,
2860 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2864 template <
typename Dispatch>
2866 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
2873 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2874 template <
typename Dispatch>
2880 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2881 VULKAN_HPP_ASSERT( d.vkDestroyPipelineCache &&
"Function <vkDestroyPipelineCache> requires <VK_VERSION_1_0>" );
2884 d.vkDestroyPipelineCache( m_device,
2886 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
2890 template <
typename Dispatch>
2897 return static_cast<Result>( d.vkGetPipelineCacheData( m_device,
static_cast<VkPipelineCache>( pipelineCache ), pDataSize, pData ) );
2900 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2901 template <
typename U
int8_tAllocator,
typename Dispatch>
2903 Device::getPipelineCacheData( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, Dispatch
const & d )
const
2906 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2907 VULKAN_HPP_ASSERT( d.vkGetPipelineCacheData &&
"Function <vkGetPipelineCacheData> requires <VK_VERSION_1_0>" );
2910 std::vector<uint8_t, Uint8_tAllocator> data;
2917 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
2919 data.resize( dataSize );
2921 d.vkGetPipelineCacheData( m_device,
static_cast<VkPipelineCache>( pipelineCache ), &dataSize,
reinterpret_cast<void *
>( data.data() ) ) );
2923 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
2926 if ( dataSize < data.size() )
2928 data.resize( dataSize );
2933 template <
typename Uint8_tAllocator,
2935 typename std::enable_if<std::is_same<typename Uint8_tAllocator::value_type, uint8_t>::value,
int>::type>
2937 Device::getPipelineCacheData( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, Uint8_tAllocator & uint8_tAllocator, Dispatch
const & d )
const
2940 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2941 VULKAN_HPP_ASSERT( d.vkGetPipelineCacheData &&
"Function <vkGetPipelineCacheData> requires <VK_VERSION_1_0>" );
2944 std::vector<uint8_t, Uint8_tAllocator> data( uint8_tAllocator );
2951 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
2953 data.resize( dataSize );
2955 d.vkGetPipelineCacheData( m_device,
static_cast<VkPipelineCache>( pipelineCache ), &dataSize,
reinterpret_cast<void *
>( data.data() ) ) );
2957 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
2960 if ( dataSize < data.size() )
2962 data.resize( dataSize );
2968 template <
typename Dispatch>
2970 uint32_t srcCacheCount,
2971 const VULKAN_HPP_NAMESPACE::PipelineCache * pSrcCaches,
2975 return static_cast<Result>(
2976 d.vkMergePipelineCaches( m_device,
static_cast<VkPipelineCache>( dstCache ), srcCacheCount,
reinterpret_cast<const VkPipelineCache *
>( pSrcCaches ) ) );
2979 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
2980 template <
typename Dispatch>
2982 Device::mergePipelineCaches( VULKAN_HPP_NAMESPACE::PipelineCache dstCache,
2983 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::PipelineCache>
const & srcCaches,
2984 Dispatch
const & d )
const
2987 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
2988 VULKAN_HPP_ASSERT( d.vkMergePipelineCaches &&
"Function <vkMergePipelineCaches> requires <VK_VERSION_1_0>" );
2999 template <
typename Dispatch>
3001 uint32_t createInfoCount,
3002 const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo * pCreateInfos,
3003 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3004 VULKAN_HPP_NAMESPACE::Pipeline * pPipelines,
3008 return static_cast<Result>( d.vkCreateGraphicsPipelines( m_device,
3013 reinterpret_cast<VkPipeline *
>( pPipelines ) ) );
3016 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3017 template <
typename PipelineAllocator,
typename Dispatch>
3019 Device::createGraphicsPipelines( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3020 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo>
const & createInfos,
3022 Dispatch
const & d )
const
3025 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3026 VULKAN_HPP_ASSERT( d.vkCreateGraphicsPipelines &&
"Function <vkCreateGraphicsPipelines> requires <VK_VERSION_1_0>" );
3029 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
3035 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3036 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
3039 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3044 template <
typename PipelineAllocator,
3046 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, VULKAN_HPP_NAMESPACE::Pipeline>::value,
int>::type>
3048 Device::createGraphicsPipelines( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3049 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo>
const & createInfos,
3051 PipelineAllocator & pipelineAllocator,
3052 Dispatch
const & d )
const
3055 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3056 VULKAN_HPP_ASSERT( d.vkCreateGraphicsPipelines &&
"Function <vkCreateGraphicsPipelines> requires <VK_VERSION_1_0>" );
3059 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
3065 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3066 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
3069 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3074 template <
typename Dispatch>
3077 const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo & createInfo,
3079 Dispatch
const & d )
const
3082 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3083 VULKAN_HPP_ASSERT( d.vkCreateGraphicsPipelines &&
"Function <vkCreateGraphicsPipelines> requires <VK_VERSION_1_0>" );
3086 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
3092 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3093 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
3096 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3101 # ifndef VULKAN_HPP_NO_SMART_HANDLE
3102 template <
typename Dispatch,
typename PipelineAllocator>
3104 Device::createGraphicsPipelinesUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3105 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo>
const & createInfos,
3107 Dispatch
const & d )
const
3110 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3111 VULKAN_HPP_ASSERT( d.vkCreateGraphicsPipelines &&
"Function <vkCreateGraphicsPipelines> requires <VK_VERSION_1_0>" );
3114 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
3120 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3121 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
3124 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3125 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines;
3126 uniquePipelines.reserve( createInfos.size() );
3128 for (
auto const & pipeline : pipelines )
3138 typename PipelineAllocator,
3139 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>>::value,
int>::type>
3141 Device::createGraphicsPipelinesUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3142 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo>
const & createInfos,
3144 PipelineAllocator & pipelineAllocator,
3145 Dispatch
const & d )
const
3148 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3149 VULKAN_HPP_ASSERT( d.vkCreateGraphicsPipelines &&
"Function <vkCreateGraphicsPipelines> requires <VK_VERSION_1_0>" );
3152 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
3158 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3159 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
3162 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3163 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines( pipelineAllocator );
3164 uniquePipelines.reserve( createInfos.size() );
3166 for (
auto const & pipeline : pipelines )
3174 template <
typename Dispatch>
3176 Device::createGraphicsPipelineUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3177 const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo & createInfo,
3179 Dispatch
const & d )
const
3182 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3183 VULKAN_HPP_ASSERT( d.vkCreateGraphicsPipelines &&
"Function <vkCreateGraphicsPipelines> requires <VK_VERSION_1_0>" );
3186 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
3192 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3193 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
3196 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3205 template <
typename Dispatch>
3207 uint32_t createInfoCount,
3208 const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo * pCreateInfos,
3209 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3210 VULKAN_HPP_NAMESPACE::Pipeline * pPipelines,
3214 return static_cast<Result>( d.vkCreateComputePipelines( m_device,
3219 reinterpret_cast<VkPipeline *
>( pPipelines ) ) );
3222 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3223 template <
typename PipelineAllocator,
typename Dispatch>
3225 Device::createComputePipelines( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3226 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo>
const & createInfos,
3228 Dispatch
const & d )
const
3231 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3232 VULKAN_HPP_ASSERT( d.vkCreateComputePipelines &&
"Function <vkCreateComputePipelines> requires <VK_VERSION_1_0>" );
3235 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
3241 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3242 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
3245 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3250 template <
typename PipelineAllocator,
3252 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, VULKAN_HPP_NAMESPACE::Pipeline>::value,
int>::type>
3254 Device::createComputePipelines( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3255 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo>
const & createInfos,
3257 PipelineAllocator & pipelineAllocator,
3258 Dispatch
const & d )
const
3261 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3262 VULKAN_HPP_ASSERT( d.vkCreateComputePipelines &&
"Function <vkCreateComputePipelines> requires <VK_VERSION_1_0>" );
3265 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
3271 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3272 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
3275 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3280 template <
typename Dispatch>
3282 Device::createComputePipeline( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3283 const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo & createInfo,
3285 Dispatch
const & d )
const
3288 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3289 VULKAN_HPP_ASSERT( d.vkCreateComputePipelines &&
"Function <vkCreateComputePipelines> requires <VK_VERSION_1_0>" );
3292 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
3298 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3299 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
3302 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3307 # ifndef VULKAN_HPP_NO_SMART_HANDLE
3308 template <
typename Dispatch,
typename PipelineAllocator>
3310 Device::createComputePipelinesUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3311 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo>
const & createInfos,
3313 Dispatch
const & d )
const
3316 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3317 VULKAN_HPP_ASSERT( d.vkCreateComputePipelines &&
"Function <vkCreateComputePipelines> requires <VK_VERSION_1_0>" );
3320 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
3326 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3327 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
3330 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3331 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines;
3332 uniquePipelines.reserve( createInfos.size() );
3334 for (
auto const & pipeline : pipelines )
3344 typename PipelineAllocator,
3345 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>>::value,
int>::type>
3347 Device::createComputePipelinesUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3348 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo>
const & createInfos,
3350 PipelineAllocator & pipelineAllocator,
3351 Dispatch
const & d )
const
3354 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3355 VULKAN_HPP_ASSERT( d.vkCreateComputePipelines &&
"Function <vkCreateComputePipelines> requires <VK_VERSION_1_0>" );
3358 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
3364 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3365 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
3368 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3369 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines( pipelineAllocator );
3370 uniquePipelines.reserve( createInfos.size() );
3372 for (
auto const & pipeline : pipelines )
3380 template <
typename Dispatch>
3382 Device::createComputePipelineUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
3383 const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo & createInfo,
3385 Dispatch
const & d )
const
3388 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3389 VULKAN_HPP_ASSERT( d.vkCreateComputePipelines &&
"Function <vkCreateComputePipelines> requires <VK_VERSION_1_0>" );
3392 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
3398 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3399 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
3402 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
3411 template <
typename Dispatch>
3413 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3420 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3421 template <
typename Dispatch>
3427 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3428 VULKAN_HPP_ASSERT( d.vkDestroyPipeline &&
"Function <vkDestroyPipeline> requires <VK_VERSION_1_0>" );
3431 d.vkDestroyPipeline( m_device,
3433 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3437 template <
typename Dispatch>
3439 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3446 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3447 template <
typename Dispatch>
3453 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3454 VULKAN_HPP_ASSERT( d.vkDestroyPipeline &&
"Function <vkDestroyPipeline> requires <VK_VERSION_1_0>" );
3457 d.vkDestroyPipeline( m_device,
3459 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3463 template <
typename Dispatch>
3465 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3466 VULKAN_HPP_NAMESPACE::PipelineLayout * pPipelineLayout,
3470 return static_cast<Result>( d.vkCreatePipelineLayout( m_device,
3476 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3477 template <
typename Dispatch>
3479 Device::createPipelineLayout(
const VULKAN_HPP_NAMESPACE::PipelineLayoutCreateInfo & createInfo,
3481 Dispatch
const & d )
const
3484 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3485 VULKAN_HPP_ASSERT( d.vkCreatePipelineLayout &&
"Function <vkCreatePipelineLayout> requires <VK_VERSION_1_0>" );
3488 VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout;
3490 d.vkCreatePipelineLayout( m_device,
3492 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3499 # ifndef VULKAN_HPP_NO_SMART_HANDLE
3500 template <
typename Dispatch>
3502 Device::createPipelineLayoutUnique(
const VULKAN_HPP_NAMESPACE::PipelineLayoutCreateInfo & createInfo,
3504 Dispatch
const & d )
const
3507 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3508 VULKAN_HPP_ASSERT( d.vkCreatePipelineLayout &&
"Function <vkCreatePipelineLayout> requires <VK_VERSION_1_0>" );
3511 VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout;
3513 d.vkCreatePipelineLayout( m_device,
3515 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3525 template <
typename Dispatch>
3526 VULKAN_HPP_INLINE void Device::destroyPipelineLayout( VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout,
3527 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3534 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3535 template <
typename Dispatch>
3536 VULKAN_HPP_INLINE void Device::destroyPipelineLayout( VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout,
3541 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3542 VULKAN_HPP_ASSERT( d.vkDestroyPipelineLayout &&
"Function <vkDestroyPipelineLayout> requires <VK_VERSION_1_0>" );
3545 d.vkDestroyPipelineLayout( m_device,
3547 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3551 template <
typename Dispatch>
3553 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3560 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3561 template <
typename Dispatch>
3567 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3568 VULKAN_HPP_ASSERT( d.vkDestroyPipelineLayout &&
"Function <vkDestroyPipelineLayout> requires <VK_VERSION_1_0>" );
3571 d.vkDestroyPipelineLayout( m_device,
3573 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3577 template <
typename Dispatch>
3579 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3580 VULKAN_HPP_NAMESPACE::Sampler * pSampler,
3584 return static_cast<Result>( d.vkCreateSampler( m_device,
3587 reinterpret_cast<VkSampler *
>( pSampler ) ) );
3590 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3591 template <
typename Dispatch>
3596 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3597 VULKAN_HPP_ASSERT( d.vkCreateSampler &&
"Function <vkCreateSampler> requires <VK_VERSION_1_0>" );
3600 VULKAN_HPP_NAMESPACE::Sampler sampler;
3602 d.vkCreateSampler( m_device,
3604 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3605 reinterpret_cast<VkSampler *
>( &sampler ) ) );
3611 # ifndef VULKAN_HPP_NO_SMART_HANDLE
3612 template <
typename Dispatch>
3617 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3618 VULKAN_HPP_ASSERT( d.vkCreateSampler &&
"Function <vkCreateSampler> requires <VK_VERSION_1_0>" );
3621 VULKAN_HPP_NAMESPACE::Sampler sampler;
3623 d.vkCreateSampler( m_device,
3625 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3626 reinterpret_cast<VkSampler *
>( &sampler ) ) );
3635 template <
typename Dispatch>
3637 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3644 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3645 template <
typename Dispatch>
3651 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3652 VULKAN_HPP_ASSERT( d.vkDestroySampler &&
"Function <vkDestroySampler> requires <VK_VERSION_1_0>" );
3655 d.vkDestroySampler( m_device,
3657 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3661 template <
typename Dispatch>
3663 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3670 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3671 template <
typename Dispatch>
3677 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3678 VULKAN_HPP_ASSERT( d.vkDestroySampler &&
"Function <vkDestroySampler> requires <VK_VERSION_1_0>" );
3681 d.vkDestroySampler( m_device,
3683 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3687 template <
typename Dispatch>
3689 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3690 VULKAN_HPP_NAMESPACE::DescriptorSetLayout * pSetLayout,
3694 return static_cast<Result>( d.vkCreateDescriptorSetLayout( m_device,
3700 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3701 template <
typename Dispatch>
3705 Dispatch
const & d )
const
3708 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3709 VULKAN_HPP_ASSERT( d.vkCreateDescriptorSetLayout &&
"Function <vkCreateDescriptorSetLayout> requires <VK_VERSION_1_0>" );
3712 VULKAN_HPP_NAMESPACE::DescriptorSetLayout setLayout;
3716 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3723 # ifndef VULKAN_HPP_NO_SMART_HANDLE
3724 template <
typename Dispatch>
3726 Device::createDescriptorSetLayoutUnique(
const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo & createInfo,
3728 Dispatch
const & d )
const
3731 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3732 VULKAN_HPP_ASSERT( d.vkCreateDescriptorSetLayout &&
"Function <vkCreateDescriptorSetLayout> requires <VK_VERSION_1_0>" );
3735 VULKAN_HPP_NAMESPACE::DescriptorSetLayout setLayout;
3739 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3749 template <
typename Dispatch>
3750 VULKAN_HPP_INLINE void Device::destroyDescriptorSetLayout( VULKAN_HPP_NAMESPACE::DescriptorSetLayout descriptorSetLayout,
3751 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3755 d.vkDestroyDescriptorSetLayout(
3759 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3760 template <
typename Dispatch>
3761 VULKAN_HPP_INLINE void Device::destroyDescriptorSetLayout( VULKAN_HPP_NAMESPACE::DescriptorSetLayout descriptorSetLayout,
3766 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3767 VULKAN_HPP_ASSERT( d.vkDestroyDescriptorSetLayout &&
"Function <vkDestroyDescriptorSetLayout> requires <VK_VERSION_1_0>" );
3770 d.vkDestroyDescriptorSetLayout(
3773 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3777 template <
typename Dispatch>
3778 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::DescriptorSetLayout descriptorSetLayout,
3779 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3783 d.vkDestroyDescriptorSetLayout(
3787 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3788 template <
typename Dispatch>
3789 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::DescriptorSetLayout descriptorSetLayout,
3794 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3795 VULKAN_HPP_ASSERT( d.vkDestroyDescriptorSetLayout &&
"Function <vkDestroyDescriptorSetLayout> requires <VK_VERSION_1_0>" );
3798 d.vkDestroyDescriptorSetLayout(
3801 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3805 template <
typename Dispatch>
3807 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3808 VULKAN_HPP_NAMESPACE::DescriptorPool * pDescriptorPool,
3812 return static_cast<Result>( d.vkCreateDescriptorPool( m_device,
3818 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3819 template <
typename Dispatch>
3823 Dispatch
const & d )
const
3826 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3827 VULKAN_HPP_ASSERT( d.vkCreateDescriptorPool &&
"Function <vkCreateDescriptorPool> requires <VK_VERSION_1_0>" );
3830 VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool;
3832 d.vkCreateDescriptorPool( m_device,
3834 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3841 # ifndef VULKAN_HPP_NO_SMART_HANDLE
3842 template <
typename Dispatch>
3844 Device::createDescriptorPoolUnique(
const VULKAN_HPP_NAMESPACE::DescriptorPoolCreateInfo & createInfo,
3846 Dispatch
const & d )
const
3849 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3850 VULKAN_HPP_ASSERT( d.vkCreateDescriptorPool &&
"Function <vkCreateDescriptorPool> requires <VK_VERSION_1_0>" );
3853 VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool;
3855 d.vkCreateDescriptorPool( m_device,
3857 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
3867 template <
typename Dispatch>
3868 VULKAN_HPP_INLINE void Device::destroyDescriptorPool( VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool,
3869 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3876 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3877 template <
typename Dispatch>
3878 VULKAN_HPP_INLINE void Device::destroyDescriptorPool( VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool,
3883 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3884 VULKAN_HPP_ASSERT( d.vkDestroyDescriptorPool &&
"Function <vkDestroyDescriptorPool> requires <VK_VERSION_1_0>" );
3887 d.vkDestroyDescriptorPool( m_device,
3889 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3893 template <
typename Dispatch>
3895 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
3902 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3903 template <
typename Dispatch>
3909 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3910 VULKAN_HPP_ASSERT( d.vkDestroyDescriptorPool &&
"Function <vkDestroyDescriptorPool> requires <VK_VERSION_1_0>" );
3913 d.vkDestroyDescriptorPool( m_device,
3915 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
3919 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
3920 template <
typename Dispatch>
3926 return static_cast<Result>(
3930 template <
typename Dispatch>
3931 VULKAN_HPP_INLINE void Device::resetDescriptorPool( VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool,
3936 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3937 VULKAN_HPP_ASSERT( d.vkResetDescriptorPool &&
"Function <vkResetDescriptorPool> requires <VK_VERSION_1_0>" );
3944 template <
typename Dispatch>
3946 VULKAN_HPP_NAMESPACE::DescriptorSet * pDescriptorSets,
3950 return static_cast<Result>( d.vkAllocateDescriptorSets(
3954 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
3955 template <
typename DescriptorSetAllocator,
typename Dispatch>
3957 Device::allocateDescriptorSets(
const VULKAN_HPP_NAMESPACE::DescriptorSetAllocateInfo & allocateInfo, Dispatch
const & d )
const
3960 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3961 VULKAN_HPP_ASSERT( d.vkAllocateDescriptorSets &&
"Function <vkAllocateDescriptorSets> requires <VK_VERSION_1_0>" );
3964 std::vector<VULKAN_HPP_NAMESPACE::DescriptorSet, DescriptorSetAllocator> descriptorSets( allocateInfo.descriptorSetCount );
3972 template <
typename DescriptorSetAllocator,
3974 typename std::enable_if<std::is_same<typename DescriptorSetAllocator::value_type, VULKAN_HPP_NAMESPACE::DescriptorSet>::value,
int>::type>
3976 Device::allocateDescriptorSets(
const VULKAN_HPP_NAMESPACE::DescriptorSetAllocateInfo & allocateInfo,
3977 DescriptorSetAllocator & descriptorSetAllocator,
3978 Dispatch
const & d )
const
3981 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
3982 VULKAN_HPP_ASSERT( d.vkAllocateDescriptorSets &&
"Function <vkAllocateDescriptorSets> requires <VK_VERSION_1_0>" );
3985 std::vector<VULKAN_HPP_NAMESPACE::DescriptorSet, DescriptorSetAllocator> descriptorSets( allocateInfo.descriptorSetCount, descriptorSetAllocator );
3993 # ifndef VULKAN_HPP_NO_SMART_HANDLE
3994 template <
typename Dispatch,
typename DescriptorSetAllocator>
3997 Device::allocateDescriptorSetsUnique(
const VULKAN_HPP_NAMESPACE::DescriptorSetAllocateInfo & allocateInfo, Dispatch
const & d )
const
4000 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4001 VULKAN_HPP_ASSERT( d.vkAllocateDescriptorSets &&
"Function <vkAllocateDescriptorSets> requires <VK_VERSION_1_0>" );
4004 std::vector<VULKAN_HPP_NAMESPACE::DescriptorSet> descriptorSets( allocateInfo.descriptorSetCount );
4008 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::DescriptorSet, Dispatch>, DescriptorSetAllocator> uniqueDescriptorSets;
4009 uniqueDescriptorSets.reserve( allocateInfo.descriptorSetCount );
4011 for (
auto const & descriptorSet : descriptorSets )
4020 typename DescriptorSetAllocator,
4021 typename std::enable_if<std::is_same<typename DescriptorSetAllocator::value_type, UniqueHandle<VULKAN_HPP_NAMESPACE::DescriptorSet, Dispatch>>::value,
4025 Device::allocateDescriptorSetsUnique(
const VULKAN_HPP_NAMESPACE::DescriptorSetAllocateInfo & allocateInfo,
4026 DescriptorSetAllocator & descriptorSetAllocator,
4027 Dispatch
const & d )
const
4030 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4031 VULKAN_HPP_ASSERT( d.vkAllocateDescriptorSets &&
"Function <vkAllocateDescriptorSets> requires <VK_VERSION_1_0>" );
4034 std::vector<VULKAN_HPP_NAMESPACE::DescriptorSet> descriptorSets( allocateInfo.descriptorSetCount );
4038 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::DescriptorSet, Dispatch>, DescriptorSetAllocator> uniqueDescriptorSets( descriptorSetAllocator );
4039 uniqueDescriptorSets.reserve( allocateInfo.descriptorSetCount );
4041 for (
auto const & descriptorSet : descriptorSets )
4050 template <
typename Dispatch>
4052 uint32_t descriptorSetCount,
4053 const VULKAN_HPP_NAMESPACE::DescriptorSet * pDescriptorSets,
4057 return static_cast<Result>( d.vkFreeDescriptorSets(
4061 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4062 template <
typename Dispatch>
4063 VULKAN_HPP_INLINE void Device::freeDescriptorSets( VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool,
4064 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DescriptorSet>
const & descriptorSets,
4068 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4069 VULKAN_HPP_ASSERT( d.vkFreeDescriptorSets &&
"Function <vkFreeDescriptorSets> requires <VK_VERSION_1_0>" );
4072 d.vkFreeDescriptorSets(
4073 m_device,
static_cast<VkDescriptorPool>( descriptorPool ), descriptorSets.size(),
reinterpret_cast<const VkDescriptorSet *
>( descriptorSets.data() ) );
4077 template <
typename Dispatch>
4079 uint32_t descriptorSetCount,
4080 const VULKAN_HPP_NAMESPACE::DescriptorSet * pDescriptorSets,
4084 return static_cast<Result>( d.vkFreeDescriptorSets(
4088 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4089 template <
typename Dispatch>
4090 VULKAN_HPP_INLINE void( Device::free )( VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool,
4091 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DescriptorSet>
const & descriptorSets,
4095 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4096 VULKAN_HPP_ASSERT( d.vkFreeDescriptorSets &&
"Function <vkFreeDescriptorSets> requires <VK_VERSION_1_0>" );
4099 d.vkFreeDescriptorSets(
4100 m_device,
static_cast<VkDescriptorPool>( descriptorPool ), descriptorSets.size(),
reinterpret_cast<const VkDescriptorSet *
>( descriptorSets.data() ) );
4104 template <
typename Dispatch>
4106 const VULKAN_HPP_NAMESPACE::WriteDescriptorSet * pDescriptorWrites,
4107 uint32_t descriptorCopyCount,
4108 const VULKAN_HPP_NAMESPACE::CopyDescriptorSet * pDescriptorCopies,
4112 d.vkUpdateDescriptorSets( m_device,
4113 descriptorWriteCount,
4115 descriptorCopyCount,
4119 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4120 template <
typename Dispatch>
4123 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::CopyDescriptorSet>
const & descriptorCopies,
4127 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4128 VULKAN_HPP_ASSERT( d.vkUpdateDescriptorSets &&
"Function <vkUpdateDescriptorSets> requires <VK_VERSION_1_0>" );
4131 d.vkUpdateDescriptorSets( m_device,
4132 descriptorWrites.size(),
4134 descriptorCopies.size(),
4139 template <
typename Dispatch>
4141 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
4142 VULKAN_HPP_NAMESPACE::Framebuffer * pFramebuffer,
4146 return static_cast<Result>( d.vkCreateFramebuffer( m_device,
4152 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4153 template <
typename Dispatch>
4155 Device::createFramebuffer(
const VULKAN_HPP_NAMESPACE::FramebufferCreateInfo & createInfo,
4157 Dispatch
const & d )
const
4160 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4161 VULKAN_HPP_ASSERT( d.vkCreateFramebuffer &&
"Function <vkCreateFramebuffer> requires <VK_VERSION_1_0>" );
4164 VULKAN_HPP_NAMESPACE::Framebuffer framebuffer;
4166 d.vkCreateFramebuffer( m_device,
4168 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
4175 # ifndef VULKAN_HPP_NO_SMART_HANDLE
4176 template <
typename Dispatch>
4178 Device::createFramebufferUnique(
const VULKAN_HPP_NAMESPACE::FramebufferCreateInfo & createInfo,
4180 Dispatch
const & d )
const
4183 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4184 VULKAN_HPP_ASSERT( d.vkCreateFramebuffer &&
"Function <vkCreateFramebuffer> requires <VK_VERSION_1_0>" );
4187 VULKAN_HPP_NAMESPACE::Framebuffer framebuffer;
4189 d.vkCreateFramebuffer( m_device,
4191 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
4201 template <
typename Dispatch>
4203 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
4210 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4211 template <
typename Dispatch>
4217 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4218 VULKAN_HPP_ASSERT( d.vkDestroyFramebuffer &&
"Function <vkDestroyFramebuffer> requires <VK_VERSION_1_0>" );
4221 d.vkDestroyFramebuffer( m_device,
4223 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
4227 template <
typename Dispatch>
4229 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
4236 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4237 template <
typename Dispatch>
4243 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4244 VULKAN_HPP_ASSERT( d.vkDestroyFramebuffer &&
"Function <vkDestroyFramebuffer> requires <VK_VERSION_1_0>" );
4247 d.vkDestroyFramebuffer( m_device,
4249 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
4253 template <
typename Dispatch>
4255 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
4256 VULKAN_HPP_NAMESPACE::RenderPass * pRenderPass,
4260 return static_cast<Result>( d.vkCreateRenderPass( m_device,
4266 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4267 template <
typename Dispatch>
4271 Dispatch
const & d )
const
4274 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4275 VULKAN_HPP_ASSERT( d.vkCreateRenderPass &&
"Function <vkCreateRenderPass> requires <VK_VERSION_1_0>" );
4278 VULKAN_HPP_NAMESPACE::RenderPass renderPass;
4280 d.vkCreateRenderPass( m_device,
4282 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
4289 # ifndef VULKAN_HPP_NO_SMART_HANDLE
4290 template <
typename Dispatch>
4292 Device::createRenderPassUnique(
const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo & createInfo,
4294 Dispatch
const & d )
const
4297 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4298 VULKAN_HPP_ASSERT( d.vkCreateRenderPass &&
"Function <vkCreateRenderPass> requires <VK_VERSION_1_0>" );
4301 VULKAN_HPP_NAMESPACE::RenderPass renderPass;
4303 d.vkCreateRenderPass( m_device,
4305 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
4315 template <
typename Dispatch>
4317 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
4324 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4325 template <
typename Dispatch>
4331 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4332 VULKAN_HPP_ASSERT( d.vkDestroyRenderPass &&
"Function <vkDestroyRenderPass> requires <VK_VERSION_1_0>" );
4335 d.vkDestroyRenderPass( m_device,
4337 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
4341 template <
typename Dispatch>
4343 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
4350 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4351 template <
typename Dispatch>
4357 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4358 VULKAN_HPP_ASSERT( d.vkDestroyRenderPass &&
"Function <vkDestroyRenderPass> requires <VK_VERSION_1_0>" );
4361 d.vkDestroyRenderPass( m_device,
4363 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
4367 template <
typename Dispatch>
4368 VULKAN_HPP_INLINE void Device::getRenderAreaGranularity( VULKAN_HPP_NAMESPACE::RenderPass renderPass,
4369 VULKAN_HPP_NAMESPACE::Extent2D * pGranularity,
4373 d.vkGetRenderAreaGranularity( m_device,
static_cast<VkRenderPass>( renderPass ),
reinterpret_cast<VkExtent2D *
>( pGranularity ) );
4376 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4377 template <
typename Dispatch>
4382 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4383 VULKAN_HPP_ASSERT( d.vkGetRenderAreaGranularity &&
"Function <vkGetRenderAreaGranularity> requires <VK_VERSION_1_0>" );
4386 VULKAN_HPP_NAMESPACE::Extent2D granularity;
4387 d.vkGetRenderAreaGranularity( m_device,
static_cast<VkRenderPass>( renderPass ),
reinterpret_cast<VkExtent2D *
>( &granularity ) );
4393 template <
typename Dispatch>
4395 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
4396 VULKAN_HPP_NAMESPACE::CommandPool * pCommandPool,
4400 return static_cast<Result>( d.vkCreateCommandPool( m_device,
4406 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4407 template <
typename Dispatch>
4409 Device::createCommandPool(
const VULKAN_HPP_NAMESPACE::CommandPoolCreateInfo & createInfo,
4411 Dispatch
const & d )
const
4414 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4415 VULKAN_HPP_ASSERT( d.vkCreateCommandPool &&
"Function <vkCreateCommandPool> requires <VK_VERSION_1_0>" );
4418 VULKAN_HPP_NAMESPACE::CommandPool commandPool;
4420 d.vkCreateCommandPool( m_device,
4422 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
4429 # ifndef VULKAN_HPP_NO_SMART_HANDLE
4430 template <
typename Dispatch>
4432 Device::createCommandPoolUnique(
const VULKAN_HPP_NAMESPACE::CommandPoolCreateInfo & createInfo,
4434 Dispatch
const & d )
const
4437 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4438 VULKAN_HPP_ASSERT( d.vkCreateCommandPool &&
"Function <vkCreateCommandPool> requires <VK_VERSION_1_0>" );
4441 VULKAN_HPP_NAMESPACE::CommandPool commandPool;
4443 d.vkCreateCommandPool( m_device,
4445 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
4455 template <
typename Dispatch>
4457 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
4464 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4465 template <
typename Dispatch>
4471 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4472 VULKAN_HPP_ASSERT( d.vkDestroyCommandPool &&
"Function <vkDestroyCommandPool> requires <VK_VERSION_1_0>" );
4475 d.vkDestroyCommandPool( m_device,
4477 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
4481 template <
typename Dispatch>
4483 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
4490 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4491 template <
typename Dispatch>
4497 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4498 VULKAN_HPP_ASSERT( d.vkDestroyCommandPool &&
"Function <vkDestroyCommandPool> requires <VK_VERSION_1_0>" );
4501 d.vkDestroyCommandPool( m_device,
4503 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
4507 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
4508 template <
typename Dispatch>
4517 template <
typename Dispatch>
4522 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4523 VULKAN_HPP_ASSERT( d.vkResetCommandPool &&
"Function <vkResetCommandPool> requires <VK_VERSION_1_0>" );
4534 template <
typename Dispatch>
4536 VULKAN_HPP_NAMESPACE::CommandBuffer * pCommandBuffers,
4540 return static_cast<Result>( d.vkAllocateCommandBuffers(
4544 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4545 template <
typename CommandBufferAllocator,
typename Dispatch>
4547 Device::allocateCommandBuffers(
const VULKAN_HPP_NAMESPACE::CommandBufferAllocateInfo & allocateInfo, Dispatch
const & d )
const
4550 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4551 VULKAN_HPP_ASSERT( d.vkAllocateCommandBuffers &&
"Function <vkAllocateCommandBuffers> requires <VK_VERSION_1_0>" );
4554 std::vector<VULKAN_HPP_NAMESPACE::CommandBuffer, CommandBufferAllocator> commandBuffers( allocateInfo.commandBufferCount );
4562 template <
typename CommandBufferAllocator,
4564 typename std::enable_if<std::is_same<typename CommandBufferAllocator::value_type, VULKAN_HPP_NAMESPACE::CommandBuffer>::value,
int>::type>
4566 Device::allocateCommandBuffers(
const VULKAN_HPP_NAMESPACE::CommandBufferAllocateInfo & allocateInfo,
4567 CommandBufferAllocator & commandBufferAllocator,
4568 Dispatch
const & d )
const
4571 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4572 VULKAN_HPP_ASSERT( d.vkAllocateCommandBuffers &&
"Function <vkAllocateCommandBuffers> requires <VK_VERSION_1_0>" );
4575 std::vector<VULKAN_HPP_NAMESPACE::CommandBuffer, CommandBufferAllocator> commandBuffers( allocateInfo.commandBufferCount, commandBufferAllocator );
4583 # ifndef VULKAN_HPP_NO_SMART_HANDLE
4584 template <
typename Dispatch,
typename CommandBufferAllocator>
4587 Device::allocateCommandBuffersUnique(
const VULKAN_HPP_NAMESPACE::CommandBufferAllocateInfo & allocateInfo, Dispatch
const & d )
const
4590 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4591 VULKAN_HPP_ASSERT( d.vkAllocateCommandBuffers &&
"Function <vkAllocateCommandBuffers> requires <VK_VERSION_1_0>" );
4594 std::vector<VULKAN_HPP_NAMESPACE::CommandBuffer> commandBuffers( allocateInfo.commandBufferCount );
4598 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::CommandBuffer, Dispatch>, CommandBufferAllocator> uniqueCommandBuffers;
4599 uniqueCommandBuffers.reserve( allocateInfo.commandBufferCount );
4601 for (
auto const & commandBuffer : commandBuffers )
4610 typename CommandBufferAllocator,
4611 typename std::enable_if<std::is_same<typename CommandBufferAllocator::value_type, UniqueHandle<VULKAN_HPP_NAMESPACE::CommandBuffer, Dispatch>>::value,
4615 Device::allocateCommandBuffersUnique(
const VULKAN_HPP_NAMESPACE::CommandBufferAllocateInfo & allocateInfo,
4616 CommandBufferAllocator & commandBufferAllocator,
4617 Dispatch
const & d )
const
4620 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4621 VULKAN_HPP_ASSERT( d.vkAllocateCommandBuffers &&
"Function <vkAllocateCommandBuffers> requires <VK_VERSION_1_0>" );
4624 std::vector<VULKAN_HPP_NAMESPACE::CommandBuffer> commandBuffers( allocateInfo.commandBufferCount );
4628 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::CommandBuffer, Dispatch>, CommandBufferAllocator> uniqueCommandBuffers( commandBufferAllocator );
4629 uniqueCommandBuffers.reserve( allocateInfo.commandBufferCount );
4631 for (
auto const & commandBuffer : commandBuffers )
4640 template <
typename Dispatch>
4642 uint32_t commandBufferCount,
4643 const VULKAN_HPP_NAMESPACE::CommandBuffer * pCommandBuffers,
4647 d.vkFreeCommandBuffers(
4648 m_device,
static_cast<VkCommandPool>( commandPool ), commandBufferCount,
reinterpret_cast<const VkCommandBuffer *
>( pCommandBuffers ) );
4651 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4652 template <
typename Dispatch>
4654 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::CommandBuffer>
const & commandBuffers,
4658 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4659 VULKAN_HPP_ASSERT( d.vkFreeCommandBuffers &&
"Function <vkFreeCommandBuffers> requires <VK_VERSION_1_0>" );
4662 d.vkFreeCommandBuffers(
4663 m_device,
static_cast<VkCommandPool>( commandPool ), commandBuffers.size(),
reinterpret_cast<const VkCommandBuffer *
>( commandBuffers.data() ) );
4667 template <
typename Dispatch>
4668 VULKAN_HPP_INLINE void( Device::free )( VULKAN_HPP_NAMESPACE::CommandPool commandPool,
4669 uint32_t commandBufferCount,
4670 const VULKAN_HPP_NAMESPACE::CommandBuffer * pCommandBuffers,
4674 d.vkFreeCommandBuffers(
4675 m_device,
static_cast<VkCommandPool>( commandPool ), commandBufferCount,
reinterpret_cast<const VkCommandBuffer *
>( pCommandBuffers ) );
4678 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4679 template <
typename Dispatch>
4680 VULKAN_HPP_INLINE void( Device::free )( VULKAN_HPP_NAMESPACE::CommandPool commandPool,
4681 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::CommandBuffer>
const & commandBuffers,
4685 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4686 VULKAN_HPP_ASSERT( d.vkFreeCommandBuffers &&
"Function <vkFreeCommandBuffers> requires <VK_VERSION_1_0>" );
4689 d.vkFreeCommandBuffers(
4690 m_device,
static_cast<VkCommandPool>( commandPool ), commandBuffers.size(),
reinterpret_cast<const VkCommandBuffer *
>( commandBuffers.data() ) );
4694 template <
typename Dispatch>
4702 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4703 template <
typename Dispatch>
4705 CommandBuffer::begin(
const VULKAN_HPP_NAMESPACE::CommandBufferBeginInfo & beginInfo, Dispatch
const & d )
const
4708 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4709 VULKAN_HPP_ASSERT( d.vkBeginCommandBuffer &&
"Function <vkBeginCommandBuffer> requires <VK_VERSION_1_0>" );
4720 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
4721 template <
typename Dispatch>
4725 return static_cast<Result>( d.vkEndCommandBuffer( m_commandBuffer ) );
4728 template <
typename Dispatch>
4732 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4733 VULKAN_HPP_ASSERT( d.vkEndCommandBuffer &&
"Function <vkEndCommandBuffer> requires <VK_VERSION_1_0>" );
4743 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
4744 template <
typename Dispatch>
4752 template <
typename Dispatch>
4756 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4757 VULKAN_HPP_ASSERT( d.vkResetCommandBuffer &&
"Function <vkResetCommandBuffer> requires <VK_VERSION_1_0>" );
4768 template <
typename Dispatch>
4770 VULKAN_HPP_NAMESPACE::Pipeline pipeline,
4777 template <
typename Dispatch>
4779 uint32_t viewportCount,
4780 const VULKAN_HPP_NAMESPACE::Viewport * pViewports,
4784 d.vkCmdSetViewport( m_commandBuffer, firstViewport, viewportCount,
reinterpret_cast<const VkViewport *
>( pViewports ) );
4787 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4788 template <
typename Dispatch>
4790 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Viewport>
const & viewports,
4794 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4795 VULKAN_HPP_ASSERT( d.vkCmdSetViewport &&
"Function <vkCmdSetViewport> requires <VK_VERSION_1_0>" );
4798 d.vkCmdSetViewport( m_commandBuffer, firstViewport, viewports.size(),
reinterpret_cast<const VkViewport *
>( viewports.data() ) );
4802 template <
typename Dispatch>
4804 uint32_t scissorCount,
4805 const VULKAN_HPP_NAMESPACE::Rect2D * pScissors,
4809 d.vkCmdSetScissor( m_commandBuffer, firstScissor, scissorCount,
reinterpret_cast<const VkRect2D *
>( pScissors ) );
4812 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4813 template <
typename Dispatch>
4815 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D>
const & scissors,
4819 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4820 VULKAN_HPP_ASSERT( d.vkCmdSetScissor &&
"Function <vkCmdSetScissor> requires <VK_VERSION_1_0>" );
4823 d.vkCmdSetScissor( m_commandBuffer, firstScissor, scissors.size(),
reinterpret_cast<const VkRect2D *
>( scissors.data() ) );
4827 template <
typename Dispatch>
4831 d.vkCmdSetLineWidth( m_commandBuffer, lineWidth );
4834 template <
typename Dispatch>
4836 CommandBuffer::setDepthBias(
float depthBiasConstantFactor,
float depthBiasClamp,
float depthBiasSlopeFactor, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
4839 d.vkCmdSetDepthBias( m_commandBuffer, depthBiasConstantFactor, depthBiasClamp, depthBiasSlopeFactor );
4842 template <
typename Dispatch>
4846 d.vkCmdSetBlendConstants( m_commandBuffer, blendConstants );
4849 template <
typename Dispatch>
4853 d.vkCmdSetDepthBounds( m_commandBuffer, minDepthBounds, maxDepthBounds );
4856 template <
typename Dispatch>
4861 d.vkCmdSetStencilCompareMask( m_commandBuffer,
static_cast<VkStencilFaceFlags>( faceMask ), compareMask );
4864 template <
typename Dispatch>
4869 d.vkCmdSetStencilWriteMask( m_commandBuffer,
static_cast<VkStencilFaceFlags>( faceMask ), writeMask );
4872 template <
typename Dispatch>
4877 d.vkCmdSetStencilReference( m_commandBuffer,
static_cast<VkStencilFaceFlags>( faceMask ), reference );
4880 template <
typename Dispatch>
4882 VULKAN_HPP_NAMESPACE::PipelineLayout layout,
4884 uint32_t descriptorSetCount,
4885 const VULKAN_HPP_NAMESPACE::DescriptorSet * pDescriptorSets,
4886 uint32_t dynamicOffsetCount,
4887 const uint32_t * pDynamicOffsets,
4891 d.vkCmdBindDescriptorSets( m_commandBuffer,
4901 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4902 template <
typename Dispatch>
4904 VULKAN_HPP_NAMESPACE::PipelineLayout layout,
4906 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DescriptorSet>
const & descriptorSets,
4907 VULKAN_HPP_NAMESPACE::ArrayProxy<const uint32_t>
const & dynamicOffsets,
4911 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4912 VULKAN_HPP_ASSERT( d.vkCmdBindDescriptorSets &&
"Function <vkCmdBindDescriptorSets> requires <VK_VERSION_1_0>" );
4915 d.vkCmdBindDescriptorSets( m_commandBuffer,
4919 descriptorSets.size(),
4921 dynamicOffsets.size(),
4922 dynamicOffsets.data() );
4926 template <
typename Dispatch>
4933 d.vkCmdBindIndexBuffer( m_commandBuffer,
static_cast<VkBuffer>( buffer ),
static_cast<VkDeviceSize>( offset ),
static_cast<VkIndexType>( indexType ) );
4936 template <
typename Dispatch>
4938 uint32_t bindingCount,
4939 const VULKAN_HPP_NAMESPACE::Buffer * pBuffers,
4944 d.vkCmdBindVertexBuffers(
4945 m_commandBuffer, firstBinding, bindingCount,
reinterpret_cast<const VkBuffer *
>( pBuffers ),
reinterpret_cast<const VkDeviceSize *
>( pOffsets ) );
4948 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
4949 template <
typename Dispatch>
4951 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer>
const & buffers,
4952 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & offsets,
4956 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
4957 VULKAN_HPP_ASSERT( d.vkCmdBindVertexBuffers &&
"Function <vkCmdBindVertexBuffers> requires <VK_VERSION_1_0>" );
4959 # ifdef VULKAN_HPP_NO_EXCEPTIONS
4962 if ( buffers.size() != offsets.size() )
4968 d.vkCmdBindVertexBuffers( m_commandBuffer,
4971 reinterpret_cast<const VkBuffer *
>( buffers.data() ),
4972 reinterpret_cast<const VkDeviceSize *
>( offsets.data() ) );
4976 template <
typename Dispatch>
4978 uint32_t vertexCount, uint32_t instanceCount, uint32_t firstVertex, uint32_t firstInstance, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
4981 d.vkCmdDraw( m_commandBuffer, vertexCount, instanceCount, firstVertex, firstInstance );
4984 template <
typename Dispatch>
4986 uint32_t instanceCount,
4987 uint32_t firstIndex,
4988 int32_t vertexOffset,
4989 uint32_t firstInstance,
4993 d.vkCmdDrawIndexed( m_commandBuffer, indexCount, instanceCount, firstIndex, vertexOffset, firstInstance );
4996 template <
typename Dispatch>
5004 d.vkCmdDrawIndirect( m_commandBuffer,
static_cast<VkBuffer>( buffer ),
static_cast<VkDeviceSize>( offset ), drawCount, stride );
5007 template <
typename Dispatch>
5015 d.vkCmdDrawIndexedIndirect( m_commandBuffer,
static_cast<VkBuffer>( buffer ),
static_cast<VkDeviceSize>( offset ), drawCount, stride );
5018 template <
typename Dispatch>
5020 CommandBuffer::dispatch( uint32_t groupCountX, uint32_t groupCountY, uint32_t groupCountZ, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
5023 d.vkCmdDispatch( m_commandBuffer, groupCountX, groupCountY, groupCountZ );
5026 template <
typename Dispatch>
5032 d.vkCmdDispatchIndirect( m_commandBuffer,
static_cast<VkBuffer>( buffer ),
static_cast<VkDeviceSize>( offset ) );
5035 template <
typename Dispatch>
5037 VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
5038 uint32_t regionCount,
5039 const VULKAN_HPP_NAMESPACE::BufferCopy * pRegions,
5043 d.vkCmdCopyBuffer( m_commandBuffer,
5044 static_cast<VkBuffer>( srcBuffer ),
5045 static_cast<VkBuffer>( dstBuffer ),
5050 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5051 template <
typename Dispatch>
5053 VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
5054 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferCopy>
const & regions,
5058 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5059 VULKAN_HPP_ASSERT( d.vkCmdCopyBuffer &&
"Function <vkCmdCopyBuffer> requires <VK_VERSION_1_0>" );
5062 d.vkCmdCopyBuffer( m_commandBuffer,
5063 static_cast<VkBuffer>( srcBuffer ),
5064 static_cast<VkBuffer>( dstBuffer ),
5066 reinterpret_cast<const VkBufferCopy *
>( regions.data() ) );
5070 template <
typename Dispatch>
5073 VULKAN_HPP_NAMESPACE::Image dstImage,
5075 uint32_t regionCount,
5076 const VULKAN_HPP_NAMESPACE::ImageCopy * pRegions,
5080 d.vkCmdCopyImage( m_commandBuffer,
5081 static_cast<VkImage>( srcImage ),
5083 static_cast<VkImage>( dstImage ),
5086 reinterpret_cast<const VkImageCopy *
>( pRegions ) );
5089 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5090 template <
typename Dispatch>
5093 VULKAN_HPP_NAMESPACE::Image dstImage,
5095 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageCopy>
const & regions,
5099 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5100 VULKAN_HPP_ASSERT( d.vkCmdCopyImage &&
"Function <vkCmdCopyImage> requires <VK_VERSION_1_0>" );
5103 d.vkCmdCopyImage( m_commandBuffer,
5104 static_cast<VkImage>( srcImage ),
5106 static_cast<VkImage>( dstImage ),
5109 reinterpret_cast<const VkImageCopy *
>( regions.data() ) );
5113 template <
typename Dispatch>
5116 VULKAN_HPP_NAMESPACE::Image dstImage,
5118 uint32_t regionCount,
5119 const VULKAN_HPP_NAMESPACE::ImageBlit * pRegions,
5124 d.vkCmdBlitImage( m_commandBuffer,
5125 static_cast<VkImage>( srcImage ),
5127 static_cast<VkImage>( dstImage ),
5130 reinterpret_cast<const VkImageBlit *
>( pRegions ),
5134 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5135 template <
typename Dispatch>
5138 VULKAN_HPP_NAMESPACE::Image dstImage,
5140 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageBlit>
const & regions,
5145 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5146 VULKAN_HPP_ASSERT( d.vkCmdBlitImage &&
"Function <vkCmdBlitImage> requires <VK_VERSION_1_0>" );
5149 d.vkCmdBlitImage( m_commandBuffer,
5150 static_cast<VkImage>( srcImage ),
5152 static_cast<VkImage>( dstImage ),
5155 reinterpret_cast<const VkImageBlit *
>( regions.data() ),
5160 template <
typename Dispatch>
5162 VULKAN_HPP_NAMESPACE::Image dstImage,
5164 uint32_t regionCount,
5165 const VULKAN_HPP_NAMESPACE::BufferImageCopy * pRegions,
5169 d.vkCmdCopyBufferToImage( m_commandBuffer,
5170 static_cast<VkBuffer>( srcBuffer ),
5171 static_cast<VkImage>( dstImage ),
5177 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5178 template <
typename Dispatch>
5180 VULKAN_HPP_NAMESPACE::Image dstImage,
5182 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferImageCopy>
const & regions,
5186 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5187 VULKAN_HPP_ASSERT( d.vkCmdCopyBufferToImage &&
"Function <vkCmdCopyBufferToImage> requires <VK_VERSION_1_0>" );
5190 d.vkCmdCopyBufferToImage( m_commandBuffer,
5191 static_cast<VkBuffer>( srcBuffer ),
5192 static_cast<VkImage>( dstImage ),
5199 template <
typename Dispatch>
5202 VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
5203 uint32_t regionCount,
5204 const VULKAN_HPP_NAMESPACE::BufferImageCopy * pRegions,
5208 d.vkCmdCopyImageToBuffer( m_commandBuffer,
5209 static_cast<VkImage>( srcImage ),
5211 static_cast<VkBuffer>( dstBuffer ),
5216 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5217 template <
typename Dispatch>
5220 VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
5221 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferImageCopy>
const & regions,
5225 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5226 VULKAN_HPP_ASSERT( d.vkCmdCopyImageToBuffer &&
"Function <vkCmdCopyImageToBuffer> requires <VK_VERSION_1_0>" );
5229 d.vkCmdCopyImageToBuffer( m_commandBuffer,
5230 static_cast<VkImage>( srcImage ),
5232 static_cast<VkBuffer>( dstBuffer ),
5238 template <
typename Dispatch>
5246 d.vkCmdUpdateBuffer(
5250 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5251 template <
typename DataType,
typename Dispatch>
5254 VULKAN_HPP_NAMESPACE::ArrayProxy<const DataType>
const & data,
5258 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5259 VULKAN_HPP_ASSERT( d.vkCmdUpdateBuffer &&
"Function <vkCmdUpdateBuffer> requires <VK_VERSION_1_0>" );
5262 d.vkCmdUpdateBuffer( m_commandBuffer,
5263 static_cast<VkBuffer>( dstBuffer ),
5265 data.size() *
sizeof( DataType ),
5266 reinterpret_cast<const void *
>( data.data() ) );
5270 template <
typename Dispatch>
5281 template <
typename Dispatch>
5284 const VULKAN_HPP_NAMESPACE::ClearColorValue * pColor,
5285 uint32_t rangeCount,
5286 const VULKAN_HPP_NAMESPACE::ImageSubresourceRange * pRanges,
5290 d.vkCmdClearColorImage( m_commandBuffer,
5291 static_cast<VkImage>( image ),
5298 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5299 template <
typename Dispatch>
5302 const VULKAN_HPP_NAMESPACE::ClearColorValue & color,
5303 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageSubresourceRange>
const & ranges,
5307 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5308 VULKAN_HPP_ASSERT( d.vkCmdClearColorImage &&
"Function <vkCmdClearColorImage> requires <VK_VERSION_1_0>" );
5311 d.vkCmdClearColorImage( m_commandBuffer,
5312 static_cast<VkImage>( image ),
5320 template <
typename Dispatch>
5323 const VULKAN_HPP_NAMESPACE::ClearDepthStencilValue * pDepthStencil,
5324 uint32_t rangeCount,
5325 const VULKAN_HPP_NAMESPACE::ImageSubresourceRange * pRanges,
5329 d.vkCmdClearDepthStencilImage( m_commandBuffer,
5330 static_cast<VkImage>( image ),
5337 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5338 template <
typename Dispatch>
5340 CommandBuffer::clearDepthStencilImage( VULKAN_HPP_NAMESPACE::Image image,
5342 const VULKAN_HPP_NAMESPACE::ClearDepthStencilValue & depthStencil,
5343 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageSubresourceRange>
const & ranges,
5347 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5348 VULKAN_HPP_ASSERT( d.vkCmdClearDepthStencilImage &&
"Function <vkCmdClearDepthStencilImage> requires <VK_VERSION_1_0>" );
5351 d.vkCmdClearDepthStencilImage( m_commandBuffer,
5352 static_cast<VkImage>( image ),
5360 template <
typename Dispatch>
5362 const VULKAN_HPP_NAMESPACE::ClearAttachment * pAttachments,
5364 const VULKAN_HPP_NAMESPACE::ClearRect * pRects,
5368 d.vkCmdClearAttachments( m_commandBuffer,
5372 reinterpret_cast<const VkClearRect *
>( pRects ) );
5375 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5376 template <
typename Dispatch>
5377 VULKAN_HPP_INLINE void CommandBuffer::clearAttachments( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ClearAttachment>
const & attachments,
5378 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ClearRect>
const & rects,
5382 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5383 VULKAN_HPP_ASSERT( d.vkCmdClearAttachments &&
"Function <vkCmdClearAttachments> requires <VK_VERSION_1_0>" );
5386 d.vkCmdClearAttachments( m_commandBuffer,
5390 reinterpret_cast<const VkClearRect *
>( rects.data() ) );
5394 template <
typename Dispatch>
5397 VULKAN_HPP_NAMESPACE::Image dstImage,
5399 uint32_t regionCount,
5400 const VULKAN_HPP_NAMESPACE::ImageResolve * pRegions,
5404 d.vkCmdResolveImage( m_commandBuffer,
5405 static_cast<VkImage>( srcImage ),
5407 static_cast<VkImage>( dstImage ),
5413 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5414 template <
typename Dispatch>
5417 VULKAN_HPP_NAMESPACE::Image dstImage,
5419 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageResolve>
const & regions,
5423 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5424 VULKAN_HPP_ASSERT( d.vkCmdResolveImage &&
"Function <vkCmdResolveImage> requires <VK_VERSION_1_0>" );
5427 d.vkCmdResolveImage( m_commandBuffer,
5428 static_cast<VkImage>( srcImage ),
5430 static_cast<VkImage>( dstImage ),
5437 template <
typename Dispatch>
5446 template <
typename Dispatch>
5455 template <
typename Dispatch>
5457 const VULKAN_HPP_NAMESPACE::Event * pEvents,
5460 uint32_t memoryBarrierCount,
5461 const VULKAN_HPP_NAMESPACE::MemoryBarrier * pMemoryBarriers,
5462 uint32_t bufferMemoryBarrierCount,
5463 const VULKAN_HPP_NAMESPACE::BufferMemoryBarrier * pBufferMemoryBarriers,
5464 uint32_t imageMemoryBarrierCount,
5465 const VULKAN_HPP_NAMESPACE::ImageMemoryBarrier * pImageMemoryBarriers,
5469 d.vkCmdWaitEvents( m_commandBuffer,
5471 reinterpret_cast<const VkEvent *
>( pEvents ),
5476 bufferMemoryBarrierCount,
5478 imageMemoryBarrierCount,
5482 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5483 template <
typename Dispatch>
5485 CommandBuffer::waitEvents( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Event>
const & events,
5488 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::MemoryBarrier>
const & memoryBarriers,
5489 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferMemoryBarrier>
const & bufferMemoryBarriers,
5490 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageMemoryBarrier>
const & imageMemoryBarriers,
5494 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5495 VULKAN_HPP_ASSERT( d.vkCmdWaitEvents &&
"Function <vkCmdWaitEvents> requires <VK_VERSION_1_0>" );
5498 d.vkCmdWaitEvents( m_commandBuffer,
5500 reinterpret_cast<const VkEvent *
>( events.data() ),
5503 memoryBarriers.size(),
5505 bufferMemoryBarriers.size(),
5507 imageMemoryBarriers.
size(),
5512 template <
typename Dispatch>
5516 uint32_t memoryBarrierCount,
5517 const VULKAN_HPP_NAMESPACE::MemoryBarrier * pMemoryBarriers,
5518 uint32_t bufferMemoryBarrierCount,
5519 const VULKAN_HPP_NAMESPACE::BufferMemoryBarrier * pBufferMemoryBarriers,
5520 uint32_t imageMemoryBarrierCount,
5521 const VULKAN_HPP_NAMESPACE::ImageMemoryBarrier * pImageMemoryBarriers,
5525 d.vkCmdPipelineBarrier( m_commandBuffer,
5531 bufferMemoryBarrierCount,
5533 imageMemoryBarrierCount,
5537 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5538 template <
typename Dispatch>
5543 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::MemoryBarrier>
const & memoryBarriers,
5544 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferMemoryBarrier>
const & bufferMemoryBarriers,
5545 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageMemoryBarrier>
const & imageMemoryBarriers,
5549 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5550 VULKAN_HPP_ASSERT( d.vkCmdPipelineBarrier &&
"Function <vkCmdPipelineBarrier> requires <VK_VERSION_1_0>" );
5553 d.vkCmdPipelineBarrier( m_commandBuffer,
5557 memoryBarriers.size(),
5559 bufferMemoryBarriers.size(),
5561 imageMemoryBarriers.
size(),
5566 template <
typename Dispatch>
5576 template <
typename Dispatch>
5580 d.vkCmdEndQuery( m_commandBuffer,
static_cast<VkQueryPool>( queryPool ), query );
5583 template <
typename Dispatch>
5585 uint32_t firstQuery,
5586 uint32_t queryCount,
5590 d.vkCmdResetQueryPool( m_commandBuffer,
static_cast<VkQueryPool>( queryPool ), firstQuery, queryCount );
5593 template <
typename Dispatch>
5595 VULKAN_HPP_NAMESPACE::QueryPool queryPool,
5603 template <
typename Dispatch>
5604 VULKAN_HPP_INLINE void CommandBuffer::copyQueryPoolResults( VULKAN_HPP_NAMESPACE::QueryPool queryPool,
5605 uint32_t firstQuery,
5606 uint32_t queryCount,
5607 VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
5614 d.vkCmdCopyQueryPoolResults( m_commandBuffer,
5618 static_cast<VkBuffer>( dstBuffer ),
5624 template <
typename Dispatch>
5629 const void * pValues,
5636 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5637 template <
typename ValuesType,
typename Dispatch>
5641 VULKAN_HPP_NAMESPACE::ArrayProxy<const ValuesType>
const & values,
5645 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5646 VULKAN_HPP_ASSERT( d.vkCmdPushConstants &&
"Function <vkCmdPushConstants> requires <VK_VERSION_1_0>" );
5649 d.vkCmdPushConstants( m_commandBuffer,
5653 values.size() *
sizeof( ValuesType ),
5654 reinterpret_cast<const void *
>( values.data() ) );
5658 template <
typename Dispatch>
5659 VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass(
const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo * pRenderPassBegin,
5667 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5668 template <
typename Dispatch>
5669 VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass(
const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo & renderPassBegin,
5674 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5675 VULKAN_HPP_ASSERT( d.vkCmdBeginRenderPass &&
"Function <vkCmdBeginRenderPass> requires <VK_VERSION_1_0>" );
5682 template <
typename Dispatch>
5686 d.vkCmdNextSubpass( m_commandBuffer,
static_cast<VkSubpassContents>( contents ) );
5689 template <
typename Dispatch>
5693 d.vkCmdEndRenderPass( m_commandBuffer );
5696 template <
typename Dispatch>
5698 const VULKAN_HPP_NAMESPACE::CommandBuffer * pCommandBuffers,
5702 d.vkCmdExecuteCommands( m_commandBuffer, commandBufferCount,
reinterpret_cast<const VkCommandBuffer *
>( pCommandBuffers ) );
5705 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5706 template <
typename Dispatch>
5707 VULKAN_HPP_INLINE void CommandBuffer::executeCommands( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::CommandBuffer>
const & commandBuffers,
5711 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5712 VULKAN_HPP_ASSERT( d.vkCmdExecuteCommands &&
"Function <vkCmdExecuteCommands> requires <VK_VERSION_1_0>" );
5715 d.vkCmdExecuteCommands( m_commandBuffer, commandBuffers.size(),
reinterpret_cast<const VkCommandBuffer *
>( commandBuffers.data() ) );
5721 template <
typename Dispatch>
5725 return static_cast<Result>( d.vkEnumerateInstanceVersion( pApiVersion ) );
5728 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5729 template <
typename Dispatch>
5733 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5734 VULKAN_HPP_ASSERT( d.vkEnumerateInstanceVersion &&
"Function <vkEnumerateInstanceVersion> requires <VK_VERSION_1_1>" );
5737 uint32_t apiVersion;
5745 template <
typename Dispatch>
5747 const VULKAN_HPP_NAMESPACE::BindBufferMemoryInfo * pBindInfos,
5751 return static_cast<Result>( d.vkBindBufferMemory2( m_device, bindInfoCount,
reinterpret_cast<const VkBindBufferMemoryInfo *
>( pBindInfos ) ) );
5754 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5755 template <
typename Dispatch>
5757 Device::bindBufferMemory2( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BindBufferMemoryInfo>
const & bindInfos, Dispatch
const & d )
const
5760 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5761 VULKAN_HPP_ASSERT( d.vkBindBufferMemory2 &&
"Function <vkBindBufferMemory2> requires <VK_KHR_bind_memory2> or <VK_VERSION_1_1>" );
5765 d.vkBindBufferMemory2( m_device, bindInfos.size(),
reinterpret_cast<const VkBindBufferMemoryInfo *
>( bindInfos.data() ) ) );
5772 template <
typename Dispatch>
5774 const VULKAN_HPP_NAMESPACE::BindImageMemoryInfo * pBindInfos,
5778 return static_cast<Result>( d.vkBindImageMemory2( m_device, bindInfoCount,
reinterpret_cast<const VkBindImageMemoryInfo *
>( pBindInfos ) ) );
5781 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5782 template <
typename Dispatch>
5784 Device::bindImageMemory2( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BindImageMemoryInfo>
const & bindInfos, Dispatch
const & d )
const
5787 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5788 VULKAN_HPP_ASSERT( d.vkBindImageMemory2 &&
"Function <vkBindImageMemory2> requires <VK_KHR_bind_memory2> or <VK_VERSION_1_1>" );
5792 d.vkBindImageMemory2( m_device, bindInfos.size(),
reinterpret_cast<const VkBindImageMemoryInfo *
>( bindInfos.data() ) ) );
5799 template <
typename Dispatch>
5801 uint32_t localDeviceIndex,
5802 uint32_t remoteDeviceIndex,
5807 d.vkGetDeviceGroupPeerMemoryFeatures(
5808 m_device, heapIndex, localDeviceIndex, remoteDeviceIndex,
reinterpret_cast<VkPeerMemoryFeatureFlags *
>( pPeerMemoryFeatures ) );
5811 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5812 template <
typename Dispatch>
5814 uint32_t heapIndex, uint32_t localDeviceIndex, uint32_t remoteDeviceIndex, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
5817 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5819 "Function <vkGetDeviceGroupPeerMemoryFeatures> requires <VK_KHR_device_group> or <VK_VERSION_1_1>" );
5823 d.vkGetDeviceGroupPeerMemoryFeatures(
5824 m_device, heapIndex, localDeviceIndex, remoteDeviceIndex,
reinterpret_cast<VkPeerMemoryFeatureFlags *
>( &peerMemoryFeatures ) );
5826 return peerMemoryFeatures;
5830 template <
typename Dispatch>
5834 d.vkCmdSetDeviceMask( m_commandBuffer, deviceMask );
5837 template <
typename Dispatch>
5839 uint32_t baseGroupY,
5840 uint32_t baseGroupZ,
5841 uint32_t groupCountX,
5842 uint32_t groupCountY,
5843 uint32_t groupCountZ,
5847 d.vkCmdDispatchBase( m_commandBuffer, baseGroupX, baseGroupY, baseGroupZ, groupCountX, groupCountY, groupCountZ );
5850 template <
typename Dispatch>
5852 Instance::enumeratePhysicalDeviceGroups( uint32_t * pPhysicalDeviceGroupCount,
5853 VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties * pPhysicalDeviceGroupProperties,
5857 return static_cast<Result>( d.vkEnumeratePhysicalDeviceGroups(
5861 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5862 template <
typename PhysicalDeviceGroupPropertiesAllocator,
typename Dispatch>
5865 Instance::enumeratePhysicalDeviceGroups( Dispatch
const & d )
const
5868 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5870 "Function <vkEnumeratePhysicalDeviceGroups> requires <VK_KHR_device_group_creation> or <VK_VERSION_1_1>" );
5873 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator> physicalDeviceGroupProperties;
5874 uint32_t physicalDeviceGroupCount;
5878 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkEnumeratePhysicalDeviceGroups( m_instance, &physicalDeviceGroupCount,
nullptr ) );
5879 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && physicalDeviceGroupCount )
5881 physicalDeviceGroupProperties.resize( physicalDeviceGroupCount );
5885 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
5887 VULKAN_HPP_ASSERT( physicalDeviceGroupCount <= physicalDeviceGroupProperties.size() );
5888 if ( physicalDeviceGroupCount < physicalDeviceGroupProperties.size() )
5890 physicalDeviceGroupProperties.resize( physicalDeviceGroupCount );
5895 template <
typename PhysicalDeviceGroupPropertiesAllocator,
5897 typename std::enable_if<
5898 std::is_same<typename PhysicalDeviceGroupPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties>::value,
5902 Instance::enumeratePhysicalDeviceGroups( PhysicalDeviceGroupPropertiesAllocator & physicalDeviceGroupPropertiesAllocator, Dispatch
const & d )
const
5905 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5907 "Function <vkEnumeratePhysicalDeviceGroups> requires <VK_KHR_device_group_creation> or <VK_VERSION_1_1>" );
5910 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator> physicalDeviceGroupProperties(
5911 physicalDeviceGroupPropertiesAllocator );
5912 uint32_t physicalDeviceGroupCount;
5916 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkEnumeratePhysicalDeviceGroups( m_instance, &physicalDeviceGroupCount,
nullptr ) );
5917 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && physicalDeviceGroupCount )
5919 physicalDeviceGroupProperties.resize( physicalDeviceGroupCount );
5923 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
5925 VULKAN_HPP_ASSERT( physicalDeviceGroupCount <= physicalDeviceGroupProperties.size() );
5926 if ( physicalDeviceGroupCount < physicalDeviceGroupProperties.size() )
5928 physicalDeviceGroupProperties.resize( physicalDeviceGroupCount );
5934 template <
typename Dispatch>
5935 VULKAN_HPP_INLINE void Device::getImageMemoryRequirements2(
const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 * pInfo,
5936 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
5940 d.vkGetImageMemoryRequirements2(
5944 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5945 template <
typename Dispatch>
5947 Device::getImageMemoryRequirements2(
const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
5950 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5952 "Function <vkGetImageMemoryRequirements2> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
5955 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
5956 d.vkGetImageMemoryRequirements2(
5959 return memoryRequirements;
5962 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
5964 Device::getImageMemoryRequirements2(
const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
5967 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5969 "Function <vkGetImageMemoryRequirements2> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
5972 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
5973 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
5974 d.vkGetImageMemoryRequirements2(
5977 return structureChain;
5981 template <
typename Dispatch>
5982 VULKAN_HPP_INLINE void Device::getBufferMemoryRequirements2(
const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 * pInfo,
5983 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
5987 d.vkGetBufferMemoryRequirements2(
5991 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
5992 template <
typename Dispatch>
5994 Device::getBufferMemoryRequirements2(
const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
5997 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
5999 "Function <vkGetBufferMemoryRequirements2> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
6002 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
6003 d.vkGetBufferMemoryRequirements2(
6006 return memoryRequirements;
6009 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
6011 Device::getBufferMemoryRequirements2(
const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
6014 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6016 "Function <vkGetBufferMemoryRequirements2> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
6019 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
6020 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
6021 d.vkGetBufferMemoryRequirements2(
6024 return structureChain;
6028 template <
typename Dispatch>
6029 VULKAN_HPP_INLINE void Device::getImageSparseMemoryRequirements2(
const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 * pInfo,
6030 uint32_t * pSparseMemoryRequirementCount,
6031 VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2 * pSparseMemoryRequirements,
6035 d.vkGetImageSparseMemoryRequirements2( m_device,
6037 pSparseMemoryRequirementCount,
6041 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6042 template <
typename SparseImageMemoryRequirements2Allocator,
typename Dispatch>
6044 Device::getImageSparseMemoryRequirements2(
const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 & info, Dispatch
const & d )
const
6047 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6049 "Function <vkGetImageSparseMemoryRequirements2> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
6052 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements;
6053 uint32_t sparseMemoryRequirementCount;
6054 d.vkGetImageSparseMemoryRequirements2(
6056 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
6057 d.vkGetImageSparseMemoryRequirements2( m_device,
6059 &sparseMemoryRequirementCount,
6062 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
6063 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
6065 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
6067 return sparseMemoryRequirements;
6070 template <
typename SparseImageMemoryRequirements2Allocator,
6072 typename std::enable_if<
6073 std::is_same<typename SparseImageMemoryRequirements2Allocator::value_type, VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2>::value,
6076 Device::getImageSparseMemoryRequirements2(
const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 & info,
6077 SparseImageMemoryRequirements2Allocator & sparseImageMemoryRequirements2Allocator,
6078 Dispatch
const & d )
const
6081 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6083 "Function <vkGetImageSparseMemoryRequirements2> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
6086 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements(
6087 sparseImageMemoryRequirements2Allocator );
6088 uint32_t sparseMemoryRequirementCount;
6089 d.vkGetImageSparseMemoryRequirements2(
6091 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
6092 d.vkGetImageSparseMemoryRequirements2( m_device,
6094 &sparseMemoryRequirementCount,
6097 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
6098 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
6100 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
6102 return sparseMemoryRequirements;
6106 template <
typename Dispatch>
6113 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6114 template <
typename Dispatch>
6119 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6121 "Function <vkGetPhysicalDeviceFeatures2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6124 VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 features;
6130 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
6135 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6137 "Function <vkGetPhysicalDeviceFeatures2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6140 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
6141 VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 & features = structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2>();
6144 return structureChain;
6148 template <
typename Dispatch>
6149 VULKAN_HPP_INLINE void PhysicalDevice::getProperties2( VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 * pProperties,
6156 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6157 template <
typename Dispatch>
6162 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6164 "Function <vkGetPhysicalDeviceProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6167 VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 properties;
6173 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
6178 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6180 "Function <vkGetPhysicalDeviceProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6183 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
6184 VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 & properties = structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2>();
6187 return structureChain;
6191 template <
typename Dispatch>
6193 VULKAN_HPP_NAMESPACE::FormatProperties2 * pFormatProperties,
6197 d.vkGetPhysicalDeviceFormatProperties2( m_physicalDevice,
static_cast<VkFormat>( format ),
reinterpret_cast<VkFormatProperties2 *
>( pFormatProperties ) );
6200 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6201 template <
typename Dispatch>
6206 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6208 "Function <vkGetPhysicalDeviceFormatProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6211 VULKAN_HPP_NAMESPACE::FormatProperties2 formatProperties;
6212 d.vkGetPhysicalDeviceFormatProperties2( m_physicalDevice,
static_cast<VkFormat>( format ),
reinterpret_cast<VkFormatProperties2 *
>( &formatProperties ) );
6214 return formatProperties;
6217 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
6222 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6224 "Function <vkGetPhysicalDeviceFormatProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6227 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
6228 VULKAN_HPP_NAMESPACE::FormatProperties2 & formatProperties = structureChain.template get<VULKAN_HPP_NAMESPACE::FormatProperties2>();
6229 d.vkGetPhysicalDeviceFormatProperties2( m_physicalDevice,
static_cast<VkFormat>( format ),
reinterpret_cast<VkFormatProperties2 *
>( &formatProperties ) );
6231 return structureChain;
6235 template <
typename Dispatch>
6237 PhysicalDevice::getImageFormatProperties2(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 * pImageFormatInfo,
6238 VULKAN_HPP_NAMESPACE::ImageFormatProperties2 * pImageFormatProperties,
6242 return static_cast<Result>( d.vkGetPhysicalDeviceImageFormatProperties2( m_physicalDevice,
6247 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6248 template <
typename Dispatch>
6250 PhysicalDevice::getImageFormatProperties2(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 & imageFormatInfo, Dispatch
const & d )
const
6253 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6255 "Function <vkGetPhysicalDeviceImageFormatProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6258 VULKAN_HPP_NAMESPACE::ImageFormatProperties2 imageFormatProperties;
6260 d.vkGetPhysicalDeviceImageFormatProperties2( m_physicalDevice,
6268 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
6270 PhysicalDevice::getImageFormatProperties2(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 & imageFormatInfo, Dispatch
const & d )
const
6273 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6275 "Function <vkGetPhysicalDeviceImageFormatProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6279 VULKAN_HPP_NAMESPACE::ImageFormatProperties2 & imageFormatProperties = structureChain.template get<VULKAN_HPP_NAMESPACE::ImageFormatProperties2>();
6281 d.vkGetPhysicalDeviceImageFormatProperties2( m_physicalDevice,
6290 template <
typename Dispatch>
6291 VULKAN_HPP_INLINE void PhysicalDevice::getQueueFamilyProperties2( uint32_t * pQueueFamilyPropertyCount,
6292 VULKAN_HPP_NAMESPACE::QueueFamilyProperties2 * pQueueFamilyProperties,
6296 d.vkGetPhysicalDeviceQueueFamilyProperties2(
6297 m_physicalDevice, pQueueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( pQueueFamilyProperties ) );
6300 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6301 template <
typename QueueFamilyProperties2Allocator,
typename Dispatch>
6303 PhysicalDevice::getQueueFamilyProperties2( Dispatch
const & d )
const
6306 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6308 "Function <vkGetPhysicalDeviceQueueFamilyProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6311 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2, QueueFamilyProperties2Allocator> queueFamilyProperties;
6312 uint32_t queueFamilyPropertyCount;
6313 d.vkGetPhysicalDeviceQueueFamilyProperties2( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
6314 queueFamilyProperties.resize( queueFamilyPropertyCount );
6315 d.vkGetPhysicalDeviceQueueFamilyProperties2(
6316 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( queueFamilyProperties.data() ) );
6319 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
6321 queueFamilyProperties.resize( queueFamilyPropertyCount );
6323 return queueFamilyProperties;
6327 typename QueueFamilyProperties2Allocator,
6329 typename std::enable_if<std::is_same<typename QueueFamilyProperties2Allocator::value_type, VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>::value,
int>::type>
6331 PhysicalDevice::getQueueFamilyProperties2( QueueFamilyProperties2Allocator & queueFamilyProperties2Allocator, Dispatch
const & d )
const
6334 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6336 "Function <vkGetPhysicalDeviceQueueFamilyProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6339 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2, QueueFamilyProperties2Allocator> queueFamilyProperties( queueFamilyProperties2Allocator );
6340 uint32_t queueFamilyPropertyCount;
6341 d.vkGetPhysicalDeviceQueueFamilyProperties2( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
6342 queueFamilyProperties.resize( queueFamilyPropertyCount );
6343 d.vkGetPhysicalDeviceQueueFamilyProperties2(
6344 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( queueFamilyProperties.data() ) );
6347 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
6349 queueFamilyProperties.resize( queueFamilyPropertyCount );
6351 return queueFamilyProperties;
6354 template <
typename StructureChain,
typename StructureChainAllocator,
typename Dispatch>
6356 PhysicalDevice::getQueueFamilyProperties2( Dispatch
const & d )
const
6359 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6361 "Function <vkGetPhysicalDeviceQueueFamilyProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6364 std::vector<StructureChain, StructureChainAllocator> structureChains;
6365 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2> queueFamilyProperties;
6366 uint32_t queueFamilyPropertyCount;
6367 d.vkGetPhysicalDeviceQueueFamilyProperties2( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
6368 structureChains.resize( queueFamilyPropertyCount );
6369 queueFamilyProperties.resize( queueFamilyPropertyCount );
6370 for ( uint32_t i = 0; i < queueFamilyPropertyCount; i++ )
6372 queueFamilyProperties[i].pNext = structureChains[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>().pNext;
6374 d.vkGetPhysicalDeviceQueueFamilyProperties2(
6375 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( queueFamilyProperties.data() ) );
6378 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
6380 structureChains.resize( queueFamilyPropertyCount );
6382 for ( uint32_t i = 0; i < queueFamilyPropertyCount; i++ )
6384 structureChains[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>() = queueFamilyProperties[i];
6386 return structureChains;
6390 typename StructureChainAllocator,
6392 typename std::enable_if<std::is_same<typename StructureChainAllocator::value_type, StructureChain>::value,
int>::type>
6394 PhysicalDevice::getQueueFamilyProperties2( StructureChainAllocator & structureChainAllocator, Dispatch
const & d )
const
6397 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6399 "Function <vkGetPhysicalDeviceQueueFamilyProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6402 std::vector<StructureChain, StructureChainAllocator> structureChains( structureChainAllocator );
6403 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2> queueFamilyProperties;
6404 uint32_t queueFamilyPropertyCount;
6405 d.vkGetPhysicalDeviceQueueFamilyProperties2( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
6406 structureChains.resize( queueFamilyPropertyCount );
6407 queueFamilyProperties.resize( queueFamilyPropertyCount );
6408 for ( uint32_t i = 0; i < queueFamilyPropertyCount; i++ )
6410 queueFamilyProperties[i].pNext = structureChains[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>().pNext;
6412 d.vkGetPhysicalDeviceQueueFamilyProperties2(
6413 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( queueFamilyProperties.data() ) );
6416 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
6418 structureChains.resize( queueFamilyPropertyCount );
6420 for ( uint32_t i = 0; i < queueFamilyPropertyCount; i++ )
6422 structureChains[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>() = queueFamilyProperties[i];
6424 return structureChains;
6428 template <
typename Dispatch>
6429 VULKAN_HPP_INLINE void PhysicalDevice::getMemoryProperties2( VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 * pMemoryProperties,
6436 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6437 template <
typename Dispatch>
6442 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6444 "Function <vkGetPhysicalDeviceMemoryProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6447 VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 memoryProperties;
6450 return memoryProperties;
6453 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
6458 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6460 "Function <vkGetPhysicalDeviceMemoryProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6463 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
6464 VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 & memoryProperties =
6465 structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2>();
6468 return structureChain;
6472 template <
typename Dispatch>
6473 VULKAN_HPP_INLINE void PhysicalDevice::getSparseImageFormatProperties2(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 * pFormatInfo,
6474 uint32_t * pPropertyCount,
6475 VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2 * pProperties,
6479 d.vkGetPhysicalDeviceSparseImageFormatProperties2( m_physicalDevice,
6485 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6486 template <
typename SparseImageFormatProperties2Allocator,
typename Dispatch>
6488 PhysicalDevice::getSparseImageFormatProperties2(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 & formatInfo, Dispatch
const & d )
const
6491 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6493 "Function <vkGetPhysicalDeviceSparseImageFormatProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6496 std::vector<VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> properties;
6497 uint32_t propertyCount;
6498 d.vkGetPhysicalDeviceSparseImageFormatProperties2(
6500 properties.resize( propertyCount );
6501 d.vkGetPhysicalDeviceSparseImageFormatProperties2( m_physicalDevice,
6507 if ( propertyCount < properties.size() )
6509 properties.resize( propertyCount );
6515 typename SparseImageFormatProperties2Allocator,
6517 typename std::enable_if<std::is_same<typename SparseImageFormatProperties2Allocator::value_type, VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2>::value,
6520 PhysicalDevice::getSparseImageFormatProperties2(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 & formatInfo,
6521 SparseImageFormatProperties2Allocator & sparseImageFormatProperties2Allocator,
6522 Dispatch
const & d )
const
6525 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6527 "Function <vkGetPhysicalDeviceSparseImageFormatProperties2> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
6530 std::vector<VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> properties( sparseImageFormatProperties2Allocator );
6531 uint32_t propertyCount;
6532 d.vkGetPhysicalDeviceSparseImageFormatProperties2(
6534 properties.resize( propertyCount );
6535 d.vkGetPhysicalDeviceSparseImageFormatProperties2( m_physicalDevice,
6541 if ( propertyCount < properties.size() )
6543 properties.resize( propertyCount );
6549 template <
typename Dispatch>
6558 template <
typename Dispatch>
6559 VULKAN_HPP_INLINE void Device::getQueue2(
const VULKAN_HPP_NAMESPACE::DeviceQueueInfo2 * pQueueInfo,
6560 VULKAN_HPP_NAMESPACE::Queue * pQueue,
6564 d.vkGetDeviceQueue2( m_device,
reinterpret_cast<const VkDeviceQueueInfo2 *
>( pQueueInfo ),
reinterpret_cast<VkQueue *
>( pQueue ) );
6567 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6568 template <
typename Dispatch>
6573 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6574 VULKAN_HPP_ASSERT( d.vkGetDeviceQueue2 &&
"Function <vkGetDeviceQueue2> requires <VK_VERSION_1_1>" );
6577 VULKAN_HPP_NAMESPACE::Queue queue;
6578 d.vkGetDeviceQueue2( m_device,
reinterpret_cast<const VkDeviceQueueInfo2 *
>( &queueInfo ),
reinterpret_cast<VkQueue *
>( &queue ) );
6584 template <
typename Dispatch>
6586 Device::createSamplerYcbcrConversion(
const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo * pCreateInfo,
6587 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
6588 VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion * pYcbcrConversion,
6592 return static_cast<Result>( d.vkCreateSamplerYcbcrConversion( m_device,
6598 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6599 template <
typename Dispatch>
6601 Device::createSamplerYcbcrConversion(
const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo & createInfo,
6603 Dispatch
const & d )
const
6606 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6608 "Function <vkCreateSamplerYcbcrConversion> requires <VK_KHR_sampler_ycbcr_conversion> or <VK_VERSION_1_1>" );
6611 VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion;
6615 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
6622 # ifndef VULKAN_HPP_NO_SMART_HANDLE
6623 template <
typename Dispatch>
6625 Device::createSamplerYcbcrConversionUnique(
const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo & createInfo,
6627 Dispatch
const & d )
const
6630 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6632 "Function <vkCreateSamplerYcbcrConversion> requires <VK_KHR_sampler_ycbcr_conversion> or <VK_VERSION_1_1>" );
6635 VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion;
6639 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
6649 template <
typename Dispatch>
6650 VULKAN_HPP_INLINE void Device::destroySamplerYcbcrConversion( VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion,
6651 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
6655 d.vkDestroySamplerYcbcrConversion(
6659 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6660 template <
typename Dispatch>
6661 VULKAN_HPP_INLINE void Device::destroySamplerYcbcrConversion( VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion,
6666 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6668 "Function <vkDestroySamplerYcbcrConversion> requires <VK_KHR_sampler_ycbcr_conversion> or <VK_VERSION_1_1>" );
6671 d.vkDestroySamplerYcbcrConversion(
6674 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
6678 template <
typename Dispatch>
6679 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion,
6680 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
6684 d.vkDestroySamplerYcbcrConversion(
6688 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6689 template <
typename Dispatch>
6690 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion,
6695 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6697 "Function <vkDestroySamplerYcbcrConversion> requires <VK_KHR_sampler_ycbcr_conversion> or <VK_VERSION_1_1>" );
6700 d.vkDestroySamplerYcbcrConversion(
6703 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
6707 template <
typename Dispatch>
6709 Device::createDescriptorUpdateTemplate(
const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo * pCreateInfo,
6710 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
6711 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate * pDescriptorUpdateTemplate,
6715 return static_cast<Result>( d.vkCreateDescriptorUpdateTemplate( m_device,
6721 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6722 template <
typename Dispatch>
6724 Device::createDescriptorUpdateTemplate(
const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo & createInfo,
6726 Dispatch
const & d )
const
6729 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6731 "Function <vkCreateDescriptorUpdateTemplate> requires <VK_KHR_descriptor_update_template> or <VK_VERSION_1_1>" );
6734 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate;
6738 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
6745 # ifndef VULKAN_HPP_NO_SMART_HANDLE
6746 template <
typename Dispatch>
6748 Device::createDescriptorUpdateTemplateUnique(
const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo & createInfo,
6750 Dispatch
const & d )
const
6753 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6755 "Function <vkCreateDescriptorUpdateTemplate> requires <VK_KHR_descriptor_update_template> or <VK_VERSION_1_1>" );
6758 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate;
6762 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
6773 template <
typename Dispatch>
6774 VULKAN_HPP_INLINE void Device::destroyDescriptorUpdateTemplate( VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
6775 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
6779 d.vkDestroyDescriptorUpdateTemplate(
6783 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6784 template <
typename Dispatch>
6785 VULKAN_HPP_INLINE void Device::destroyDescriptorUpdateTemplate( VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
6790 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6792 "Function <vkDestroyDescriptorUpdateTemplate> requires <VK_KHR_descriptor_update_template> or <VK_VERSION_1_1>" );
6795 d.vkDestroyDescriptorUpdateTemplate(
6798 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
6802 template <
typename Dispatch>
6803 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
6804 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
6808 d.vkDestroyDescriptorUpdateTemplate(
6812 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6813 template <
typename Dispatch>
6814 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
6819 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6821 "Function <vkDestroyDescriptorUpdateTemplate> requires <VK_KHR_descriptor_update_template> or <VK_VERSION_1_1>" );
6824 d.vkDestroyDescriptorUpdateTemplate(
6827 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
6831 template <
typename Dispatch>
6832 VULKAN_HPP_INLINE void Device::updateDescriptorSetWithTemplate( VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet,
6833 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
6838 d.vkUpdateDescriptorSetWithTemplate(
6842 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6843 template <
typename DataType,
typename Dispatch>
6844 VULKAN_HPP_INLINE void Device::updateDescriptorSetWithTemplate( VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet,
6845 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
6846 DataType
const & data,
6850 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6852 "Function <vkUpdateDescriptorSetWithTemplate> requires <VK_KHR_descriptor_update_template> or <VK_VERSION_1_1>" );
6855 d.vkUpdateDescriptorSetWithTemplate( m_device,
6858 reinterpret_cast<const void *
>( &data ) );
6862 template <
typename Dispatch>
6863 VULKAN_HPP_INLINE void PhysicalDevice::getExternalBufferProperties(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalBufferInfo * pExternalBufferInfo,
6864 VULKAN_HPP_NAMESPACE::ExternalBufferProperties * pExternalBufferProperties,
6868 d.vkGetPhysicalDeviceExternalBufferProperties( m_physicalDevice,
6873 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6874 template <
typename Dispatch>
6876 PhysicalDevice::getExternalBufferProperties(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalBufferInfo & externalBufferInfo,
6880 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6882 "Function <vkGetPhysicalDeviceExternalBufferProperties> requires <VK_KHR_external_memory_capabilities> or <VK_VERSION_1_1>" );
6885 VULKAN_HPP_NAMESPACE::ExternalBufferProperties externalBufferProperties;
6886 d.vkGetPhysicalDeviceExternalBufferProperties( m_physicalDevice,
6890 return externalBufferProperties;
6894 template <
typename Dispatch>
6895 VULKAN_HPP_INLINE void PhysicalDevice::getExternalFenceProperties(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalFenceInfo * pExternalFenceInfo,
6896 VULKAN_HPP_NAMESPACE::ExternalFenceProperties * pExternalFenceProperties,
6900 d.vkGetPhysicalDeviceExternalFenceProperties( m_physicalDevice,
6905 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6906 template <
typename Dispatch>
6908 PhysicalDevice::getExternalFenceProperties(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalFenceInfo & externalFenceInfo,
6912 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6914 "Function <vkGetPhysicalDeviceExternalFenceProperties> requires <VK_KHR_external_fence_capabilities> or <VK_VERSION_1_1>" );
6917 VULKAN_HPP_NAMESPACE::ExternalFenceProperties externalFenceProperties;
6918 d.vkGetPhysicalDeviceExternalFenceProperties( m_physicalDevice,
6922 return externalFenceProperties;
6926 template <
typename Dispatch>
6928 PhysicalDevice::getExternalSemaphoreProperties(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalSemaphoreInfo * pExternalSemaphoreInfo,
6929 VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties * pExternalSemaphoreProperties,
6933 d.vkGetPhysicalDeviceExternalSemaphoreProperties( m_physicalDevice,
6938 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6939 template <
typename Dispatch>
6941 PhysicalDevice::getExternalSemaphoreProperties(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalSemaphoreInfo & externalSemaphoreInfo,
6945 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6947 "Function <vkGetPhysicalDeviceExternalSemaphoreProperties> requires <VK_KHR_external_semaphore_capabilities> or <VK_VERSION_1_1>" );
6950 VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties externalSemaphoreProperties;
6951 d.vkGetPhysicalDeviceExternalSemaphoreProperties( m_physicalDevice,
6955 return externalSemaphoreProperties;
6959 template <
typename Dispatch>
6960 VULKAN_HPP_INLINE void Device::getDescriptorSetLayoutSupport(
const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo * pCreateInfo,
6961 VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport * pSupport,
6965 d.vkGetDescriptorSetLayoutSupport(
6969 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
6970 template <
typename Dispatch>
6972 Device::getDescriptorSetLayoutSupport(
const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo & createInfo,
6976 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6977 VULKAN_HPP_ASSERT( d.vkGetDescriptorSetLayoutSupport &&
"Function <vkGetDescriptorSetLayoutSupport> requires <VK_KHR_maintenance3> or <VK_VERSION_1_1>" );
6980 VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport support;
6981 d.vkGetDescriptorSetLayoutSupport(
6987 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
6989 Device::getDescriptorSetLayoutSupport(
const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo & createInfo,
6993 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
6994 VULKAN_HPP_ASSERT( d.vkGetDescriptorSetLayoutSupport &&
"Function <vkGetDescriptorSetLayoutSupport> requires <VK_KHR_maintenance3> or <VK_VERSION_1_1>" );
6997 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
6998 VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport & support = structureChain.template get<VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport>();
6999 d.vkGetDescriptorSetLayoutSupport(
7002 return structureChain;
7008 template <
typename Dispatch>
7011 VULKAN_HPP_NAMESPACE::Buffer countBuffer,
7013 uint32_t maxDrawCount,
7018 d.vkCmdDrawIndirectCount( m_commandBuffer,
7021 static_cast<VkBuffer>( countBuffer ),
7027 template <
typename Dispatch>
7028 VULKAN_HPP_INLINE void CommandBuffer::drawIndexedIndirectCount( VULKAN_HPP_NAMESPACE::Buffer buffer,
7030 VULKAN_HPP_NAMESPACE::Buffer countBuffer,
7032 uint32_t maxDrawCount,
7037 d.vkCmdDrawIndexedIndirectCount( m_commandBuffer,
7040 static_cast<VkBuffer>( countBuffer ),
7046 template <
typename Dispatch>
7048 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
7049 VULKAN_HPP_NAMESPACE::RenderPass * pRenderPass,
7053 return static_cast<Result>( d.vkCreateRenderPass2( m_device,
7059 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7060 template <
typename Dispatch>
7062 Device::createRenderPass2(
const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 & createInfo,
7064 Dispatch
const & d )
const
7067 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7068 VULKAN_HPP_ASSERT( d.vkCreateRenderPass2 &&
"Function <vkCreateRenderPass2> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
7071 VULKAN_HPP_NAMESPACE::RenderPass renderPass;
7073 d.vkCreateRenderPass2( m_device,
7075 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
7082 # ifndef VULKAN_HPP_NO_SMART_HANDLE
7083 template <
typename Dispatch>
7085 Device::createRenderPass2Unique(
const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 & createInfo,
7087 Dispatch
const & d )
const
7090 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7091 VULKAN_HPP_ASSERT( d.vkCreateRenderPass2 &&
"Function <vkCreateRenderPass2> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
7094 VULKAN_HPP_NAMESPACE::RenderPass renderPass;
7096 d.vkCreateRenderPass2( m_device,
7098 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
7108 template <
typename Dispatch>
7109 VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass2(
const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo * pRenderPassBegin,
7110 const VULKAN_HPP_NAMESPACE::SubpassBeginInfo * pSubpassBeginInfo,
7114 d.vkCmdBeginRenderPass2(
7118 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7119 template <
typename Dispatch>
7120 VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass2(
const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo & renderPassBegin,
7121 const VULKAN_HPP_NAMESPACE::SubpassBeginInfo & subpassBeginInfo,
7125 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7126 VULKAN_HPP_ASSERT( d.vkCmdBeginRenderPass2 &&
"Function <vkCmdBeginRenderPass2> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
7129 d.vkCmdBeginRenderPass2(
7134 template <
typename Dispatch>
7135 VULKAN_HPP_INLINE void CommandBuffer::nextSubpass2(
const VULKAN_HPP_NAMESPACE::SubpassBeginInfo * pSubpassBeginInfo,
7136 const VULKAN_HPP_NAMESPACE::SubpassEndInfo * pSubpassEndInfo,
7140 d.vkCmdNextSubpass2(
7144 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7145 template <
typename Dispatch>
7146 VULKAN_HPP_INLINE void CommandBuffer::nextSubpass2(
const VULKAN_HPP_NAMESPACE::SubpassBeginInfo & subpassBeginInfo,
7147 const VULKAN_HPP_NAMESPACE::SubpassEndInfo & subpassEndInfo,
7151 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7152 VULKAN_HPP_ASSERT( d.vkCmdNextSubpass2 &&
"Function <vkCmdNextSubpass2> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
7155 d.vkCmdNextSubpass2(
7160 template <
typename Dispatch>
7161 VULKAN_HPP_INLINE void CommandBuffer::endRenderPass2(
const VULKAN_HPP_NAMESPACE::SubpassEndInfo * pSubpassEndInfo,
7165 d.vkCmdEndRenderPass2( m_commandBuffer,
reinterpret_cast<const VkSubpassEndInfo *
>( pSubpassEndInfo ) );
7168 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7169 template <
typename Dispatch>
7170 VULKAN_HPP_INLINE void CommandBuffer::endRenderPass2(
const VULKAN_HPP_NAMESPACE::SubpassEndInfo & subpassEndInfo,
7174 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7175 VULKAN_HPP_ASSERT( d.vkCmdEndRenderPass2 &&
"Function <vkCmdEndRenderPass2> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
7178 d.vkCmdEndRenderPass2( m_commandBuffer,
reinterpret_cast<const VkSubpassEndInfo *
>( &subpassEndInfo ) );
7182 template <
typename Dispatch>
7184 Device::resetQueryPool( VULKAN_HPP_NAMESPACE::QueryPool queryPool, uint32_t firstQuery, uint32_t queryCount, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
7187 d.vkResetQueryPool( m_device,
static_cast<VkQueryPool>( queryPool ), firstQuery, queryCount );
7190 template <
typename Dispatch>
7196 return static_cast<Result>( d.vkGetSemaphoreCounterValue( m_device,
static_cast<VkSemaphore>( semaphore ), pValue ) );
7199 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7200 template <
typename Dispatch>
7202 Dispatch
const & d )
const
7205 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7206 VULKAN_HPP_ASSERT( d.vkGetSemaphoreCounterValue &&
"Function <vkGetSemaphoreCounterValue> requires <VK_KHR_timeline_semaphore> or <VK_VERSION_1_2>" );
7218 template <
typename Dispatch>
7224 return static_cast<Result>( d.vkWaitSemaphores( m_device,
reinterpret_cast<const VkSemaphoreWaitInfo *
>( pWaitInfo ), timeout ) );
7227 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7228 template <
typename Dispatch>
7230 Device::waitSemaphores(
const VULKAN_HPP_NAMESPACE::SemaphoreWaitInfo & waitInfo, uint64_t timeout, Dispatch
const & d )
const
7233 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7234 VULKAN_HPP_ASSERT( d.vkWaitSemaphores &&
"Function <vkWaitSemaphores> requires <VK_KHR_timeline_semaphore> or <VK_VERSION_1_2>" );
7240 result,
VULKAN_HPP_NAMESPACE_STRING "::Device::waitSemaphores", { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eTimeout } );
7246 template <
typename Dispatch>
7254 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7255 template <
typename Dispatch>
7257 Device::signalSemaphore(
const VULKAN_HPP_NAMESPACE::SemaphoreSignalInfo & signalInfo, Dispatch
const & d )
const
7260 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7261 VULKAN_HPP_ASSERT( d.vkSignalSemaphore &&
"Function <vkSignalSemaphore> requires <VK_KHR_timeline_semaphore> or <VK_VERSION_1_2>" );
7272 template <
typename Dispatch>
7280 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7281 template <
typename Dispatch>
7286 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7288 "Function <vkGetBufferDeviceAddress> requires <VK_EXT_buffer_device_address> or <VK_KHR_buffer_device_address> or <VK_VERSION_1_2>" );
7297 template <
typename Dispatch>
7298 VULKAN_HPP_INLINE uint64_t Device::getBufferOpaqueCaptureAddress(
const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo * pInfo,
7305 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7306 template <
typename Dispatch>
7307 VULKAN_HPP_INLINE uint64_t Device::getBufferOpaqueCaptureAddress(
const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo & info,
7311 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7313 "Function <vkGetBufferOpaqueCaptureAddress> requires <VK_KHR_buffer_device_address> or <VK_VERSION_1_2>" );
7316 uint64_t result = d.vkGetBufferOpaqueCaptureAddress( m_device,
reinterpret_cast<const VkBufferDeviceAddressInfo *
>( &info ) );
7322 template <
typename Dispatch>
7323 VULKAN_HPP_INLINE uint64_t Device::getMemoryOpaqueCaptureAddress(
const VULKAN_HPP_NAMESPACE::DeviceMemoryOpaqueCaptureAddressInfo * pInfo,
7330 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7331 template <
typename Dispatch>
7332 VULKAN_HPP_INLINE uint64_t Device::getMemoryOpaqueCaptureAddress(
const VULKAN_HPP_NAMESPACE::DeviceMemoryOpaqueCaptureAddressInfo & info,
7336 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7338 "Function <vkGetDeviceMemoryOpaqueCaptureAddress> requires <VK_KHR_buffer_device_address> or <VK_VERSION_1_2>" );
7349 template <
typename Dispatch>
7351 VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties * pToolProperties,
7355 return static_cast<Result>(
7359 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7360 template <
typename PhysicalDeviceToolPropertiesAllocator,
typename Dispatch>
7363 PhysicalDevice::getToolProperties( Dispatch
const & d )
const
7366 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7368 "Function <vkGetPhysicalDeviceToolProperties> requires <VK_EXT_tooling_info> or <VK_VERSION_1_3>" );
7371 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator> toolProperties;
7377 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && toolCount )
7379 toolProperties.resize( toolCount );
7381 d.vkGetPhysicalDeviceToolProperties( m_physicalDevice, &toolCount,
reinterpret_cast<VkPhysicalDeviceToolProperties *
>( toolProperties.data() ) ) );
7383 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
7386 if ( toolCount < toolProperties.size() )
7388 toolProperties.resize( toolCount );
7394 typename PhysicalDeviceToolPropertiesAllocator,
7396 typename std::enable_if<std::is_same<typename PhysicalDeviceToolPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties>::value,
7400 PhysicalDevice::getToolProperties( PhysicalDeviceToolPropertiesAllocator & physicalDeviceToolPropertiesAllocator, Dispatch
const & d )
const
7403 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7405 "Function <vkGetPhysicalDeviceToolProperties> requires <VK_EXT_tooling_info> or <VK_VERSION_1_3>" );
7408 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator> toolProperties(
7409 physicalDeviceToolPropertiesAllocator );
7415 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && toolCount )
7417 toolProperties.resize( toolCount );
7419 d.vkGetPhysicalDeviceToolProperties( m_physicalDevice, &toolCount,
reinterpret_cast<VkPhysicalDeviceToolProperties *
>( toolProperties.data() ) ) );
7421 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
7424 if ( toolCount < toolProperties.size() )
7426 toolProperties.resize( toolCount );
7432 template <
typename Dispatch>
7434 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
7435 VULKAN_HPP_NAMESPACE::PrivateDataSlot * pPrivateDataSlot,
7439 return static_cast<Result>( d.vkCreatePrivateDataSlot( m_device,
7445 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7446 template <
typename Dispatch>
7448 Device::createPrivateDataSlot(
const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo & createInfo,
7450 Dispatch
const & d )
const
7453 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7454 VULKAN_HPP_ASSERT( d.vkCreatePrivateDataSlot &&
"Function <vkCreatePrivateDataSlot> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
7457 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot;
7459 d.vkCreatePrivateDataSlot( m_device,
7461 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
7468 # ifndef VULKAN_HPP_NO_SMART_HANDLE
7469 template <
typename Dispatch>
7471 Device::createPrivateDataSlotUnique(
const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo & createInfo,
7473 Dispatch
const & d )
const
7476 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7477 VULKAN_HPP_ASSERT( d.vkCreatePrivateDataSlot &&
"Function <vkCreatePrivateDataSlot> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
7480 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot;
7482 d.vkCreatePrivateDataSlot( m_device,
7484 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
7494 template <
typename Dispatch>
7495 VULKAN_HPP_INLINE void Device::destroyPrivateDataSlot( VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
7496 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
7503 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7504 template <
typename Dispatch>
7505 VULKAN_HPP_INLINE void Device::destroyPrivateDataSlot( VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
7510 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7511 VULKAN_HPP_ASSERT( d.vkDestroyPrivateDataSlot &&
"Function <vkDestroyPrivateDataSlot> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
7514 d.vkDestroyPrivateDataSlot(
7517 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
7521 template <
typename Dispatch>
7523 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
7530 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7531 template <
typename Dispatch>
7537 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7538 VULKAN_HPP_ASSERT( d.vkDestroyPrivateDataSlot &&
"Function <vkDestroyPrivateDataSlot> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
7541 d.vkDestroyPrivateDataSlot(
7544 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
7548 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
7549 template <
typename Dispatch>
7551 uint64_t objectHandle,
7552 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
7557 return static_cast<Result>(
7558 d.vkSetPrivateData( m_device,
static_cast<VkObjectType>( objectType_ ), objectHandle,
static_cast<VkPrivateDataSlot>( privateDataSlot ), data ) );
7561 template <
typename Dispatch>
7563 uint64_t objectHandle,
7564 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
7566 Dispatch
const & d )
const
7569 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7570 VULKAN_HPP_ASSERT( d.vkSetPrivateData &&
"Function <vkSetPrivateData> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
7574 d.vkSetPrivateData( m_device,
static_cast<VkObjectType>( objectType_ ), objectHandle,
static_cast<VkPrivateDataSlot>( privateDataSlot ), data ) );
7581 template <
typename Dispatch>
7583 uint64_t objectHandle,
7584 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
7589 d.vkGetPrivateData( m_device,
static_cast<VkObjectType>( objectType_ ), objectHandle,
static_cast<VkPrivateDataSlot>( privateDataSlot ), pData );
7592 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7593 template <
typename Dispatch>
7595 uint64_t objectHandle,
7596 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
7600 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7601 VULKAN_HPP_ASSERT( d.vkGetPrivateData &&
"Function <vkGetPrivateData> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
7605 d.vkGetPrivateData( m_device,
static_cast<VkObjectType>( objectType_ ), objectHandle,
static_cast<VkPrivateDataSlot>( privateDataSlot ), &data );
7611 template <
typename Dispatch>
7613 const VULKAN_HPP_NAMESPACE::DependencyInfo * pDependencyInfo,
7617 d.vkCmdSetEvent2( m_commandBuffer,
static_cast<VkEvent>( event ),
reinterpret_cast<const VkDependencyInfo *
>( pDependencyInfo ) );
7620 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7621 template <
typename Dispatch>
7623 const VULKAN_HPP_NAMESPACE::DependencyInfo & dependencyInfo,
7627 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7628 VULKAN_HPP_ASSERT( d.vkCmdSetEvent2 &&
"Function <vkCmdSetEvent2> requires <VK_KHR_synchronization2> or <VK_VERSION_1_3>" );
7631 d.vkCmdSetEvent2( m_commandBuffer,
static_cast<VkEvent>( event ),
reinterpret_cast<const VkDependencyInfo *
>( &dependencyInfo ) );
7635 template <
typename Dispatch>
7644 template <
typename Dispatch>
7646 const VULKAN_HPP_NAMESPACE::Event * pEvents,
7647 const VULKAN_HPP_NAMESPACE::DependencyInfo * pDependencyInfos,
7652 m_commandBuffer, eventCount,
reinterpret_cast<const VkEvent *
>( pEvents ),
reinterpret_cast<const VkDependencyInfo *
>( pDependencyInfos ) );
7655 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7656 template <
typename Dispatch>
7657 VULKAN_HPP_INLINE void CommandBuffer::waitEvents2( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Event>
const & events,
7658 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DependencyInfo>
const & dependencyInfos,
7662 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7663 VULKAN_HPP_ASSERT( d.vkCmdWaitEvents2 &&
"Function <vkCmdWaitEvents2> requires <VK_KHR_synchronization2> or <VK_VERSION_1_3>" );
7665 # ifdef VULKAN_HPP_NO_EXCEPTIONS
7668 if ( events.size() != dependencyInfos.size() )
7674 d.vkCmdWaitEvents2( m_commandBuffer,
7676 reinterpret_cast<const VkEvent *
>( events.data() ),
7681 template <
typename Dispatch>
7682 VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier2(
const VULKAN_HPP_NAMESPACE::DependencyInfo * pDependencyInfo,
7686 d.vkCmdPipelineBarrier2( m_commandBuffer,
reinterpret_cast<const VkDependencyInfo *
>( pDependencyInfo ) );
7689 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7690 template <
typename Dispatch>
7691 VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier2(
const VULKAN_HPP_NAMESPACE::DependencyInfo & dependencyInfo,
7695 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7696 VULKAN_HPP_ASSERT( d.vkCmdPipelineBarrier2 &&
"Function <vkCmdPipelineBarrier2> requires <VK_KHR_synchronization2> or <VK_VERSION_1_3>" );
7699 d.vkCmdPipelineBarrier2( m_commandBuffer,
reinterpret_cast<const VkDependencyInfo *
>( &dependencyInfo ) );
7703 template <
typename Dispatch>
7705 VULKAN_HPP_NAMESPACE::QueryPool queryPool,
7713 template <
typename Dispatch>
7715 const VULKAN_HPP_NAMESPACE::SubmitInfo2 * pSubmits,
7716 VULKAN_HPP_NAMESPACE::Fence fence,
7720 return static_cast<Result>( d.vkQueueSubmit2( m_queue, submitCount,
reinterpret_cast<const VkSubmitInfo2 *
>( pSubmits ),
static_cast<VkFence>( fence ) ) );
7723 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7724 template <
typename Dispatch>
7726 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SubmitInfo2>
const & submits, VULKAN_HPP_NAMESPACE::Fence fence, Dispatch
const & d )
const
7729 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7730 VULKAN_HPP_ASSERT( d.vkQueueSubmit2 &&
"Function <vkQueueSubmit2> requires <VK_KHR_synchronization2> or <VK_VERSION_1_3>" );
7734 d.vkQueueSubmit2( m_queue, submits.size(),
reinterpret_cast<const VkSubmitInfo2 *
>( submits.data() ),
static_cast<VkFence>( fence ) ) );
7741 template <
typename Dispatch>
7742 VULKAN_HPP_INLINE void CommandBuffer::copyBuffer2(
const VULKAN_HPP_NAMESPACE::CopyBufferInfo2 * pCopyBufferInfo,
7746 d.vkCmdCopyBuffer2( m_commandBuffer,
reinterpret_cast<const VkCopyBufferInfo2 *
>( pCopyBufferInfo ) );
7749 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7750 template <
typename Dispatch>
7751 VULKAN_HPP_INLINE void CommandBuffer::copyBuffer2(
const VULKAN_HPP_NAMESPACE::CopyBufferInfo2 & copyBufferInfo,
7755 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7756 VULKAN_HPP_ASSERT( d.vkCmdCopyBuffer2 &&
"Function <vkCmdCopyBuffer2> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
7759 d.vkCmdCopyBuffer2( m_commandBuffer,
reinterpret_cast<const VkCopyBufferInfo2 *
>( ©BufferInfo ) );
7763 template <
typename Dispatch>
7767 d.vkCmdCopyImage2( m_commandBuffer,
reinterpret_cast<const VkCopyImageInfo2 *
>( pCopyImageInfo ) );
7770 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7771 template <
typename Dispatch>
7775 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7776 VULKAN_HPP_ASSERT( d.vkCmdCopyImage2 &&
"Function <vkCmdCopyImage2> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
7779 d.vkCmdCopyImage2( m_commandBuffer,
reinterpret_cast<const VkCopyImageInfo2 *
>( ©ImageInfo ) );
7783 template <
typename Dispatch>
7784 VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage2(
const VULKAN_HPP_NAMESPACE::CopyBufferToImageInfo2 * pCopyBufferToImageInfo,
7788 d.vkCmdCopyBufferToImage2( m_commandBuffer,
reinterpret_cast<const VkCopyBufferToImageInfo2 *
>( pCopyBufferToImageInfo ) );
7791 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7792 template <
typename Dispatch>
7793 VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage2(
const VULKAN_HPP_NAMESPACE::CopyBufferToImageInfo2 & copyBufferToImageInfo,
7797 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7798 VULKAN_HPP_ASSERT( d.vkCmdCopyBufferToImage2 &&
"Function <vkCmdCopyBufferToImage2> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
7801 d.vkCmdCopyBufferToImage2( m_commandBuffer,
reinterpret_cast<const VkCopyBufferToImageInfo2 *
>( ©BufferToImageInfo ) );
7805 template <
typename Dispatch>
7806 VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer2(
const VULKAN_HPP_NAMESPACE::CopyImageToBufferInfo2 * pCopyImageToBufferInfo,
7810 d.vkCmdCopyImageToBuffer2( m_commandBuffer,
reinterpret_cast<const VkCopyImageToBufferInfo2 *
>( pCopyImageToBufferInfo ) );
7813 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7814 template <
typename Dispatch>
7815 VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer2(
const VULKAN_HPP_NAMESPACE::CopyImageToBufferInfo2 & copyImageToBufferInfo,
7819 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7820 VULKAN_HPP_ASSERT( d.vkCmdCopyImageToBuffer2 &&
"Function <vkCmdCopyImageToBuffer2> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
7823 d.vkCmdCopyImageToBuffer2( m_commandBuffer,
reinterpret_cast<const VkCopyImageToBufferInfo2 *
>( ©ImageToBufferInfo ) );
7827 template <
typename Dispatch>
7831 d.vkCmdBlitImage2( m_commandBuffer,
reinterpret_cast<const VkBlitImageInfo2 *
>( pBlitImageInfo ) );
7834 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7835 template <
typename Dispatch>
7839 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7840 VULKAN_HPP_ASSERT( d.vkCmdBlitImage2 &&
"Function <vkCmdBlitImage2> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
7843 d.vkCmdBlitImage2( m_commandBuffer,
reinterpret_cast<const VkBlitImageInfo2 *
>( &blitImageInfo ) );
7847 template <
typename Dispatch>
7848 VULKAN_HPP_INLINE void CommandBuffer::resolveImage2(
const VULKAN_HPP_NAMESPACE::ResolveImageInfo2 * pResolveImageInfo,
7852 d.vkCmdResolveImage2( m_commandBuffer,
reinterpret_cast<const VkResolveImageInfo2 *
>( pResolveImageInfo ) );
7855 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7856 template <
typename Dispatch>
7857 VULKAN_HPP_INLINE void CommandBuffer::resolveImage2(
const VULKAN_HPP_NAMESPACE::ResolveImageInfo2 & resolveImageInfo,
7861 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7862 VULKAN_HPP_ASSERT( d.vkCmdResolveImage2 &&
"Function <vkCmdResolveImage2> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
7865 d.vkCmdResolveImage2( m_commandBuffer,
reinterpret_cast<const VkResolveImageInfo2 *
>( &resolveImageInfo ) );
7869 template <
typename Dispatch>
7870 VULKAN_HPP_INLINE void CommandBuffer::beginRendering(
const VULKAN_HPP_NAMESPACE::RenderingInfo * pRenderingInfo,
7874 d.vkCmdBeginRendering( m_commandBuffer,
reinterpret_cast<const VkRenderingInfo *
>( pRenderingInfo ) );
7877 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7878 template <
typename Dispatch>
7879 VULKAN_HPP_INLINE void CommandBuffer::beginRendering(
const VULKAN_HPP_NAMESPACE::RenderingInfo & renderingInfo,
7883 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7884 VULKAN_HPP_ASSERT( d.vkCmdBeginRendering &&
"Function <vkCmdBeginRendering> requires <VK_KHR_dynamic_rendering> or <VK_VERSION_1_3>" );
7887 d.vkCmdBeginRendering( m_commandBuffer,
reinterpret_cast<const VkRenderingInfo *
>( &renderingInfo ) );
7891 template <
typename Dispatch>
7895 d.vkCmdEndRendering( m_commandBuffer );
7898 template <
typename Dispatch>
7902 d.vkCmdSetCullMode( m_commandBuffer,
static_cast<VkCullModeFlags>( cullMode ) );
7905 template <
typename Dispatch>
7909 d.vkCmdSetFrontFace( m_commandBuffer,
static_cast<VkFrontFace>( frontFace ) );
7912 template <
typename Dispatch>
7917 d.vkCmdSetPrimitiveTopology( m_commandBuffer,
static_cast<VkPrimitiveTopology>( primitiveTopology ) );
7920 template <
typename Dispatch>
7922 const VULKAN_HPP_NAMESPACE::Viewport * pViewports,
7926 d.vkCmdSetViewportWithCount( m_commandBuffer, viewportCount,
reinterpret_cast<const VkViewport *
>( pViewports ) );
7929 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7930 template <
typename Dispatch>
7931 VULKAN_HPP_INLINE void CommandBuffer::setViewportWithCount( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Viewport>
const & viewports,
7935 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7937 "Function <vkCmdSetViewportWithCount> requires <VK_EXT_extended_dynamic_state> or <VK_EXT_shader_object> or <VK_VERSION_1_3>" );
7940 d.vkCmdSetViewportWithCount( m_commandBuffer, viewports.size(),
reinterpret_cast<const VkViewport *
>( viewports.data() ) );
7944 template <
typename Dispatch>
7946 CommandBuffer::setScissorWithCount( uint32_t scissorCount,
const VULKAN_HPP_NAMESPACE::Rect2D * pScissors, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
7949 d.vkCmdSetScissorWithCount( m_commandBuffer, scissorCount,
reinterpret_cast<const VkRect2D *
>( pScissors ) );
7952 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7953 template <
typename Dispatch>
7954 VULKAN_HPP_INLINE void CommandBuffer::setScissorWithCount( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D>
const & scissors,
7958 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7960 "Function <vkCmdSetScissorWithCount> requires <VK_EXT_extended_dynamic_state> or <VK_EXT_shader_object> or <VK_VERSION_1_3>" );
7963 d.vkCmdSetScissorWithCount( m_commandBuffer, scissors.size(),
reinterpret_cast<const VkRect2D *
>( scissors.data() ) );
7967 template <
typename Dispatch>
7969 uint32_t bindingCount,
7970 const VULKAN_HPP_NAMESPACE::Buffer * pBuffers,
7977 d.vkCmdBindVertexBuffers2( m_commandBuffer,
7980 reinterpret_cast<const VkBuffer *
>( pBuffers ),
7986 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
7987 template <
typename Dispatch>
7989 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer>
const & buffers,
7990 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & offsets,
7991 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & sizes,
7992 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & strides,
7996 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
7998 "Function <vkCmdBindVertexBuffers2> requires <VK_EXT_extended_dynamic_state> or <VK_EXT_shader_object> or <VK_VERSION_1_3>" );
8000 # ifdef VULKAN_HPP_NO_EXCEPTIONS
8005 if ( buffers.size() != offsets.size() )
8009 if ( !sizes.empty() && buffers.size() != sizes.size() )
8013 if ( !strides.empty() && buffers.size() != strides.size() )
8019 d.vkCmdBindVertexBuffers2( m_commandBuffer,
8022 reinterpret_cast<const VkBuffer *
>( buffers.data() ),
8023 reinterpret_cast<const VkDeviceSize *
>( offsets.data() ),
8024 reinterpret_cast<const VkDeviceSize *
>( sizes.data() ),
8025 reinterpret_cast<const VkDeviceSize *
>( strides.data() ) );
8029 template <
typename Dispatch>
8033 d.vkCmdSetDepthTestEnable( m_commandBuffer,
static_cast<VkBool32>( depthTestEnable ) );
8036 template <
typename Dispatch>
8040 d.vkCmdSetDepthWriteEnable( m_commandBuffer,
static_cast<VkBool32>( depthWriteEnable ) );
8043 template <
typename Dispatch>
8047 d.vkCmdSetDepthCompareOp( m_commandBuffer,
static_cast<VkCompareOp>( depthCompareOp ) );
8050 template <
typename Dispatch>
8055 d.vkCmdSetDepthBoundsTestEnable( m_commandBuffer,
static_cast<VkBool32>( depthBoundsTestEnable ) );
8058 template <
typename Dispatch>
8062 d.vkCmdSetStencilTestEnable( m_commandBuffer,
static_cast<VkBool32>( stencilTestEnable ) );
8065 template <
typename Dispatch>
8074 d.vkCmdSetStencilOp( m_commandBuffer,
8082 template <
typename Dispatch>
8087 d.vkCmdSetRasterizerDiscardEnable( m_commandBuffer,
static_cast<VkBool32>( rasterizerDiscardEnable ) );
8090 template <
typename Dispatch>
8094 d.vkCmdSetDepthBiasEnable( m_commandBuffer,
static_cast<VkBool32>( depthBiasEnable ) );
8097 template <
typename Dispatch>
8102 d.vkCmdSetPrimitiveRestartEnable( m_commandBuffer,
static_cast<VkBool32>( primitiveRestartEnable ) );
8105 template <
typename Dispatch>
8106 VULKAN_HPP_INLINE void Device::getBufferMemoryRequirements(
const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements * pInfo,
8107 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
8111 d.vkGetDeviceBufferMemoryRequirements(
8115 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8116 template <
typename Dispatch>
8118 Device::getBufferMemoryRequirements(
const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
8121 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8123 "Function <vkGetDeviceBufferMemoryRequirements> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
8126 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
8127 d.vkGetDeviceBufferMemoryRequirements(
8130 return memoryRequirements;
8133 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
8135 Device::getBufferMemoryRequirements(
const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
8138 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8140 "Function <vkGetDeviceBufferMemoryRequirements> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
8143 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
8144 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
8145 d.vkGetDeviceBufferMemoryRequirements(
8148 return structureChain;
8152 template <
typename Dispatch>
8153 VULKAN_HPP_INLINE void Device::getImageMemoryRequirements(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements * pInfo,
8154 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
8158 d.vkGetDeviceImageMemoryRequirements(
8162 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8163 template <
typename Dispatch>
8165 Device::getImageMemoryRequirements(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
8168 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8170 "Function <vkGetDeviceImageMemoryRequirements> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
8173 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
8174 d.vkGetDeviceImageMemoryRequirements(
8177 return memoryRequirements;
8180 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
8182 Device::getImageMemoryRequirements(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
8185 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8187 "Function <vkGetDeviceImageMemoryRequirements> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
8190 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
8191 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
8192 d.vkGetDeviceImageMemoryRequirements(
8195 return structureChain;
8199 template <
typename Dispatch>
8200 VULKAN_HPP_INLINE void Device::getImageSparseMemoryRequirements(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements * pInfo,
8201 uint32_t * pSparseMemoryRequirementCount,
8202 VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2 * pSparseMemoryRequirements,
8206 d.vkGetDeviceImageSparseMemoryRequirements( m_device,
8208 pSparseMemoryRequirementCount,
8212 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8213 template <
typename SparseImageMemoryRequirements2Allocator,
typename Dispatch>
8215 Device::getImageSparseMemoryRequirements(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements & info, Dispatch
const & d )
const
8218 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8220 "Function <vkGetDeviceImageSparseMemoryRequirements> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
8223 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements;
8224 uint32_t sparseMemoryRequirementCount;
8225 d.vkGetDeviceImageSparseMemoryRequirements(
8227 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
8228 d.vkGetDeviceImageSparseMemoryRequirements( m_device,
8230 &sparseMemoryRequirementCount,
8233 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
8234 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
8236 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
8238 return sparseMemoryRequirements;
8241 template <
typename SparseImageMemoryRequirements2Allocator,
8243 typename std::enable_if<
8244 std::is_same<typename SparseImageMemoryRequirements2Allocator::value_type, VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2>::value,
8247 Device::getImageSparseMemoryRequirements(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements & info,
8248 SparseImageMemoryRequirements2Allocator & sparseImageMemoryRequirements2Allocator,
8249 Dispatch
const & d )
const
8252 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8254 "Function <vkGetDeviceImageSparseMemoryRequirements> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
8257 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements(
8258 sparseImageMemoryRequirements2Allocator );
8259 uint32_t sparseMemoryRequirementCount;
8260 d.vkGetDeviceImageSparseMemoryRequirements(
8262 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
8263 d.vkGetDeviceImageSparseMemoryRequirements( m_device,
8265 &sparseMemoryRequirementCount,
8268 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
8269 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
8271 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
8273 return sparseMemoryRequirements;
8279 template <
typename Dispatch>
8281 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
8288 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8289 template <
typename Dispatch>
8295 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8296 VULKAN_HPP_ASSERT( d.vkDestroySurfaceKHR &&
"Function <vkDestroySurfaceKHR> requires <VK_KHR_surface>" );
8299 d.vkDestroySurfaceKHR( m_instance,
8301 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
8305 template <
typename Dispatch>
8307 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
8314 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8315 template <
typename Dispatch>
8321 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8322 VULKAN_HPP_ASSERT( d.vkDestroySurfaceKHR &&
"Function <vkDestroySurfaceKHR> requires <VK_KHR_surface>" );
8325 d.vkDestroySurfaceKHR( m_instance,
8327 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
8331 template <
typename Dispatch>
8333 VULKAN_HPP_NAMESPACE::SurfaceKHR surface,
8338 return static_cast<Result>( d.vkGetPhysicalDeviceSurfaceSupportKHR(
8339 m_physicalDevice, queueFamilyIndex,
static_cast<VkSurfaceKHR>( surface ),
reinterpret_cast<VkBool32 *
>( pSupported ) ) );
8342 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8343 template <
typename Dispatch>
8345 PhysicalDevice::getSurfaceSupportKHR( uint32_t queueFamilyIndex, VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch
const & d )
const
8348 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8349 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceSurfaceSupportKHR &&
"Function <vkGetPhysicalDeviceSurfaceSupportKHR> requires <VK_KHR_surface>" );
8354 m_physicalDevice, queueFamilyIndex,
static_cast<VkSurfaceKHR>( surface ),
reinterpret_cast<VkBool32 *
>( &supported ) ) );
8361 template <
typename Dispatch>
8363 VULKAN_HPP_NAMESPACE::SurfaceCapabilitiesKHR * pSurfaceCapabilities,
8367 return static_cast<Result>( d.vkGetPhysicalDeviceSurfaceCapabilitiesKHR(
8371 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8372 template <
typename Dispatch>
8374 PhysicalDevice::getSurfaceCapabilitiesKHR( VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch
const & d )
const
8377 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8378 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceSurfaceCapabilitiesKHR &&
"Function <vkGetPhysicalDeviceSurfaceCapabilitiesKHR> requires <VK_KHR_surface>" );
8381 VULKAN_HPP_NAMESPACE::SurfaceCapabilitiesKHR surfaceCapabilities;
8390 template <
typename Dispatch>
8392 uint32_t * pSurfaceFormatCount,
8393 VULKAN_HPP_NAMESPACE::SurfaceFormatKHR * pSurfaceFormats,
8397 return static_cast<Result>( d.vkGetPhysicalDeviceSurfaceFormatsKHR(
8401 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8402 template <
typename SurfaceFormatKHRAllocator,
typename Dispatch>
8404 PhysicalDevice::getSurfaceFormatsKHR( VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch
const & d )
const
8407 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8408 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceSurfaceFormatsKHR &&
"Function <vkGetPhysicalDeviceSurfaceFormatsKHR> requires <VK_KHR_surface>" );
8411 std::vector<VULKAN_HPP_NAMESPACE::SurfaceFormatKHR, SurfaceFormatKHRAllocator> surfaceFormats;
8412 uint32_t surfaceFormatCount;
8417 d.vkGetPhysicalDeviceSurfaceFormatsKHR( m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &surfaceFormatCount,
nullptr ) );
8418 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && surfaceFormatCount )
8420 surfaceFormats.resize( surfaceFormatCount );
8422 m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &surfaceFormatCount,
reinterpret_cast<VkSurfaceFormatKHR *
>( surfaceFormats.data() ) ) );
8424 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
8427 if ( surfaceFormatCount < surfaceFormats.size() )
8429 surfaceFormats.resize( surfaceFormatCount );
8434 template <
typename SurfaceFormatKHRAllocator,
8436 typename std::enable_if<std::is_same<typename SurfaceFormatKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::SurfaceFormatKHR>::value,
int>::type>
8438 PhysicalDevice::getSurfaceFormatsKHR( VULKAN_HPP_NAMESPACE::SurfaceKHR surface,
8439 SurfaceFormatKHRAllocator & surfaceFormatKHRAllocator,
8440 Dispatch
const & d )
const
8443 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8444 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceSurfaceFormatsKHR &&
"Function <vkGetPhysicalDeviceSurfaceFormatsKHR> requires <VK_KHR_surface>" );
8447 std::vector<VULKAN_HPP_NAMESPACE::SurfaceFormatKHR, SurfaceFormatKHRAllocator> surfaceFormats( surfaceFormatKHRAllocator );
8448 uint32_t surfaceFormatCount;
8453 d.vkGetPhysicalDeviceSurfaceFormatsKHR( m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &surfaceFormatCount,
nullptr ) );
8454 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && surfaceFormatCount )
8456 surfaceFormats.resize( surfaceFormatCount );
8458 m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &surfaceFormatCount,
reinterpret_cast<VkSurfaceFormatKHR *
>( surfaceFormats.data() ) ) );
8460 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
8463 if ( surfaceFormatCount < surfaceFormats.size() )
8465 surfaceFormats.resize( surfaceFormatCount );
8471 template <
typename Dispatch>
8473 uint32_t * pPresentModeCount,
8478 return static_cast<Result>( d.vkGetPhysicalDeviceSurfacePresentModesKHR(
8482 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8483 template <
typename PresentModeKHRAllocator,
typename Dispatch>
8485 PhysicalDevice::getSurfacePresentModesKHR( VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch
const & d )
const
8488 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8489 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceSurfacePresentModesKHR &&
"Function <vkGetPhysicalDeviceSurfacePresentModesKHR> requires <VK_KHR_surface>" );
8492 std::vector<VULKAN_HPP_NAMESPACE::PresentModeKHR, PresentModeKHRAllocator> presentModes;
8493 uint32_t presentModeCount;
8498 d.vkGetPhysicalDeviceSurfacePresentModesKHR( m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &presentModeCount,
nullptr ) );
8499 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && presentModeCount )
8501 presentModes.resize( presentModeCount );
8503 m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &presentModeCount,
reinterpret_cast<VkPresentModeKHR *
>( presentModes.data() ) ) );
8505 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
8508 if ( presentModeCount < presentModes.size() )
8510 presentModes.resize( presentModeCount );
8515 template <
typename PresentModeKHRAllocator,
8517 typename std::enable_if<std::is_same<typename PresentModeKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::PresentModeKHR>::value,
int>::type>
8519 PhysicalDevice::getSurfacePresentModesKHR( VULKAN_HPP_NAMESPACE::SurfaceKHR surface,
8520 PresentModeKHRAllocator & presentModeKHRAllocator,
8521 Dispatch
const & d )
const
8524 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8525 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceSurfacePresentModesKHR &&
"Function <vkGetPhysicalDeviceSurfacePresentModesKHR> requires <VK_KHR_surface>" );
8528 std::vector<VULKAN_HPP_NAMESPACE::PresentModeKHR, PresentModeKHRAllocator> presentModes( presentModeKHRAllocator );
8529 uint32_t presentModeCount;
8534 d.vkGetPhysicalDeviceSurfacePresentModesKHR( m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &presentModeCount,
nullptr ) );
8535 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && presentModeCount )
8537 presentModes.resize( presentModeCount );
8539 m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &presentModeCount,
reinterpret_cast<VkPresentModeKHR *
>( presentModes.data() ) ) );
8541 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
8544 if ( presentModeCount < presentModes.size() )
8546 presentModes.resize( presentModeCount );
8554 template <
typename Dispatch>
8556 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
8557 VULKAN_HPP_NAMESPACE::SwapchainKHR * pSwapchain,
8561 return static_cast<Result>( d.vkCreateSwapchainKHR( m_device,
8567 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8568 template <
typename Dispatch>
8570 Device::createSwapchainKHR(
const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR & createInfo,
8572 Dispatch
const & d )
const
8575 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8576 VULKAN_HPP_ASSERT( d.vkCreateSwapchainKHR &&
"Function <vkCreateSwapchainKHR> requires <VK_KHR_swapchain>" );
8579 VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain;
8581 d.vkCreateSwapchainKHR( m_device,
8583 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
8590 # ifndef VULKAN_HPP_NO_SMART_HANDLE
8591 template <
typename Dispatch>
8593 Device::createSwapchainKHRUnique(
const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR & createInfo,
8595 Dispatch
const & d )
const
8598 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8599 VULKAN_HPP_ASSERT( d.vkCreateSwapchainKHR &&
"Function <vkCreateSwapchainKHR> requires <VK_KHR_swapchain>" );
8602 VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain;
8604 d.vkCreateSwapchainKHR( m_device,
8606 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
8616 template <
typename Dispatch>
8618 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
8625 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8626 template <
typename Dispatch>
8632 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8633 VULKAN_HPP_ASSERT( d.vkDestroySwapchainKHR &&
"Function <vkDestroySwapchainKHR> requires <VK_KHR_swapchain>" );
8636 d.vkDestroySwapchainKHR( m_device,
8638 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
8642 template <
typename Dispatch>
8644 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
8651 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8652 template <
typename Dispatch>
8658 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8659 VULKAN_HPP_ASSERT( d.vkDestroySwapchainKHR &&
"Function <vkDestroySwapchainKHR> requires <VK_KHR_swapchain>" );
8662 d.vkDestroySwapchainKHR( m_device,
8664 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
8668 template <
typename Dispatch>
8670 uint32_t * pSwapchainImageCount,
8671 VULKAN_HPP_NAMESPACE::Image * pSwapchainImages,
8675 return static_cast<Result>(
8676 d.vkGetSwapchainImagesKHR( m_device,
static_cast<VkSwapchainKHR>( swapchain ), pSwapchainImageCount,
reinterpret_cast<VkImage *
>( pSwapchainImages ) ) );
8679 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8680 template <
typename ImageAllocator,
typename Dispatch>
8682 Device::getSwapchainImagesKHR( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch
const & d )
const
8685 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8686 VULKAN_HPP_ASSERT( d.vkGetSwapchainImagesKHR &&
"Function <vkGetSwapchainImagesKHR> requires <VK_KHR_swapchain>" );
8689 std::vector<VULKAN_HPP_NAMESPACE::Image, ImageAllocator> swapchainImages;
8690 uint32_t swapchainImageCount;
8695 d.vkGetSwapchainImagesKHR( m_device,
static_cast<VkSwapchainKHR>( swapchain ), &swapchainImageCount,
nullptr ) );
8696 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && swapchainImageCount )
8698 swapchainImages.resize( swapchainImageCount );
8700 m_device,
static_cast<VkSwapchainKHR>( swapchain ), &swapchainImageCount,
reinterpret_cast<VkImage *
>( swapchainImages.data() ) ) );
8702 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
8705 if ( swapchainImageCount < swapchainImages.size() )
8707 swapchainImages.resize( swapchainImageCount );
8712 template <
typename ImageAllocator,
8714 typename std::enable_if<std::is_same<typename ImageAllocator::value_type, VULKAN_HPP_NAMESPACE::Image>::value,
int>::type>
8716 Device::getSwapchainImagesKHR( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, ImageAllocator & imageAllocator, Dispatch
const & d )
const
8719 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8720 VULKAN_HPP_ASSERT( d.vkGetSwapchainImagesKHR &&
"Function <vkGetSwapchainImagesKHR> requires <VK_KHR_swapchain>" );
8723 std::vector<VULKAN_HPP_NAMESPACE::Image, ImageAllocator> swapchainImages( imageAllocator );
8724 uint32_t swapchainImageCount;
8729 d.vkGetSwapchainImagesKHR( m_device,
static_cast<VkSwapchainKHR>( swapchain ), &swapchainImageCount,
nullptr ) );
8730 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && swapchainImageCount )
8732 swapchainImages.resize( swapchainImageCount );
8734 m_device,
static_cast<VkSwapchainKHR>( swapchain ), &swapchainImageCount,
reinterpret_cast<VkImage *
>( swapchainImages.data() ) ) );
8736 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
8739 if ( swapchainImageCount < swapchainImages.size() )
8741 swapchainImages.resize( swapchainImageCount );
8747 template <
typename Dispatch>
8750 VULKAN_HPP_NAMESPACE::Semaphore semaphore,
8751 VULKAN_HPP_NAMESPACE::Fence fence,
8752 uint32_t * pImageIndex,
8756 return static_cast<Result>( d.vkAcquireNextImageKHR(
8760 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8761 template <
typename Dispatch>
8764 VULKAN_HPP_NAMESPACE::Semaphore semaphore,
8765 VULKAN_HPP_NAMESPACE::Fence fence,
8766 Dispatch
const & d )
const
8769 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8770 VULKAN_HPP_ASSERT( d.vkAcquireNextImageKHR &&
"Function <vkAcquireNextImageKHR> requires <VK_KHR_swapchain>" );
8773 uint32_t imageIndex;
8778 { VULKAN_HPP_NAMESPACE::Result::eSuccess,
8779 VULKAN_HPP_NAMESPACE::Result::eTimeout,
8780 VULKAN_HPP_NAMESPACE::Result::eNotReady,
8781 VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR } );
8787 template <
typename Dispatch>
8792 return static_cast<Result>( d.vkQueuePresentKHR( m_queue,
reinterpret_cast<const VkPresentInfoKHR *
>( pPresentInfo ) ) );
8795 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8796 template <
typename Dispatch>
8798 Dispatch
const & d )
const
8801 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8802 VULKAN_HPP_ASSERT( d.vkQueuePresentKHR &&
"Function <vkQueuePresentKHR> requires <VK_KHR_swapchain>" );
8808 result,
VULKAN_HPP_NAMESPACE_STRING "::Queue::presentKHR", { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR } );
8814 template <
typename Dispatch>
8816 VULKAN_HPP_NAMESPACE::DeviceGroupPresentCapabilitiesKHR * pDeviceGroupPresentCapabilities, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
8819 return static_cast<Result>(
8823 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8824 template <
typename Dispatch>
8826 Device::getGroupPresentCapabilitiesKHR( Dispatch
const & d )
const
8829 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8831 "Function <vkGetDeviceGroupPresentCapabilitiesKHR> requires <VK_KHR_device_group> or <VK_KHR_swapchain>" );
8834 VULKAN_HPP_NAMESPACE::DeviceGroupPresentCapabilitiesKHR deviceGroupPresentCapabilities;
8843 template <
typename Dispatch>
8849 return static_cast<Result>( d.vkGetDeviceGroupSurfacePresentModesKHR(
8853 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8854 template <
typename Dispatch>
8856 Device::getGroupSurfacePresentModesKHR( VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch
const & d )
const
8859 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8861 "Function <vkGetDeviceGroupSurfacePresentModesKHR> requires <VK_KHR_device_group> or <VK_KHR_swapchain>" );
8873 template <
typename Dispatch>
8875 uint32_t * pRectCount,
8876 VULKAN_HPP_NAMESPACE::Rect2D * pRects,
8880 return static_cast<Result>(
8881 d.vkGetPhysicalDevicePresentRectanglesKHR( m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), pRectCount,
reinterpret_cast<VkRect2D *
>( pRects ) ) );
8884 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8885 template <
typename Rect2DAllocator,
typename Dispatch>
8887 PhysicalDevice::getPresentRectanglesKHR( VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch
const & d )
const
8890 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8892 "Function <vkGetPhysicalDevicePresentRectanglesKHR> requires <VK_KHR_device_group> or <VK_KHR_swapchain>" );
8895 std::vector<VULKAN_HPP_NAMESPACE::Rect2D, Rect2DAllocator> rects;
8901 d.vkGetPhysicalDevicePresentRectanglesKHR( m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &rectCount,
nullptr ) );
8902 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && rectCount )
8904 rects.resize( rectCount );
8906 m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &rectCount,
reinterpret_cast<VkRect2D *
>( rects.data() ) ) );
8908 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
8911 if ( rectCount < rects.size() )
8913 rects.resize( rectCount );
8918 template <
typename Rect2DAllocator,
8920 typename std::enable_if<std::is_same<typename Rect2DAllocator::value_type, VULKAN_HPP_NAMESPACE::Rect2D>::value,
int>::type>
8922 PhysicalDevice::getPresentRectanglesKHR( VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Rect2DAllocator & rect2DAllocator, Dispatch
const & d )
const
8925 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8927 "Function <vkGetPhysicalDevicePresentRectanglesKHR> requires <VK_KHR_device_group> or <VK_KHR_swapchain>" );
8930 std::vector<VULKAN_HPP_NAMESPACE::Rect2D, Rect2DAllocator> rects( rect2DAllocator );
8936 d.vkGetPhysicalDevicePresentRectanglesKHR( m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &rectCount,
nullptr ) );
8937 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && rectCount )
8939 rects.resize( rectCount );
8941 m_physicalDevice,
static_cast<VkSurfaceKHR>( surface ), &rectCount,
reinterpret_cast<VkRect2D *
>( rects.data() ) ) );
8943 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
8946 if ( rectCount < rects.size() )
8948 rects.resize( rectCount );
8954 template <
typename Dispatch>
8956 uint32_t * pImageIndex,
8963 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
8964 template <
typename Dispatch>
8966 Dispatch
const & d )
const
8969 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
8970 VULKAN_HPP_ASSERT( d.vkAcquireNextImage2KHR &&
"Function <vkAcquireNextImage2KHR> requires <VK_KHR_device_group> or <VK_KHR_swapchain>" );
8973 uint32_t imageIndex;
8975 d.vkAcquireNextImage2KHR( m_device,
reinterpret_cast<const VkAcquireNextImageInfoKHR *
>( &acquireInfo ), &imageIndex ) );
8978 { VULKAN_HPP_NAMESPACE::Result::eSuccess,
8979 VULKAN_HPP_NAMESPACE::Result::eTimeout,
8980 VULKAN_HPP_NAMESPACE::Result::eNotReady,
8981 VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR } );
8989 template <
typename Dispatch>
8991 VULKAN_HPP_NAMESPACE::DisplayPropertiesKHR * pProperties,
8995 return static_cast<Result>(
8996 d.vkGetPhysicalDeviceDisplayPropertiesKHR( m_physicalDevice, pPropertyCount,
reinterpret_cast<VkDisplayPropertiesKHR *
>( pProperties ) ) );
8999 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9000 template <
typename DisplayPropertiesKHRAllocator,
typename Dispatch>
9002 PhysicalDevice::getDisplayPropertiesKHR( Dispatch
const & d )
const
9005 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9006 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceDisplayPropertiesKHR &&
"Function <vkGetPhysicalDeviceDisplayPropertiesKHR> requires <VK_KHR_display>" );
9009 std::vector<VULKAN_HPP_NAMESPACE::DisplayPropertiesKHR, DisplayPropertiesKHRAllocator> properties;
9010 uint32_t propertyCount;
9014 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceDisplayPropertiesKHR( m_physicalDevice, &propertyCount,
nullptr ) );
9015 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
9017 properties.resize( propertyCount );
9019 d.vkGetPhysicalDeviceDisplayPropertiesKHR( m_physicalDevice, &propertyCount,
reinterpret_cast<VkDisplayPropertiesKHR *
>( properties.data() ) ) );
9021 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
9024 if ( propertyCount < properties.size() )
9026 properties.resize( propertyCount );
9032 typename DisplayPropertiesKHRAllocator,
9034 typename std::enable_if<std::is_same<typename DisplayPropertiesKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::DisplayPropertiesKHR>::value,
int>::type>
9036 PhysicalDevice::getDisplayPropertiesKHR( DisplayPropertiesKHRAllocator & displayPropertiesKHRAllocator, Dispatch
const & d )
const
9039 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9040 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceDisplayPropertiesKHR &&
"Function <vkGetPhysicalDeviceDisplayPropertiesKHR> requires <VK_KHR_display>" );
9043 std::vector<VULKAN_HPP_NAMESPACE::DisplayPropertiesKHR, DisplayPropertiesKHRAllocator> properties( displayPropertiesKHRAllocator );
9044 uint32_t propertyCount;
9048 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceDisplayPropertiesKHR( m_physicalDevice, &propertyCount,
nullptr ) );
9049 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
9051 properties.resize( propertyCount );
9053 d.vkGetPhysicalDeviceDisplayPropertiesKHR( m_physicalDevice, &propertyCount,
reinterpret_cast<VkDisplayPropertiesKHR *
>( properties.data() ) ) );
9055 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
9058 if ( propertyCount < properties.size() )
9060 properties.resize( propertyCount );
9066 template <
typename Dispatch>
9068 VULKAN_HPP_NAMESPACE::DisplayPlanePropertiesKHR * pProperties,
9072 return static_cast<Result>(
9073 d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR( m_physicalDevice, pPropertyCount,
reinterpret_cast<VkDisplayPlanePropertiesKHR *
>( pProperties ) ) );
9076 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9077 template <
typename DisplayPlanePropertiesKHRAllocator,
typename Dispatch>
9080 PhysicalDevice::getDisplayPlanePropertiesKHR( Dispatch
const & d )
const
9083 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9084 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR &&
"Function <vkGetPhysicalDeviceDisplayPlanePropertiesKHR> requires <VK_KHR_display>" );
9087 std::vector<VULKAN_HPP_NAMESPACE::DisplayPlanePropertiesKHR, DisplayPlanePropertiesKHRAllocator> properties;
9088 uint32_t propertyCount;
9092 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR( m_physicalDevice, &propertyCount,
nullptr ) );
9093 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
9095 properties.resize( propertyCount );
9099 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
9102 if ( propertyCount < properties.size() )
9104 properties.resize( propertyCount );
9110 typename DisplayPlanePropertiesKHRAllocator,
9112 typename std::enable_if<std::is_same<typename DisplayPlanePropertiesKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::DisplayPlanePropertiesKHR>::value,
9116 PhysicalDevice::getDisplayPlanePropertiesKHR( DisplayPlanePropertiesKHRAllocator & displayPlanePropertiesKHRAllocator, Dispatch
const & d )
const
9119 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9120 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR &&
"Function <vkGetPhysicalDeviceDisplayPlanePropertiesKHR> requires <VK_KHR_display>" );
9123 std::vector<VULKAN_HPP_NAMESPACE::DisplayPlanePropertiesKHR, DisplayPlanePropertiesKHRAllocator> properties( displayPlanePropertiesKHRAllocator );
9124 uint32_t propertyCount;
9128 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR( m_physicalDevice, &propertyCount,
nullptr ) );
9129 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
9131 properties.resize( propertyCount );
9135 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
9138 if ( propertyCount < properties.size() )
9140 properties.resize( propertyCount );
9146 template <
typename Dispatch>
9148 uint32_t * pDisplayCount,
9149 VULKAN_HPP_NAMESPACE::DisplayKHR * pDisplays,
9153 return static_cast<Result>(
9154 d.vkGetDisplayPlaneSupportedDisplaysKHR( m_physicalDevice, planeIndex, pDisplayCount,
reinterpret_cast<VkDisplayKHR *
>( pDisplays ) ) );
9157 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9158 template <
typename DisplayKHRAllocator,
typename Dispatch>
9160 PhysicalDevice::getDisplayPlaneSupportedDisplaysKHR( uint32_t planeIndex, Dispatch
const & d )
const
9163 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9164 VULKAN_HPP_ASSERT( d.vkGetDisplayPlaneSupportedDisplaysKHR &&
"Function <vkGetDisplayPlaneSupportedDisplaysKHR> requires <VK_KHR_display>" );
9167 std::vector<VULKAN_HPP_NAMESPACE::DisplayKHR, DisplayKHRAllocator> displays;
9168 uint32_t displayCount;
9172 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetDisplayPlaneSupportedDisplaysKHR( m_physicalDevice, planeIndex, &displayCount,
nullptr ) );
9173 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && displayCount )
9175 displays.resize( displayCount );
9177 d.vkGetDisplayPlaneSupportedDisplaysKHR( m_physicalDevice, planeIndex, &displayCount,
reinterpret_cast<VkDisplayKHR *
>( displays.data() ) ) );
9179 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
9182 if ( displayCount < displays.size() )
9184 displays.resize( displayCount );
9189 template <
typename DisplayKHRAllocator,
9191 typename std::enable_if<std::is_same<typename DisplayKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::DisplayKHR>::value,
int>::type>
9193 PhysicalDevice::getDisplayPlaneSupportedDisplaysKHR( uint32_t planeIndex, DisplayKHRAllocator & displayKHRAllocator, Dispatch
const & d )
const
9196 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9197 VULKAN_HPP_ASSERT( d.vkGetDisplayPlaneSupportedDisplaysKHR &&
"Function <vkGetDisplayPlaneSupportedDisplaysKHR> requires <VK_KHR_display>" );
9200 std::vector<VULKAN_HPP_NAMESPACE::DisplayKHR, DisplayKHRAllocator> displays( displayKHRAllocator );
9201 uint32_t displayCount;
9205 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetDisplayPlaneSupportedDisplaysKHR( m_physicalDevice, planeIndex, &displayCount,
nullptr ) );
9206 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && displayCount )
9208 displays.resize( displayCount );
9210 d.vkGetDisplayPlaneSupportedDisplaysKHR( m_physicalDevice, planeIndex, &displayCount,
reinterpret_cast<VkDisplayKHR *
>( displays.data() ) ) );
9212 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
9215 if ( displayCount < displays.size() )
9217 displays.resize( displayCount );
9223 template <
typename Dispatch>
9225 uint32_t * pPropertyCount,
9226 VULKAN_HPP_NAMESPACE::DisplayModePropertiesKHR * pProperties,
9230 return static_cast<Result>( d.vkGetDisplayModePropertiesKHR(
9234 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9235 template <
typename DisplayModePropertiesKHRAllocator,
typename Dispatch>
9238 PhysicalDevice::getDisplayModePropertiesKHR( VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch
const & d )
const
9241 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9242 VULKAN_HPP_ASSERT( d.vkGetDisplayModePropertiesKHR &&
"Function <vkGetDisplayModePropertiesKHR> requires <VK_KHR_display>" );
9245 std::vector<VULKAN_HPP_NAMESPACE::DisplayModePropertiesKHR, DisplayModePropertiesKHRAllocator> properties;
9246 uint32_t propertyCount;
9251 d.vkGetDisplayModePropertiesKHR( m_physicalDevice,
static_cast<VkDisplayKHR>( display ), &propertyCount,
nullptr ) );
9252 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
9254 properties.resize( propertyCount );
9258 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
9261 if ( propertyCount < properties.size() )
9263 properties.resize( propertyCount );
9268 template <
typename DisplayModePropertiesKHRAllocator,
9270 typename std::enable_if<std::is_same<typename DisplayModePropertiesKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::DisplayModePropertiesKHR>::value,
9274 PhysicalDevice::getDisplayModePropertiesKHR( VULKAN_HPP_NAMESPACE::DisplayKHR display,
9275 DisplayModePropertiesKHRAllocator & displayModePropertiesKHRAllocator,
9276 Dispatch
const & d )
const
9279 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9280 VULKAN_HPP_ASSERT( d.vkGetDisplayModePropertiesKHR &&
"Function <vkGetDisplayModePropertiesKHR> requires <VK_KHR_display>" );
9283 std::vector<VULKAN_HPP_NAMESPACE::DisplayModePropertiesKHR, DisplayModePropertiesKHRAllocator> properties( displayModePropertiesKHRAllocator );
9284 uint32_t propertyCount;
9289 d.vkGetDisplayModePropertiesKHR( m_physicalDevice,
static_cast<VkDisplayKHR>( display ), &propertyCount,
nullptr ) );
9290 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
9292 properties.resize( propertyCount );
9296 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
9299 if ( propertyCount < properties.size() )
9301 properties.resize( propertyCount );
9307 template <
typename Dispatch>
9309 const VULKAN_HPP_NAMESPACE::DisplayModeCreateInfoKHR * pCreateInfo,
9310 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
9311 VULKAN_HPP_NAMESPACE::DisplayModeKHR * pMode,
9315 return static_cast<Result>( d.vkCreateDisplayModeKHR( m_physicalDevice,
9322 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9323 template <
typename Dispatch>
9325 PhysicalDevice::createDisplayModeKHR( VULKAN_HPP_NAMESPACE::DisplayKHR display,
9326 const VULKAN_HPP_NAMESPACE::DisplayModeCreateInfoKHR & createInfo,
9328 Dispatch
const & d )
const
9331 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9332 VULKAN_HPP_ASSERT( d.vkCreateDisplayModeKHR &&
"Function <vkCreateDisplayModeKHR> requires <VK_KHR_display>" );
9335 VULKAN_HPP_NAMESPACE::DisplayModeKHR mode;
9337 d.vkCreateDisplayModeKHR( m_physicalDevice,
9340 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9347 # ifndef VULKAN_HPP_NO_SMART_HANDLE
9348 template <
typename Dispatch>
9350 PhysicalDevice::createDisplayModeKHRUnique( VULKAN_HPP_NAMESPACE::DisplayKHR display,
9351 const VULKAN_HPP_NAMESPACE::DisplayModeCreateInfoKHR & createInfo,
9353 Dispatch
const & d )
const
9356 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9357 VULKAN_HPP_ASSERT( d.vkCreateDisplayModeKHR &&
"Function <vkCreateDisplayModeKHR> requires <VK_KHR_display>" );
9360 VULKAN_HPP_NAMESPACE::DisplayModeKHR mode;
9362 d.vkCreateDisplayModeKHR( m_physicalDevice,
9365 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9375 template <
typename Dispatch>
9377 PhysicalDevice::getDisplayPlaneCapabilitiesKHR( VULKAN_HPP_NAMESPACE::DisplayModeKHR mode,
9378 uint32_t planeIndex,
9379 VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilitiesKHR * pCapabilities,
9383 return static_cast<Result>( d.vkGetDisplayPlaneCapabilitiesKHR(
9387 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9388 template <
typename Dispatch>
9390 PhysicalDevice::getDisplayPlaneCapabilitiesKHR( VULKAN_HPP_NAMESPACE::DisplayModeKHR mode, uint32_t planeIndex, Dispatch
const & d )
const
9393 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9394 VULKAN_HPP_ASSERT( d.vkGetDisplayPlaneCapabilitiesKHR &&
"Function <vkGetDisplayPlaneCapabilitiesKHR> requires <VK_KHR_display>" );
9397 VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilitiesKHR capabilities;
9406 template <
typename Dispatch>
9408 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
9409 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
9413 return static_cast<Result>( d.vkCreateDisplayPlaneSurfaceKHR( m_instance,
9419 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9420 template <
typename Dispatch>
9422 Instance::createDisplayPlaneSurfaceKHR(
const VULKAN_HPP_NAMESPACE::DisplaySurfaceCreateInfoKHR & createInfo,
9424 Dispatch
const & d )
const
9427 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9428 VULKAN_HPP_ASSERT( d.vkCreateDisplayPlaneSurfaceKHR &&
"Function <vkCreateDisplayPlaneSurfaceKHR> requires <VK_KHR_display>" );
9431 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9435 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9442 # ifndef VULKAN_HPP_NO_SMART_HANDLE
9443 template <
typename Dispatch>
9445 Instance::createDisplayPlaneSurfaceKHRUnique(
const VULKAN_HPP_NAMESPACE::DisplaySurfaceCreateInfoKHR & createInfo,
9447 Dispatch
const & d )
const
9450 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9451 VULKAN_HPP_ASSERT( d.vkCreateDisplayPlaneSurfaceKHR &&
"Function <vkCreateDisplayPlaneSurfaceKHR> requires <VK_KHR_display>" );
9454 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9458 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9470 template <
typename Dispatch>
9472 const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR * pCreateInfos,
9473 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
9474 VULKAN_HPP_NAMESPACE::SwapchainKHR * pSwapchains,
9478 return static_cast<Result>( d.vkCreateSharedSwapchainsKHR( m_device,
9485 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9486 template <
typename SwapchainKHRAllocator,
typename Dispatch>
9488 Device::createSharedSwapchainsKHR( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR>
const & createInfos,
9490 Dispatch
const & d )
const
9493 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9494 VULKAN_HPP_ASSERT( d.vkCreateSharedSwapchainsKHR &&
"Function <vkCreateSharedSwapchainsKHR> requires <VK_KHR_display_swapchain>" );
9497 std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR, SwapchainKHRAllocator> swapchains( createInfos.size() );
9502 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9509 template <
typename SwapchainKHRAllocator,
9511 typename std::enable_if<std::is_same<typename SwapchainKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::SwapchainKHR>::value,
int>::type>
9513 Device::createSharedSwapchainsKHR( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR>
const & createInfos,
9515 SwapchainKHRAllocator & swapchainKHRAllocator,
9516 Dispatch
const & d )
const
9519 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9520 VULKAN_HPP_ASSERT( d.vkCreateSharedSwapchainsKHR &&
"Function <vkCreateSharedSwapchainsKHR> requires <VK_KHR_display_swapchain>" );
9523 std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR, SwapchainKHRAllocator> swapchains( createInfos.size(), swapchainKHRAllocator );
9528 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9535 template <
typename Dispatch>
9537 Device::createSharedSwapchainKHR(
const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR & createInfo,
9539 Dispatch
const & d )
const
9542 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9543 VULKAN_HPP_ASSERT( d.vkCreateSharedSwapchainsKHR &&
"Function <vkCreateSharedSwapchainsKHR> requires <VK_KHR_display_swapchain>" );
9546 VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain;
9551 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9558 # ifndef VULKAN_HPP_NO_SMART_HANDLE
9559 template <
typename Dispatch,
typename SwapchainKHRAllocator>
9562 Device::createSharedSwapchainsKHRUnique( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR>
const & createInfos,
9564 Dispatch
const & d )
const
9567 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9568 VULKAN_HPP_ASSERT( d.vkCreateSharedSwapchainsKHR &&
"Function <vkCreateSharedSwapchainsKHR> requires <VK_KHR_display_swapchain>" );
9571 std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR> swapchains( createInfos.size() );
9576 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9579 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::SwapchainKHR, Dispatch>, SwapchainKHRAllocator> uniqueSwapchains;
9580 uniqueSwapchains.reserve( createInfos.size() );
9582 for (
auto const & swapchain : swapchains )
9589 template <
typename Dispatch,
9590 typename SwapchainKHRAllocator,
9591 typename std::enable_if<std::is_same<typename SwapchainKHRAllocator::value_type, UniqueHandle<VULKAN_HPP_NAMESPACE::SwapchainKHR, Dispatch>>::value,
9595 Device::createSharedSwapchainsKHRUnique( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR>
const & createInfos,
9597 SwapchainKHRAllocator & swapchainKHRAllocator,
9598 Dispatch
const & d )
const
9601 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9602 VULKAN_HPP_ASSERT( d.vkCreateSharedSwapchainsKHR &&
"Function <vkCreateSharedSwapchainsKHR> requires <VK_KHR_display_swapchain>" );
9605 std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR> swapchains( createInfos.size() );
9610 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9613 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::SwapchainKHR, Dispatch>, SwapchainKHRAllocator> uniqueSwapchains( swapchainKHRAllocator );
9614 uniqueSwapchains.reserve( createInfos.size() );
9616 for (
auto const & swapchain : swapchains )
9623 template <
typename Dispatch>
9625 Device::createSharedSwapchainKHRUnique(
const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR & createInfo,
9627 Dispatch
const & d )
const
9630 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9631 VULKAN_HPP_ASSERT( d.vkCreateSharedSwapchainsKHR &&
"Function <vkCreateSharedSwapchainsKHR> requires <VK_KHR_display_swapchain>" );
9634 VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain;
9639 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9649 #if defined( VK_USE_PLATFORM_XLIB_KHR )
9652 template <
typename Dispatch>
9654 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
9655 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
9659 return static_cast<Result>( d.vkCreateXlibSurfaceKHR( m_instance,
9660 reinterpret_cast<const VkXlibSurfaceCreateInfoKHR *
>( pCreateInfo ),
9665 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9666 template <
typename Dispatch>
9668 Instance::createXlibSurfaceKHR(
const VULKAN_HPP_NAMESPACE::XlibSurfaceCreateInfoKHR & createInfo,
9669 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
9670 Dispatch
const & d )
const
9673 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9674 VULKAN_HPP_ASSERT( d.vkCreateXlibSurfaceKHR &&
"Function <vkCreateXlibSurfaceKHR> requires <VK_KHR_xlib_surface>" );
9677 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9679 d.vkCreateXlibSurfaceKHR( m_instance,
9680 reinterpret_cast<const VkXlibSurfaceCreateInfoKHR *
>( &createInfo ),
9681 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9688 # ifndef VULKAN_HPP_NO_SMART_HANDLE
9689 template <
typename Dispatch>
9691 Instance::createXlibSurfaceKHRUnique(
const VULKAN_HPP_NAMESPACE::XlibSurfaceCreateInfoKHR & createInfo,
9692 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
9693 Dispatch
const & d )
const
9696 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9697 VULKAN_HPP_ASSERT( d.vkCreateXlibSurfaceKHR &&
"Function <vkCreateXlibSurfaceKHR> requires <VK_KHR_xlib_surface>" );
9700 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9702 d.vkCreateXlibSurfaceKHR( m_instance,
9703 reinterpret_cast<const VkXlibSurfaceCreateInfoKHR *
>( &createInfo ),
9704 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9709 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
9714 template <
typename Dispatch>
9716 PhysicalDevice::getXlibPresentationSupportKHR( uint32_t queueFamilyIndex, Display * dpy, VisualID visualID, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
9719 return static_cast<Bool32>( d.vkGetPhysicalDeviceXlibPresentationSupportKHR( m_physicalDevice, queueFamilyIndex, dpy, visualID ) );
9722 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9723 template <
typename Dispatch>
9725 PhysicalDevice::getXlibPresentationSupportKHR( uint32_t queueFamilyIndex, Display & dpy, VisualID visualID, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
9728 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9730 "Function <vkGetPhysicalDeviceXlibPresentationSupportKHR> requires <VK_KHR_xlib_surface>" );
9733 VkBool32 result = d.vkGetPhysicalDeviceXlibPresentationSupportKHR( m_physicalDevice, queueFamilyIndex, &dpy, visualID );
9740 #if defined( VK_USE_PLATFORM_XCB_KHR )
9743 template <
typename Dispatch>
9745 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
9746 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
9750 return static_cast<Result>( d.vkCreateXcbSurfaceKHR( m_instance,
9751 reinterpret_cast<const VkXcbSurfaceCreateInfoKHR *
>( pCreateInfo ),
9756 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9757 template <
typename Dispatch>
9759 Instance::createXcbSurfaceKHR(
const VULKAN_HPP_NAMESPACE::XcbSurfaceCreateInfoKHR & createInfo,
9760 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
9761 Dispatch
const & d )
const
9764 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9765 VULKAN_HPP_ASSERT( d.vkCreateXcbSurfaceKHR &&
"Function <vkCreateXcbSurfaceKHR> requires <VK_KHR_xcb_surface>" );
9768 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9770 d.vkCreateXcbSurfaceKHR( m_instance,
9771 reinterpret_cast<const VkXcbSurfaceCreateInfoKHR *
>( &createInfo ),
9772 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9779 # ifndef VULKAN_HPP_NO_SMART_HANDLE
9780 template <
typename Dispatch>
9782 Instance::createXcbSurfaceKHRUnique(
const VULKAN_HPP_NAMESPACE::XcbSurfaceCreateInfoKHR & createInfo,
9783 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
9784 Dispatch
const & d )
const
9787 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9788 VULKAN_HPP_ASSERT( d.vkCreateXcbSurfaceKHR &&
"Function <vkCreateXcbSurfaceKHR> requires <VK_KHR_xcb_surface>" );
9791 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9793 d.vkCreateXcbSurfaceKHR( m_instance,
9794 reinterpret_cast<const VkXcbSurfaceCreateInfoKHR *
>( &createInfo ),
9795 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9800 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
9805 template <
typename Dispatch>
9807 xcb_connection_t * connection,
9808 xcb_visualid_t visual_id,
9812 return static_cast<Bool32>( d.vkGetPhysicalDeviceXcbPresentationSupportKHR( m_physicalDevice, queueFamilyIndex, connection, visual_id ) );
9815 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9816 template <
typename Dispatch>
9818 xcb_connection_t & connection,
9819 xcb_visualid_t visual_id,
9823 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9825 "Function <vkGetPhysicalDeviceXcbPresentationSupportKHR> requires <VK_KHR_xcb_surface>" );
9828 VkBool32 result = d.vkGetPhysicalDeviceXcbPresentationSupportKHR( m_physicalDevice, queueFamilyIndex, &connection, visual_id );
9835 #if defined( VK_USE_PLATFORM_WAYLAND_KHR )
9838 template <
typename Dispatch>
9840 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
9841 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
9845 return static_cast<Result>( d.vkCreateWaylandSurfaceKHR( m_instance,
9846 reinterpret_cast<const VkWaylandSurfaceCreateInfoKHR *
>( pCreateInfo ),
9851 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9852 template <
typename Dispatch>
9854 Instance::createWaylandSurfaceKHR(
const VULKAN_HPP_NAMESPACE::WaylandSurfaceCreateInfoKHR & createInfo,
9855 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
9856 Dispatch
const & d )
const
9859 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9860 VULKAN_HPP_ASSERT( d.vkCreateWaylandSurfaceKHR &&
"Function <vkCreateWaylandSurfaceKHR> requires <VK_KHR_wayland_surface>" );
9863 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9866 reinterpret_cast<const VkWaylandSurfaceCreateInfoKHR *
>( &createInfo ),
9867 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9874 # ifndef VULKAN_HPP_NO_SMART_HANDLE
9875 template <
typename Dispatch>
9877 Instance::createWaylandSurfaceKHRUnique(
const VULKAN_HPP_NAMESPACE::WaylandSurfaceCreateInfoKHR & createInfo,
9878 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
9879 Dispatch
const & d )
const
9882 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9883 VULKAN_HPP_ASSERT( d.vkCreateWaylandSurfaceKHR &&
"Function <vkCreateWaylandSurfaceKHR> requires <VK_KHR_wayland_surface>" );
9886 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9889 reinterpret_cast<const VkWaylandSurfaceCreateInfoKHR *
>( &createInfo ),
9890 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9895 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
9900 template <
typename Dispatch>
9902 struct wl_display * display,
9906 return static_cast<Bool32>( d.vkGetPhysicalDeviceWaylandPresentationSupportKHR( m_physicalDevice, queueFamilyIndex, display ) );
9909 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9910 template <
typename Dispatch>
9912 PhysicalDevice::getWaylandPresentationSupportKHR( uint32_t queueFamilyIndex,
struct wl_display & display, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
9915 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9917 "Function <vkGetPhysicalDeviceWaylandPresentationSupportKHR> requires <VK_KHR_wayland_surface>" );
9920 VkBool32 result = d.vkGetPhysicalDeviceWaylandPresentationSupportKHR( m_physicalDevice, queueFamilyIndex, &display );
9927 #if defined( VK_USE_PLATFORM_ANDROID_KHR )
9930 template <
typename Dispatch>
9932 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
9933 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
9937 return static_cast<Result>( d.vkCreateAndroidSurfaceKHR( m_instance,
9938 reinterpret_cast<const VkAndroidSurfaceCreateInfoKHR *
>( pCreateInfo ),
9943 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
9944 template <
typename Dispatch>
9946 Instance::createAndroidSurfaceKHR(
const VULKAN_HPP_NAMESPACE::AndroidSurfaceCreateInfoKHR & createInfo,
9947 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
9948 Dispatch
const & d )
const
9951 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9952 VULKAN_HPP_ASSERT( d.vkCreateAndroidSurfaceKHR &&
"Function <vkCreateAndroidSurfaceKHR> requires <VK_KHR_android_surface>" );
9955 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9958 reinterpret_cast<const VkAndroidSurfaceCreateInfoKHR *
>( &createInfo ),
9959 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9966 # ifndef VULKAN_HPP_NO_SMART_HANDLE
9967 template <
typename Dispatch>
9969 Instance::createAndroidSurfaceKHRUnique(
const VULKAN_HPP_NAMESPACE::AndroidSurfaceCreateInfoKHR & createInfo,
9970 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
9971 Dispatch
const & d )
const
9974 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
9975 VULKAN_HPP_ASSERT( d.vkCreateAndroidSurfaceKHR &&
"Function <vkCreateAndroidSurfaceKHR> requires <VK_KHR_android_surface>" );
9978 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
9981 reinterpret_cast<const VkAndroidSurfaceCreateInfoKHR *
>( &createInfo ),
9982 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
9987 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
9993 #if defined( VK_USE_PLATFORM_WIN32_KHR )
9996 template <
typename Dispatch>
9998 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
9999 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
10003 return static_cast<Result>( d.vkCreateWin32SurfaceKHR( m_instance,
10004 reinterpret_cast<const VkWin32SurfaceCreateInfoKHR *
>( pCreateInfo ),
10009 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10010 template <
typename Dispatch>
10012 Instance::createWin32SurfaceKHR(
const VULKAN_HPP_NAMESPACE::Win32SurfaceCreateInfoKHR & createInfo,
10013 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
10014 Dispatch
const & d )
const
10017 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10018 VULKAN_HPP_ASSERT( d.vkCreateWin32SurfaceKHR &&
"Function <vkCreateWin32SurfaceKHR> requires <VK_KHR_win32_surface>" );
10021 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
10023 d.vkCreateWin32SurfaceKHR( m_instance,
10024 reinterpret_cast<const VkWin32SurfaceCreateInfoKHR *
>( &createInfo ),
10025 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
10032 # ifndef VULKAN_HPP_NO_SMART_HANDLE
10033 template <
typename Dispatch>
10035 Instance::createWin32SurfaceKHRUnique(
const VULKAN_HPP_NAMESPACE::Win32SurfaceCreateInfoKHR & createInfo,
10036 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
10037 Dispatch
const & d )
const
10040 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10041 VULKAN_HPP_ASSERT( d.vkCreateWin32SurfaceKHR &&
"Function <vkCreateWin32SurfaceKHR> requires <VK_KHR_win32_surface>" );
10044 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
10046 d.vkCreateWin32SurfaceKHR( m_instance,
10047 reinterpret_cast<const VkWin32SurfaceCreateInfoKHR *
>( &createInfo ),
10048 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
10053 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
10058 template <
typename Dispatch>
10062 return static_cast<Bool32>( d.vkGetPhysicalDeviceWin32PresentationSupportKHR( m_physicalDevice, queueFamilyIndex ) );
10068 template <
typename Dispatch>
10070 Instance::createDebugReportCallbackEXT(
const VULKAN_HPP_NAMESPACE::DebugReportCallbackCreateInfoEXT * pCreateInfo,
10071 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
10072 VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT * pCallback,
10076 return static_cast<Result>( d.vkCreateDebugReportCallbackEXT( m_instance,
10082 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10083 template <
typename Dispatch>
10085 Instance::createDebugReportCallbackEXT(
const VULKAN_HPP_NAMESPACE::DebugReportCallbackCreateInfoEXT & createInfo,
10087 Dispatch
const & d )
const
10090 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10091 VULKAN_HPP_ASSERT( d.vkCreateDebugReportCallbackEXT &&
"Function <vkCreateDebugReportCallbackEXT> requires <VK_EXT_debug_report>" );
10094 VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback;
10098 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
10105 # ifndef VULKAN_HPP_NO_SMART_HANDLE
10106 template <
typename Dispatch>
10108 Instance::createDebugReportCallbackEXTUnique(
const VULKAN_HPP_NAMESPACE::DebugReportCallbackCreateInfoEXT & createInfo,
10110 Dispatch
const & d )
const
10113 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10114 VULKAN_HPP_ASSERT( d.vkCreateDebugReportCallbackEXT &&
"Function <vkCreateDebugReportCallbackEXT> requires <VK_EXT_debug_report>" );
10117 VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback;
10121 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
10131 template <
typename Dispatch>
10132 VULKAN_HPP_INLINE void Instance::destroyDebugReportCallbackEXT( VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback,
10133 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
10137 d.vkDestroyDebugReportCallbackEXT(
10141 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10142 template <
typename Dispatch>
10143 VULKAN_HPP_INLINE void Instance::destroyDebugReportCallbackEXT( VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback,
10148 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10149 VULKAN_HPP_ASSERT( d.vkDestroyDebugReportCallbackEXT &&
"Function <vkDestroyDebugReportCallbackEXT> requires <VK_EXT_debug_report>" );
10152 d.vkDestroyDebugReportCallbackEXT(
10155 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
10159 template <
typename Dispatch>
10161 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
10165 d.vkDestroyDebugReportCallbackEXT(
10169 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10170 template <
typename Dispatch>
10176 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10177 VULKAN_HPP_ASSERT( d.vkDestroyDebugReportCallbackEXT &&
"Function <vkDestroyDebugReportCallbackEXT> requires <VK_EXT_debug_report>" );
10180 d.vkDestroyDebugReportCallbackEXT(
10183 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
10187 template <
typename Dispatch>
10192 int32_t messageCode,
10193 const char * pLayerPrefix,
10194 const char * pMessage,
10198 d.vkDebugReportMessageEXT( m_instance,
10208 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10209 template <
typename Dispatch>
10214 int32_t messageCode,
10215 const std::string & layerPrefix,
10216 const std::string & message,
10220 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10221 VULKAN_HPP_ASSERT( d.vkDebugReportMessageEXT &&
"Function <vkDebugReportMessageEXT> requires <VK_EXT_debug_report>" );
10224 d.vkDebugReportMessageEXT( m_instance,
10230 layerPrefix.c_str(),
10237 template <
typename Dispatch>
10245 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10246 template <
typename Dispatch>
10248 Device::debugMarkerSetObjectTagEXT(
const VULKAN_HPP_NAMESPACE::DebugMarkerObjectTagInfoEXT & tagInfo, Dispatch
const & d )
const
10251 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10252 VULKAN_HPP_ASSERT( d.vkDebugMarkerSetObjectTagEXT &&
"Function <vkDebugMarkerSetObjectTagEXT> requires <VK_EXT_debug_marker>" );
10263 template <
typename Dispatch>
10271 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10272 template <
typename Dispatch>
10274 Device::debugMarkerSetObjectNameEXT(
const VULKAN_HPP_NAMESPACE::DebugMarkerObjectNameInfoEXT & nameInfo, Dispatch
const & d )
const
10277 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10278 VULKAN_HPP_ASSERT( d.vkDebugMarkerSetObjectNameEXT &&
"Function <vkDebugMarkerSetObjectNameEXT> requires <VK_EXT_debug_marker>" );
10289 template <
typename Dispatch>
10290 VULKAN_HPP_INLINE void CommandBuffer::debugMarkerBeginEXT(
const VULKAN_HPP_NAMESPACE::DebugMarkerMarkerInfoEXT * pMarkerInfo,
10297 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10298 template <
typename Dispatch>
10299 VULKAN_HPP_INLINE void CommandBuffer::debugMarkerBeginEXT(
const VULKAN_HPP_NAMESPACE::DebugMarkerMarkerInfoEXT & markerInfo,
10303 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10304 VULKAN_HPP_ASSERT( d.vkCmdDebugMarkerBeginEXT &&
"Function <vkCmdDebugMarkerBeginEXT> requires <VK_EXT_debug_marker>" );
10311 template <
typename Dispatch>
10315 d.vkCmdDebugMarkerEndEXT( m_commandBuffer );
10318 template <
typename Dispatch>
10319 VULKAN_HPP_INLINE void CommandBuffer::debugMarkerInsertEXT(
const VULKAN_HPP_NAMESPACE::DebugMarkerMarkerInfoEXT * pMarkerInfo,
10326 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10327 template <
typename Dispatch>
10328 VULKAN_HPP_INLINE void CommandBuffer::debugMarkerInsertEXT(
const VULKAN_HPP_NAMESPACE::DebugMarkerMarkerInfoEXT & markerInfo,
10332 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10333 VULKAN_HPP_ASSERT( d.vkCmdDebugMarkerInsertEXT &&
"Function <vkCmdDebugMarkerInsertEXT> requires <VK_EXT_debug_marker>" );
10342 template <
typename Dispatch>
10344 VULKAN_HPP_NAMESPACE::VideoCapabilitiesKHR * pCapabilities,
10348 return static_cast<Result>( d.vkGetPhysicalDeviceVideoCapabilitiesKHR(
10352 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10353 template <
typename Dispatch>
10355 PhysicalDevice::getVideoCapabilitiesKHR(
const VULKAN_HPP_NAMESPACE::VideoProfileInfoKHR & videoProfile, Dispatch
const & d )
const
10358 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10359 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceVideoCapabilitiesKHR &&
"Function <vkGetPhysicalDeviceVideoCapabilitiesKHR> requires <VK_KHR_video_queue>" );
10362 VULKAN_HPP_NAMESPACE::VideoCapabilitiesKHR capabilities;
10370 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
10372 PhysicalDevice::getVideoCapabilitiesKHR(
const VULKAN_HPP_NAMESPACE::VideoProfileInfoKHR & videoProfile, Dispatch
const & d )
const
10375 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10376 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceVideoCapabilitiesKHR &&
"Function <vkGetPhysicalDeviceVideoCapabilitiesKHR> requires <VK_KHR_video_queue>" );
10380 VULKAN_HPP_NAMESPACE::VideoCapabilitiesKHR & capabilities = structureChain.template get<VULKAN_HPP_NAMESPACE::VideoCapabilitiesKHR>();
10389 template <
typename Dispatch>
10391 PhysicalDevice::getVideoFormatPropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceVideoFormatInfoKHR * pVideoFormatInfo,
10392 uint32_t * pVideoFormatPropertyCount,
10393 VULKAN_HPP_NAMESPACE::VideoFormatPropertiesKHR * pVideoFormatProperties,
10397 return static_cast<Result>( d.vkGetPhysicalDeviceVideoFormatPropertiesKHR( m_physicalDevice,
10399 pVideoFormatPropertyCount,
10403 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10404 template <
typename V
ideoFormatPropertiesKHRAllocator,
typename Dispatch>
10407 PhysicalDevice::getVideoFormatPropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceVideoFormatInfoKHR & videoFormatInfo, Dispatch
const & d )
const
10410 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10412 "Function <vkGetPhysicalDeviceVideoFormatPropertiesKHR> requires <VK_KHR_video_queue>" );
10415 std::vector<VULKAN_HPP_NAMESPACE::VideoFormatPropertiesKHR, VideoFormatPropertiesKHRAllocator> videoFormatProperties;
10416 uint32_t videoFormatPropertyCount;
10422 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && videoFormatPropertyCount )
10424 videoFormatProperties.resize( videoFormatPropertyCount );
10426 d.vkGetPhysicalDeviceVideoFormatPropertiesKHR( m_physicalDevice,
10428 &videoFormatPropertyCount,
10431 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
10434 if ( videoFormatPropertyCount < videoFormatProperties.size() )
10436 videoFormatProperties.resize( videoFormatPropertyCount );
10441 template <
typename VideoFormatPropertiesKHRAllocator,
10443 typename std::enable_if<std::is_same<typename VideoFormatPropertiesKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::VideoFormatPropertiesKHR>::value,
10447 PhysicalDevice::getVideoFormatPropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceVideoFormatInfoKHR & videoFormatInfo,
10448 VideoFormatPropertiesKHRAllocator & videoFormatPropertiesKHRAllocator,
10449 Dispatch
const & d )
const
10452 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10454 "Function <vkGetPhysicalDeviceVideoFormatPropertiesKHR> requires <VK_KHR_video_queue>" );
10457 std::vector<VULKAN_HPP_NAMESPACE::VideoFormatPropertiesKHR, VideoFormatPropertiesKHRAllocator> videoFormatProperties( videoFormatPropertiesKHRAllocator );
10458 uint32_t videoFormatPropertyCount;
10464 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && videoFormatPropertyCount )
10466 videoFormatProperties.resize( videoFormatPropertyCount );
10468 d.vkGetPhysicalDeviceVideoFormatPropertiesKHR( m_physicalDevice,
10470 &videoFormatPropertyCount,
10473 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
10476 if ( videoFormatPropertyCount < videoFormatProperties.size() )
10478 videoFormatProperties.resize( videoFormatPropertyCount );
10484 template <
typename Dispatch>
10486 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
10487 VULKAN_HPP_NAMESPACE::VideoSessionKHR * pVideoSession,
10491 return static_cast<Result>( d.vkCreateVideoSessionKHR( m_device,
10497 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10498 template <
typename Dispatch>
10500 Device::createVideoSessionKHR(
const VULKAN_HPP_NAMESPACE::VideoSessionCreateInfoKHR & createInfo,
10502 Dispatch
const & d )
const
10505 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10506 VULKAN_HPP_ASSERT( d.vkCreateVideoSessionKHR &&
"Function <vkCreateVideoSessionKHR> requires <VK_KHR_video_queue>" );
10509 VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession;
10511 d.vkCreateVideoSessionKHR( m_device,
10513 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
10520 # ifndef VULKAN_HPP_NO_SMART_HANDLE
10521 template <
typename Dispatch>
10523 Device::createVideoSessionKHRUnique(
const VULKAN_HPP_NAMESPACE::VideoSessionCreateInfoKHR & createInfo,
10525 Dispatch
const & d )
const
10528 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10529 VULKAN_HPP_ASSERT( d.vkCreateVideoSessionKHR &&
"Function <vkCreateVideoSessionKHR> requires <VK_KHR_video_queue>" );
10532 VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession;
10534 d.vkCreateVideoSessionKHR( m_device,
10536 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
10546 template <
typename Dispatch>
10547 VULKAN_HPP_INLINE void Device::destroyVideoSessionKHR( VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession,
10548 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
10555 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10556 template <
typename Dispatch>
10557 VULKAN_HPP_INLINE void Device::destroyVideoSessionKHR( VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession,
10562 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10563 VULKAN_HPP_ASSERT( d.vkDestroyVideoSessionKHR &&
"Function <vkDestroyVideoSessionKHR> requires <VK_KHR_video_queue>" );
10566 d.vkDestroyVideoSessionKHR(
10569 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
10573 template <
typename Dispatch>
10575 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
10582 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10583 template <
typename Dispatch>
10589 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10590 VULKAN_HPP_ASSERT( d.vkDestroyVideoSessionKHR &&
"Function <vkDestroyVideoSessionKHR> requires <VK_KHR_video_queue>" );
10593 d.vkDestroyVideoSessionKHR(
10596 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
10600 template <
typename Dispatch>
10602 Device::getVideoSessionMemoryRequirementsKHR( VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession,
10603 uint32_t * pMemoryRequirementsCount,
10604 VULKAN_HPP_NAMESPACE::VideoSessionMemoryRequirementsKHR * pMemoryRequirements,
10608 return static_cast<Result>( d.vkGetVideoSessionMemoryRequirementsKHR( m_device,
10610 pMemoryRequirementsCount,
10614 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10615 template <
typename V
ideoSessionMemoryRequirementsKHRAllocator,
typename Dispatch>
10618 Device::getVideoSessionMemoryRequirementsKHR( VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, Dispatch
const & d )
const
10621 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10622 VULKAN_HPP_ASSERT( d.vkGetVideoSessionMemoryRequirementsKHR &&
"Function <vkGetVideoSessionMemoryRequirementsKHR> requires <VK_KHR_video_queue>" );
10625 std::vector<VULKAN_HPP_NAMESPACE::VideoSessionMemoryRequirementsKHR, VideoSessionMemoryRequirementsKHRAllocator> memoryRequirements;
10626 uint32_t memoryRequirementsCount;
10631 d.vkGetVideoSessionMemoryRequirementsKHR( m_device,
static_cast<VkVideoSessionKHR>( videoSession ), &memoryRequirementsCount,
nullptr ) );
10632 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && memoryRequirementsCount )
10634 memoryRequirements.resize( memoryRequirementsCount );
10636 d.vkGetVideoSessionMemoryRequirementsKHR( m_device,
10638 &memoryRequirementsCount,
10641 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
10644 if ( memoryRequirementsCount < memoryRequirements.size() )
10646 memoryRequirements.resize( memoryRequirementsCount );
10648 return memoryRequirements;
10651 template <
typename VideoSessionMemoryRequirementsKHRAllocator,
10653 typename std::enable_if<
10654 std::is_same<typename VideoSessionMemoryRequirementsKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::VideoSessionMemoryRequirementsKHR>::value,
10658 Device::getVideoSessionMemoryRequirementsKHR( VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession,
10659 VideoSessionMemoryRequirementsKHRAllocator & videoSessionMemoryRequirementsKHRAllocator,
10660 Dispatch
const & d )
const
10663 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10664 VULKAN_HPP_ASSERT( d.vkGetVideoSessionMemoryRequirementsKHR &&
"Function <vkGetVideoSessionMemoryRequirementsKHR> requires <VK_KHR_video_queue>" );
10667 std::vector<VULKAN_HPP_NAMESPACE::VideoSessionMemoryRequirementsKHR, VideoSessionMemoryRequirementsKHRAllocator> memoryRequirements(
10668 videoSessionMemoryRequirementsKHRAllocator );
10669 uint32_t memoryRequirementsCount;
10674 d.vkGetVideoSessionMemoryRequirementsKHR( m_device,
static_cast<VkVideoSessionKHR>( videoSession ), &memoryRequirementsCount,
nullptr ) );
10675 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && memoryRequirementsCount )
10677 memoryRequirements.resize( memoryRequirementsCount );
10679 d.vkGetVideoSessionMemoryRequirementsKHR( m_device,
10681 &memoryRequirementsCount,
10684 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
10687 if ( memoryRequirementsCount < memoryRequirements.size() )
10689 memoryRequirements.resize( memoryRequirementsCount );
10691 return memoryRequirements;
10695 template <
typename Dispatch>
10697 Device::bindVideoSessionMemoryKHR( VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession,
10698 uint32_t bindSessionMemoryInfoCount,
10699 const VULKAN_HPP_NAMESPACE::BindVideoSessionMemoryInfoKHR * pBindSessionMemoryInfos,
10703 return static_cast<Result>( d.vkBindVideoSessionMemoryKHR( m_device,
10705 bindSessionMemoryInfoCount,
10709 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10710 template <
typename Dispatch>
10712 VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession,
10713 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BindVideoSessionMemoryInfoKHR>
const & bindSessionMemoryInfos,
10714 Dispatch
const & d )
const
10717 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10718 VULKAN_HPP_ASSERT( d.vkBindVideoSessionMemoryKHR &&
"Function <vkBindVideoSessionMemoryKHR> requires <VK_KHR_video_queue>" );
10722 d.vkBindVideoSessionMemoryKHR( m_device,
10724 bindSessionMemoryInfos.size(),
10732 template <
typename Dispatch>
10734 Device::createVideoSessionParametersKHR(
const VULKAN_HPP_NAMESPACE::VideoSessionParametersCreateInfoKHR * pCreateInfo,
10735 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
10736 VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR * pVideoSessionParameters,
10740 return static_cast<Result>( d.vkCreateVideoSessionParametersKHR( m_device,
10746 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10747 template <
typename Dispatch>
10749 Device::createVideoSessionParametersKHR(
const VULKAN_HPP_NAMESPACE::VideoSessionParametersCreateInfoKHR & createInfo,
10751 Dispatch
const & d )
const
10754 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10755 VULKAN_HPP_ASSERT( d.vkCreateVideoSessionParametersKHR &&
"Function <vkCreateVideoSessionParametersKHR> requires <VK_KHR_video_queue>" );
10758 VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters;
10762 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
10769 # ifndef VULKAN_HPP_NO_SMART_HANDLE
10770 template <
typename Dispatch>
10772 Device::createVideoSessionParametersKHRUnique(
const VULKAN_HPP_NAMESPACE::VideoSessionParametersCreateInfoKHR & createInfo,
10774 Dispatch
const & d )
const
10777 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10778 VULKAN_HPP_ASSERT( d.vkCreateVideoSessionParametersKHR &&
"Function <vkCreateVideoSessionParametersKHR> requires <VK_KHR_video_queue>" );
10781 VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters;
10785 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
10796 template <
typename Dispatch>
10798 Device::updateVideoSessionParametersKHR( VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters,
10799 const VULKAN_HPP_NAMESPACE::VideoSessionParametersUpdateInfoKHR * pUpdateInfo,
10803 return static_cast<Result>( d.vkUpdateVideoSessionParametersKHR( m_device,
10808 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10809 template <
typename Dispatch>
10811 Device::updateVideoSessionParametersKHR( VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters,
10812 const VULKAN_HPP_NAMESPACE::VideoSessionParametersUpdateInfoKHR & updateInfo,
10813 Dispatch
const & d )
const
10816 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10817 VULKAN_HPP_ASSERT( d.vkUpdateVideoSessionParametersKHR &&
"Function <vkUpdateVideoSessionParametersKHR> requires <VK_KHR_video_queue>" );
10821 d.vkUpdateVideoSessionParametersKHR( m_device,
10830 template <
typename Dispatch>
10831 VULKAN_HPP_INLINE void Device::destroyVideoSessionParametersKHR( VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters,
10832 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
10836 d.vkDestroyVideoSessionParametersKHR(
10840 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10841 template <
typename Dispatch>
10842 VULKAN_HPP_INLINE void Device::destroyVideoSessionParametersKHR( VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters,
10847 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10848 VULKAN_HPP_ASSERT( d.vkDestroyVideoSessionParametersKHR &&
"Function <vkDestroyVideoSessionParametersKHR> requires <VK_KHR_video_queue>" );
10851 d.vkDestroyVideoSessionParametersKHR(
10854 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
10858 template <
typename Dispatch>
10859 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters,
10860 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
10864 d.vkDestroyVideoSessionParametersKHR(
10868 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10869 template <
typename Dispatch>
10870 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters,
10875 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10876 VULKAN_HPP_ASSERT( d.vkDestroyVideoSessionParametersKHR &&
"Function <vkDestroyVideoSessionParametersKHR> requires <VK_KHR_video_queue>" );
10879 d.vkDestroyVideoSessionParametersKHR(
10882 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
10886 template <
typename Dispatch>
10887 VULKAN_HPP_INLINE void CommandBuffer::beginVideoCodingKHR(
const VULKAN_HPP_NAMESPACE::VideoBeginCodingInfoKHR * pBeginInfo,
10894 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10895 template <
typename Dispatch>
10896 VULKAN_HPP_INLINE void CommandBuffer::beginVideoCodingKHR(
const VULKAN_HPP_NAMESPACE::VideoBeginCodingInfoKHR & beginInfo,
10900 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10901 VULKAN_HPP_ASSERT( d.vkCmdBeginVideoCodingKHR &&
"Function <vkCmdBeginVideoCodingKHR> requires <VK_KHR_video_queue>" );
10908 template <
typename Dispatch>
10909 VULKAN_HPP_INLINE void CommandBuffer::endVideoCodingKHR(
const VULKAN_HPP_NAMESPACE::VideoEndCodingInfoKHR * pEndCodingInfo,
10913 d.vkCmdEndVideoCodingKHR( m_commandBuffer,
reinterpret_cast<const VkVideoEndCodingInfoKHR *
>( pEndCodingInfo ) );
10916 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10917 template <
typename Dispatch>
10918 VULKAN_HPP_INLINE void CommandBuffer::endVideoCodingKHR(
const VULKAN_HPP_NAMESPACE::VideoEndCodingInfoKHR & endCodingInfo,
10922 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10923 VULKAN_HPP_ASSERT( d.vkCmdEndVideoCodingKHR &&
"Function <vkCmdEndVideoCodingKHR> requires <VK_KHR_video_queue>" );
10926 d.vkCmdEndVideoCodingKHR( m_commandBuffer,
reinterpret_cast<const VkVideoEndCodingInfoKHR *
>( &endCodingInfo ) );
10930 template <
typename Dispatch>
10931 VULKAN_HPP_INLINE void CommandBuffer::controlVideoCodingKHR(
const VULKAN_HPP_NAMESPACE::VideoCodingControlInfoKHR * pCodingControlInfo,
10938 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10939 template <
typename Dispatch>
10940 VULKAN_HPP_INLINE void CommandBuffer::controlVideoCodingKHR(
const VULKAN_HPP_NAMESPACE::VideoCodingControlInfoKHR & codingControlInfo,
10944 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10945 VULKAN_HPP_ASSERT( d.vkCmdControlVideoCodingKHR &&
"Function <vkCmdControlVideoCodingKHR> requires <VK_KHR_video_queue>" );
10954 template <
typename Dispatch>
10955 VULKAN_HPP_INLINE void CommandBuffer::decodeVideoKHR(
const VULKAN_HPP_NAMESPACE::VideoDecodeInfoKHR * pDecodeInfo,
10959 d.vkCmdDecodeVideoKHR( m_commandBuffer,
reinterpret_cast<const VkVideoDecodeInfoKHR *
>( pDecodeInfo ) );
10962 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10963 template <
typename Dispatch>
10964 VULKAN_HPP_INLINE void CommandBuffer::decodeVideoKHR(
const VULKAN_HPP_NAMESPACE::VideoDecodeInfoKHR & decodeInfo,
10968 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
10969 VULKAN_HPP_ASSERT( d.vkCmdDecodeVideoKHR &&
"Function <vkCmdDecodeVideoKHR> requires <VK_KHR_video_decode_queue>" );
10972 d.vkCmdDecodeVideoKHR( m_commandBuffer,
reinterpret_cast<const VkVideoDecodeInfoKHR *
>( &decodeInfo ) );
10978 template <
typename Dispatch>
10980 uint32_t bindingCount,
10981 const VULKAN_HPP_NAMESPACE::Buffer * pBuffers,
10987 d.vkCmdBindTransformFeedbackBuffersEXT( m_commandBuffer,
10990 reinterpret_cast<const VkBuffer *
>( pBuffers ),
10995 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
10996 template <
typename Dispatch>
10998 CommandBuffer::bindTransformFeedbackBuffersEXT( uint32_t firstBinding,
10999 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer>
const & buffers,
11000 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & offsets,
11001 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & sizes,
11005 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11006 VULKAN_HPP_ASSERT( d.vkCmdBindTransformFeedbackBuffersEXT &&
"Function <vkCmdBindTransformFeedbackBuffersEXT> requires <VK_EXT_transform_feedback>" );
11008 # ifdef VULKAN_HPP_NO_EXCEPTIONS
11012 if ( buffers.size() != offsets.size() )
11016 if ( !sizes.empty() && buffers.size() != sizes.size() )
11022 d.vkCmdBindTransformFeedbackBuffersEXT( m_commandBuffer,
11025 reinterpret_cast<const VkBuffer *
>( buffers.data() ),
11026 reinterpret_cast<const VkDeviceSize *
>( offsets.data() ),
11027 reinterpret_cast<const VkDeviceSize *
>( sizes.data() ) );
11031 template <
typename Dispatch>
11033 uint32_t counterBufferCount,
11034 const VULKAN_HPP_NAMESPACE::Buffer * pCounterBuffers,
11039 d.vkCmdBeginTransformFeedbackEXT( m_commandBuffer,
11040 firstCounterBuffer,
11041 counterBufferCount,
11042 reinterpret_cast<const VkBuffer *
>( pCounterBuffers ),
11043 reinterpret_cast<const VkDeviceSize *
>( pCounterBufferOffsets ) );
11046 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11047 template <
typename Dispatch>
11049 CommandBuffer::beginTransformFeedbackEXT( uint32_t firstCounterBuffer,
11050 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer>
const & counterBuffers,
11051 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & counterBufferOffsets,
11055 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11056 VULKAN_HPP_ASSERT( d.vkCmdBeginTransformFeedbackEXT &&
"Function <vkCmdBeginTransformFeedbackEXT> requires <VK_EXT_transform_feedback>" );
11058 # ifdef VULKAN_HPP_NO_EXCEPTIONS
11059 VULKAN_HPP_ASSERT( counterBufferOffsets.empty() || counterBuffers.size() == counterBufferOffsets.size() );
11061 if ( !counterBufferOffsets.empty() && counterBuffers.size() != counterBufferOffsets.size() )
11067 d.vkCmdBeginTransformFeedbackEXT( m_commandBuffer,
11068 firstCounterBuffer,
11069 counterBuffers.size(),
11070 reinterpret_cast<const VkBuffer *
>( counterBuffers.data() ),
11071 reinterpret_cast<const VkDeviceSize *
>( counterBufferOffsets.data() ) );
11075 template <
typename Dispatch>
11077 uint32_t counterBufferCount,
11078 const VULKAN_HPP_NAMESPACE::Buffer * pCounterBuffers,
11083 d.vkCmdEndTransformFeedbackEXT( m_commandBuffer,
11084 firstCounterBuffer,
11085 counterBufferCount,
11086 reinterpret_cast<const VkBuffer *
>( pCounterBuffers ),
11087 reinterpret_cast<const VkDeviceSize *
>( pCounterBufferOffsets ) );
11090 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11091 template <
typename Dispatch>
11093 CommandBuffer::endTransformFeedbackEXT( uint32_t firstCounterBuffer,
11094 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer>
const & counterBuffers,
11095 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & counterBufferOffsets,
11099 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11100 VULKAN_HPP_ASSERT( d.vkCmdEndTransformFeedbackEXT &&
"Function <vkCmdEndTransformFeedbackEXT> requires <VK_EXT_transform_feedback>" );
11102 # ifdef VULKAN_HPP_NO_EXCEPTIONS
11103 VULKAN_HPP_ASSERT( counterBufferOffsets.empty() || counterBuffers.size() == counterBufferOffsets.size() );
11105 if ( !counterBufferOffsets.empty() && counterBuffers.size() != counterBufferOffsets.size() )
11111 d.vkCmdEndTransformFeedbackEXT( m_commandBuffer,
11112 firstCounterBuffer,
11113 counterBuffers.size(),
11114 reinterpret_cast<const VkBuffer *
>( counterBuffers.data() ),
11115 reinterpret_cast<const VkDeviceSize *
>( counterBufferOffsets.data() ) );
11119 template <
typename Dispatch>
11130 template <
typename Dispatch>
11132 CommandBuffer::endQueryIndexedEXT( VULKAN_HPP_NAMESPACE::QueryPool queryPool, uint32_t query, uint32_t index, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
11135 d.vkCmdEndQueryIndexedEXT( m_commandBuffer,
static_cast<VkQueryPool>( queryPool ), query, index );
11138 template <
typename Dispatch>
11140 uint32_t firstInstance,
11141 VULKAN_HPP_NAMESPACE::Buffer counterBuffer,
11143 uint32_t counterOffset,
11144 uint32_t vertexStride,
11148 d.vkCmdDrawIndirectByteCountEXT( m_commandBuffer,
11151 static_cast<VkBuffer>( counterBuffer ),
11159 template <
typename Dispatch>
11161 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
11162 VULKAN_HPP_NAMESPACE::CuModuleNVX * pModule,
11166 return static_cast<Result>( d.vkCreateCuModuleNVX( m_device,
11172 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11173 template <
typename Dispatch>
11175 Device::createCuModuleNVX(
const VULKAN_HPP_NAMESPACE::CuModuleCreateInfoNVX & createInfo,
11177 Dispatch
const & d )
const
11180 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11181 VULKAN_HPP_ASSERT( d.vkCreateCuModuleNVX &&
"Function <vkCreateCuModuleNVX> requires <VK_NVX_binary_import>" );
11184 VULKAN_HPP_NAMESPACE::CuModuleNVX
module;
11186 d.vkCreateCuModuleNVX( m_device,
11188 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
11195 # ifndef VULKAN_HPP_NO_SMART_HANDLE
11196 template <
typename Dispatch>
11198 Device::createCuModuleNVXUnique(
const VULKAN_HPP_NAMESPACE::CuModuleCreateInfoNVX & createInfo,
11200 Dispatch
const & d )
const
11203 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11204 VULKAN_HPP_ASSERT( d.vkCreateCuModuleNVX &&
"Function <vkCreateCuModuleNVX> requires <VK_NVX_binary_import>" );
11207 VULKAN_HPP_NAMESPACE::CuModuleNVX
module;
11209 d.vkCreateCuModuleNVX( m_device,
11211 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
11221 template <
typename Dispatch>
11223 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
11224 VULKAN_HPP_NAMESPACE::CuFunctionNVX * pFunction,
11228 return static_cast<Result>( d.vkCreateCuFunctionNVX( m_device,
11234 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11235 template <
typename Dispatch>
11237 Device::createCuFunctionNVX(
const VULKAN_HPP_NAMESPACE::CuFunctionCreateInfoNVX & createInfo,
11239 Dispatch
const & d )
const
11242 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11243 VULKAN_HPP_ASSERT( d.vkCreateCuFunctionNVX &&
"Function <vkCreateCuFunctionNVX> requires <VK_NVX_binary_import>" );
11246 VULKAN_HPP_NAMESPACE::CuFunctionNVX
function;
11248 d.vkCreateCuFunctionNVX( m_device,
11250 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
11257 # ifndef VULKAN_HPP_NO_SMART_HANDLE
11258 template <
typename Dispatch>
11260 Device::createCuFunctionNVXUnique(
const VULKAN_HPP_NAMESPACE::CuFunctionCreateInfoNVX & createInfo,
11262 Dispatch
const & d )
const
11265 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11266 VULKAN_HPP_ASSERT( d.vkCreateCuFunctionNVX &&
"Function <vkCreateCuFunctionNVX> requires <VK_NVX_binary_import>" );
11269 VULKAN_HPP_NAMESPACE::CuFunctionNVX
function;
11271 d.vkCreateCuFunctionNVX( m_device,
11273 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
11283 template <
typename Dispatch>
11285 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
11292 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11293 template <
typename Dispatch>
11299 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11300 VULKAN_HPP_ASSERT( d.vkDestroyCuModuleNVX &&
"Function <vkDestroyCuModuleNVX> requires <VK_NVX_binary_import>" );
11303 d.vkDestroyCuModuleNVX( m_device,
11305 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
11309 template <
typename Dispatch>
11311 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
11318 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11319 template <
typename Dispatch>
11325 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11326 VULKAN_HPP_ASSERT( d.vkDestroyCuModuleNVX &&
"Function <vkDestroyCuModuleNVX> requires <VK_NVX_binary_import>" );
11329 d.vkDestroyCuModuleNVX( m_device,
11331 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
11335 template <
typename Dispatch>
11337 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
11344 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11345 template <
typename Dispatch>
11351 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11352 VULKAN_HPP_ASSERT( d.vkDestroyCuFunctionNVX &&
"Function <vkDestroyCuFunctionNVX> requires <VK_NVX_binary_import>" );
11355 d.vkDestroyCuFunctionNVX( m_device,
11357 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
11361 template <
typename Dispatch>
11363 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
11370 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11371 template <
typename Dispatch>
11377 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11378 VULKAN_HPP_ASSERT( d.vkDestroyCuFunctionNVX &&
"Function <vkDestroyCuFunctionNVX> requires <VK_NVX_binary_import>" );
11381 d.vkDestroyCuFunctionNVX( m_device,
11383 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
11387 template <
typename Dispatch>
11388 VULKAN_HPP_INLINE void CommandBuffer::cuLaunchKernelNVX(
const VULKAN_HPP_NAMESPACE::CuLaunchInfoNVX * pLaunchInfo,
11392 d.vkCmdCuLaunchKernelNVX( m_commandBuffer,
reinterpret_cast<const VkCuLaunchInfoNVX *
>( pLaunchInfo ) );
11395 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11396 template <
typename Dispatch>
11397 VULKAN_HPP_INLINE void CommandBuffer::cuLaunchKernelNVX(
const VULKAN_HPP_NAMESPACE::CuLaunchInfoNVX & launchInfo,
11401 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11402 VULKAN_HPP_ASSERT( d.vkCmdCuLaunchKernelNVX &&
"Function <vkCmdCuLaunchKernelNVX> requires <VK_NVX_binary_import>" );
11405 d.vkCmdCuLaunchKernelNVX( m_commandBuffer,
reinterpret_cast<const VkCuLaunchInfoNVX *
>( &launchInfo ) );
11411 template <
typename Dispatch>
11412 VULKAN_HPP_INLINE uint32_t Device::getImageViewHandleNVX(
const VULKAN_HPP_NAMESPACE::ImageViewHandleInfoNVX * pInfo,
11419 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11420 template <
typename Dispatch>
11421 VULKAN_HPP_INLINE uint32_t Device::getImageViewHandleNVX(
const VULKAN_HPP_NAMESPACE::ImageViewHandleInfoNVX & info,
11425 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11426 VULKAN_HPP_ASSERT( d.vkGetImageViewHandleNVX &&
"Function <vkGetImageViewHandleNVX> requires <VK_NVX_image_view_handle>" );
11429 uint32_t result = d.vkGetImageViewHandleNVX( m_device,
reinterpret_cast<const VkImageViewHandleInfoNVX *
>( &info ) );
11435 template <
typename Dispatch>
11437 VULKAN_HPP_NAMESPACE::ImageViewAddressPropertiesNVX * pProperties,
11441 return static_cast<Result>(
11445 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11446 template <
typename Dispatch>
11448 Device::getImageViewAddressNVX( VULKAN_HPP_NAMESPACE::ImageView imageView, Dispatch
const & d )
const
11451 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11452 VULKAN_HPP_ASSERT( d.vkGetImageViewAddressNVX &&
"Function <vkGetImageViewAddressNVX> requires <VK_NVX_image_view_handle>" );
11455 VULKAN_HPP_NAMESPACE::ImageViewAddressPropertiesNVX properties;
11466 template <
typename Dispatch>
11469 VULKAN_HPP_NAMESPACE::Buffer countBuffer,
11471 uint32_t maxDrawCount,
11476 d.vkCmdDrawIndirectCountAMD( m_commandBuffer,
11479 static_cast<VkBuffer>( countBuffer ),
11485 template <
typename Dispatch>
11488 VULKAN_HPP_NAMESPACE::Buffer countBuffer,
11490 uint32_t maxDrawCount,
11495 d.vkCmdDrawIndexedIndirectCountAMD( m_commandBuffer,
11498 static_cast<VkBuffer>( countBuffer ),
11506 template <
typename Dispatch>
11510 size_t * pInfoSize,
11515 return static_cast<Result>( d.vkGetShaderInfoAMD( m_device,
11523 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11524 template <
typename U
int8_tAllocator,
typename Dispatch>
11526 Device::getShaderInfoAMD( VULKAN_HPP_NAMESPACE::Pipeline pipeline,
11529 Dispatch
const & d )
const
11532 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11533 VULKAN_HPP_ASSERT( d.vkGetShaderInfoAMD &&
"Function <vkGetShaderInfoAMD> requires <VK_AMD_shader_info>" );
11536 std::vector<uint8_t, Uint8_tAllocator> info;
11547 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && infoSize )
11549 info.resize( infoSize );
11555 reinterpret_cast<void *
>( info.data() ) ) );
11557 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
11560 if ( infoSize < info.size() )
11562 info.resize( infoSize );
11567 template <
typename Uint8_tAllocator,
11569 typename std::enable_if<std::is_same<typename Uint8_tAllocator::value_type, uint8_t>::value,
int>::type>
11571 Device::getShaderInfoAMD( VULKAN_HPP_NAMESPACE::Pipeline pipeline,
11574 Uint8_tAllocator & uint8_tAllocator,
11575 Dispatch
const & d )
const
11578 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11579 VULKAN_HPP_ASSERT( d.vkGetShaderInfoAMD &&
"Function <vkGetShaderInfoAMD> requires <VK_AMD_shader_info>" );
11582 std::vector<uint8_t, Uint8_tAllocator> info( uint8_tAllocator );
11593 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && infoSize )
11595 info.resize( infoSize );
11601 reinterpret_cast<void *
>( info.data() ) ) );
11603 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
11606 if ( infoSize < info.size() )
11608 info.resize( infoSize );
11616 template <
typename Dispatch>
11617 VULKAN_HPP_INLINE void CommandBuffer::beginRenderingKHR(
const VULKAN_HPP_NAMESPACE::RenderingInfo * pRenderingInfo,
11621 d.vkCmdBeginRenderingKHR( m_commandBuffer,
reinterpret_cast<const VkRenderingInfo *
>( pRenderingInfo ) );
11624 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11625 template <
typename Dispatch>
11626 VULKAN_HPP_INLINE void CommandBuffer::beginRenderingKHR(
const VULKAN_HPP_NAMESPACE::RenderingInfo & renderingInfo,
11630 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11631 VULKAN_HPP_ASSERT( d.vkCmdBeginRenderingKHR &&
"Function <vkCmdBeginRenderingKHR> requires <VK_KHR_dynamic_rendering> or <VK_VERSION_1_3>" );
11634 d.vkCmdBeginRenderingKHR( m_commandBuffer,
reinterpret_cast<const VkRenderingInfo *
>( &renderingInfo ) );
11638 template <
typename Dispatch>
11642 d.vkCmdEndRenderingKHR( m_commandBuffer );
11645 #if defined( VK_USE_PLATFORM_GGP )
11648 template <
typename Dispatch>
11650 Instance::createStreamDescriptorSurfaceGGP(
const VULKAN_HPP_NAMESPACE::StreamDescriptorSurfaceCreateInfoGGP * pCreateInfo,
11651 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
11652 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
11656 return static_cast<Result>( d.vkCreateStreamDescriptorSurfaceGGP( m_instance,
11657 reinterpret_cast<const VkStreamDescriptorSurfaceCreateInfoGGP *
>( pCreateInfo ),
11662 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11663 template <
typename Dispatch>
11665 Instance::createStreamDescriptorSurfaceGGP(
const VULKAN_HPP_NAMESPACE::StreamDescriptorSurfaceCreateInfoGGP & createInfo,
11666 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
11667 Dispatch
const & d )
const
11670 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11671 VULKAN_HPP_ASSERT( d.vkCreateStreamDescriptorSurfaceGGP &&
"Function <vkCreateStreamDescriptorSurfaceGGP> requires <VK_GGP_stream_descriptor_surface>" );
11674 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
11677 reinterpret_cast<const VkStreamDescriptorSurfaceCreateInfoGGP *
>( &createInfo ),
11678 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
11685 # ifndef VULKAN_HPP_NO_SMART_HANDLE
11686 template <
typename Dispatch>
11688 Instance::createStreamDescriptorSurfaceGGPUnique(
const VULKAN_HPP_NAMESPACE::StreamDescriptorSurfaceCreateInfoGGP & createInfo,
11689 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
11690 Dispatch
const & d )
const
11693 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11694 VULKAN_HPP_ASSERT( d.vkCreateStreamDescriptorSurfaceGGP &&
"Function <vkCreateStreamDescriptorSurfaceGGP> requires <VK_GGP_stream_descriptor_surface>" );
11697 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
11700 reinterpret_cast<const VkStreamDescriptorSurfaceCreateInfoGGP *
>( &createInfo ),
11701 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
11706 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
11714 template <
typename Dispatch>
11722 VULKAN_HPP_NAMESPACE::ExternalImageFormatPropertiesNV * pExternalImageFormatProperties,
11726 return static_cast<Result>(
11727 d.vkGetPhysicalDeviceExternalImageFormatPropertiesNV( m_physicalDevice,
11737 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11738 template <
typename Dispatch>
11746 Dispatch
const & d )
const
11749 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11751 "Function <vkGetPhysicalDeviceExternalImageFormatPropertiesNV> requires <VK_NV_external_memory_capabilities>" );
11754 VULKAN_HPP_NAMESPACE::ExternalImageFormatPropertiesNV externalImageFormatProperties;
11756 d.vkGetPhysicalDeviceExternalImageFormatPropertiesNV( m_physicalDevice,
11770 #if defined( VK_USE_PLATFORM_WIN32_KHR )
11773 template <
typename Dispatch>
11780 return static_cast<Result>(
11784 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11785 template <
typename Dispatch>
11790 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11791 VULKAN_HPP_ASSERT( d.vkGetMemoryWin32HandleNV &&
"Function <vkGetMemoryWin32HandleNV> requires <VK_NV_external_memory_win32>" );
11806 template <
typename Dispatch>
11807 VULKAN_HPP_INLINE void PhysicalDevice::getFeatures2KHR( VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 * pFeatures,
11814 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11815 template <
typename Dispatch>
11820 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11822 "Function <vkGetPhysicalDeviceFeatures2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
11825 VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 features;
11831 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
11836 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11838 "Function <vkGetPhysicalDeviceFeatures2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
11841 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
11842 VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 & features = structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2>();
11845 return structureChain;
11849 template <
typename Dispatch>
11850 VULKAN_HPP_INLINE void PhysicalDevice::getProperties2KHR( VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 * pProperties,
11857 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11858 template <
typename Dispatch>
11863 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11865 "Function <vkGetPhysicalDeviceProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
11868 VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 properties;
11874 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
11879 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11881 "Function <vkGetPhysicalDeviceProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
11884 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
11885 VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 & properties = structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2>();
11888 return structureChain;
11892 template <
typename Dispatch>
11894 VULKAN_HPP_NAMESPACE::FormatProperties2 * pFormatProperties,
11898 d.vkGetPhysicalDeviceFormatProperties2KHR(
11902 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11903 template <
typename Dispatch>
11908 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11910 "Function <vkGetPhysicalDeviceFormatProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
11913 VULKAN_HPP_NAMESPACE::FormatProperties2 formatProperties;
11914 d.vkGetPhysicalDeviceFormatProperties2KHR(
11917 return formatProperties;
11920 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
11925 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11927 "Function <vkGetPhysicalDeviceFormatProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
11930 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
11931 VULKAN_HPP_NAMESPACE::FormatProperties2 & formatProperties = structureChain.template get<VULKAN_HPP_NAMESPACE::FormatProperties2>();
11932 d.vkGetPhysicalDeviceFormatProperties2KHR(
11935 return structureChain;
11939 template <
typename Dispatch>
11941 PhysicalDevice::getImageFormatProperties2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 * pImageFormatInfo,
11942 VULKAN_HPP_NAMESPACE::ImageFormatProperties2 * pImageFormatProperties,
11946 return static_cast<Result>( d.vkGetPhysicalDeviceImageFormatProperties2KHR( m_physicalDevice,
11951 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
11952 template <
typename Dispatch>
11954 PhysicalDevice::getImageFormatProperties2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 & imageFormatInfo, Dispatch
const & d )
const
11957 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11959 "Function <vkGetPhysicalDeviceImageFormatProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
11962 VULKAN_HPP_NAMESPACE::ImageFormatProperties2 imageFormatProperties;
11964 d.vkGetPhysicalDeviceImageFormatProperties2KHR( m_physicalDevice,
11972 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
11974 PhysicalDevice::getImageFormatProperties2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 & imageFormatInfo, Dispatch
const & d )
const
11977 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
11979 "Function <vkGetPhysicalDeviceImageFormatProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
11983 VULKAN_HPP_NAMESPACE::ImageFormatProperties2 & imageFormatProperties = structureChain.template get<VULKAN_HPP_NAMESPACE::ImageFormatProperties2>();
11985 d.vkGetPhysicalDeviceImageFormatProperties2KHR( m_physicalDevice,
11994 template <
typename Dispatch>
11995 VULKAN_HPP_INLINE void PhysicalDevice::getQueueFamilyProperties2KHR( uint32_t * pQueueFamilyPropertyCount,
11996 VULKAN_HPP_NAMESPACE::QueueFamilyProperties2 * pQueueFamilyProperties,
12000 d.vkGetPhysicalDeviceQueueFamilyProperties2KHR(
12001 m_physicalDevice, pQueueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( pQueueFamilyProperties ) );
12004 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12005 template <
typename QueueFamilyProperties2Allocator,
typename Dispatch>
12007 PhysicalDevice::getQueueFamilyProperties2KHR( Dispatch
const & d )
const
12010 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12012 "Function <vkGetPhysicalDeviceQueueFamilyProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
12015 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2, QueueFamilyProperties2Allocator> queueFamilyProperties;
12016 uint32_t queueFamilyPropertyCount;
12017 d.vkGetPhysicalDeviceQueueFamilyProperties2KHR( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
12018 queueFamilyProperties.resize( queueFamilyPropertyCount );
12019 d.vkGetPhysicalDeviceQueueFamilyProperties2KHR(
12020 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( queueFamilyProperties.data() ) );
12023 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
12025 queueFamilyProperties.resize( queueFamilyPropertyCount );
12027 return queueFamilyProperties;
12031 typename QueueFamilyProperties2Allocator,
12033 typename std::enable_if<std::is_same<typename QueueFamilyProperties2Allocator::value_type, VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>::value,
int>::type>
12035 PhysicalDevice::getQueueFamilyProperties2KHR( QueueFamilyProperties2Allocator & queueFamilyProperties2Allocator, Dispatch
const & d )
const
12038 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12040 "Function <vkGetPhysicalDeviceQueueFamilyProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
12043 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2, QueueFamilyProperties2Allocator> queueFamilyProperties( queueFamilyProperties2Allocator );
12044 uint32_t queueFamilyPropertyCount;
12045 d.vkGetPhysicalDeviceQueueFamilyProperties2KHR( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
12046 queueFamilyProperties.resize( queueFamilyPropertyCount );
12047 d.vkGetPhysicalDeviceQueueFamilyProperties2KHR(
12048 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( queueFamilyProperties.data() ) );
12051 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
12053 queueFamilyProperties.resize( queueFamilyPropertyCount );
12055 return queueFamilyProperties;
12058 template <
typename StructureChain,
typename StructureChainAllocator,
typename Dispatch>
12060 PhysicalDevice::getQueueFamilyProperties2KHR( Dispatch
const & d )
const
12063 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12065 "Function <vkGetPhysicalDeviceQueueFamilyProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
12068 std::vector<StructureChain, StructureChainAllocator> structureChains;
12069 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2> queueFamilyProperties;
12070 uint32_t queueFamilyPropertyCount;
12071 d.vkGetPhysicalDeviceQueueFamilyProperties2KHR( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
12072 structureChains.resize( queueFamilyPropertyCount );
12073 queueFamilyProperties.resize( queueFamilyPropertyCount );
12074 for ( uint32_t i = 0; i < queueFamilyPropertyCount; i++ )
12076 queueFamilyProperties[i].pNext = structureChains[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>().pNext;
12078 d.vkGetPhysicalDeviceQueueFamilyProperties2KHR(
12079 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( queueFamilyProperties.data() ) );
12082 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
12084 structureChains.resize( queueFamilyPropertyCount );
12086 for ( uint32_t i = 0; i < queueFamilyPropertyCount; i++ )
12088 structureChains[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>() = queueFamilyProperties[i];
12090 return structureChains;
12094 typename StructureChainAllocator,
12096 typename std::enable_if<std::is_same<typename StructureChainAllocator::value_type, StructureChain>::value,
int>::type>
12098 PhysicalDevice::getQueueFamilyProperties2KHR( StructureChainAllocator & structureChainAllocator, Dispatch
const & d )
const
12101 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12103 "Function <vkGetPhysicalDeviceQueueFamilyProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
12106 std::vector<StructureChain, StructureChainAllocator> structureChains( structureChainAllocator );
12107 std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2> queueFamilyProperties;
12108 uint32_t queueFamilyPropertyCount;
12109 d.vkGetPhysicalDeviceQueueFamilyProperties2KHR( m_physicalDevice, &queueFamilyPropertyCount,
nullptr );
12110 structureChains.resize( queueFamilyPropertyCount );
12111 queueFamilyProperties.resize( queueFamilyPropertyCount );
12112 for ( uint32_t i = 0; i < queueFamilyPropertyCount; i++ )
12114 queueFamilyProperties[i].pNext = structureChains[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>().pNext;
12116 d.vkGetPhysicalDeviceQueueFamilyProperties2KHR(
12117 m_physicalDevice, &queueFamilyPropertyCount,
reinterpret_cast<VkQueueFamilyProperties2 *
>( queueFamilyProperties.data() ) );
12120 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
12122 structureChains.resize( queueFamilyPropertyCount );
12124 for ( uint32_t i = 0; i < queueFamilyPropertyCount; i++ )
12126 structureChains[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>() = queueFamilyProperties[i];
12128 return structureChains;
12132 template <
typename Dispatch>
12133 VULKAN_HPP_INLINE void PhysicalDevice::getMemoryProperties2KHR( VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 * pMemoryProperties,
12140 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12141 template <
typename Dispatch>
12146 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12148 "Function <vkGetPhysicalDeviceMemoryProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
12151 VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 memoryProperties;
12154 return memoryProperties;
12157 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
12162 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12164 "Function <vkGetPhysicalDeviceMemoryProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
12167 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
12168 VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 & memoryProperties =
12169 structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2>();
12172 return structureChain;
12176 template <
typename Dispatch>
12177 VULKAN_HPP_INLINE void PhysicalDevice::getSparseImageFormatProperties2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 * pFormatInfo,
12178 uint32_t * pPropertyCount,
12179 VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2 * pProperties,
12183 d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR( m_physicalDevice,
12189 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12190 template <
typename SparseImageFormatProperties2Allocator,
typename Dispatch>
12192 PhysicalDevice::getSparseImageFormatProperties2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 & formatInfo,
12193 Dispatch
const & d )
const
12196 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12198 "Function <vkGetPhysicalDeviceSparseImageFormatProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
12201 std::vector<VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> properties;
12202 uint32_t propertyCount;
12203 d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR(
12205 properties.resize( propertyCount );
12206 d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR( m_physicalDevice,
12212 if ( propertyCount < properties.size() )
12214 properties.resize( propertyCount );
12220 typename SparseImageFormatProperties2Allocator,
12222 typename std::enable_if<std::is_same<typename SparseImageFormatProperties2Allocator::value_type, VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2>::value,
12225 PhysicalDevice::getSparseImageFormatProperties2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 & formatInfo,
12226 SparseImageFormatProperties2Allocator & sparseImageFormatProperties2Allocator,
12227 Dispatch
const & d )
const
12230 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12232 "Function <vkGetPhysicalDeviceSparseImageFormatProperties2KHR> requires <VK_KHR_get_physical_device_properties2> or <VK_VERSION_1_1>" );
12235 std::vector<VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> properties( sparseImageFormatProperties2Allocator );
12236 uint32_t propertyCount;
12237 d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR(
12239 properties.resize( propertyCount );
12240 d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR( m_physicalDevice,
12246 if ( propertyCount < properties.size() )
12248 properties.resize( propertyCount );
12256 template <
typename Dispatch>
12258 uint32_t localDeviceIndex,
12259 uint32_t remoteDeviceIndex,
12264 d.vkGetDeviceGroupPeerMemoryFeaturesKHR(
12265 m_device, heapIndex, localDeviceIndex, remoteDeviceIndex,
reinterpret_cast<VkPeerMemoryFeatureFlags *
>( pPeerMemoryFeatures ) );
12268 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12269 template <
typename Dispatch>
12271 uint32_t heapIndex, uint32_t localDeviceIndex, uint32_t remoteDeviceIndex, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
12274 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12276 "Function <vkGetDeviceGroupPeerMemoryFeaturesKHR> requires <VK_KHR_device_group> or <VK_VERSION_1_1>" );
12280 d.vkGetDeviceGroupPeerMemoryFeaturesKHR(
12281 m_device, heapIndex, localDeviceIndex, remoteDeviceIndex,
reinterpret_cast<VkPeerMemoryFeatureFlags *
>( &peerMemoryFeatures ) );
12283 return peerMemoryFeatures;
12287 template <
typename Dispatch>
12291 d.vkCmdSetDeviceMaskKHR( m_commandBuffer, deviceMask );
12294 template <
typename Dispatch>
12296 uint32_t baseGroupY,
12297 uint32_t baseGroupZ,
12298 uint32_t groupCountX,
12299 uint32_t groupCountY,
12300 uint32_t groupCountZ,
12304 d.vkCmdDispatchBaseKHR( m_commandBuffer, baseGroupX, baseGroupY, baseGroupZ, groupCountX, groupCountY, groupCountZ );
12307 #if defined( VK_USE_PLATFORM_VI_NN )
12310 template <
typename Dispatch>
12312 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
12313 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
12317 return static_cast<Result>( d.vkCreateViSurfaceNN( m_instance,
12318 reinterpret_cast<const VkViSurfaceCreateInfoNN *
>( pCreateInfo ),
12323 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12324 template <
typename Dispatch>
12326 Instance::createViSurfaceNN(
const VULKAN_HPP_NAMESPACE::ViSurfaceCreateInfoNN & createInfo,
12327 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
12328 Dispatch
const & d )
const
12331 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12332 VULKAN_HPP_ASSERT( d.vkCreateViSurfaceNN &&
"Function <vkCreateViSurfaceNN> requires <VK_NN_vi_surface>" );
12335 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
12337 d.vkCreateViSurfaceNN( m_instance,
12338 reinterpret_cast<const VkViSurfaceCreateInfoNN *
>( &createInfo ),
12339 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
12346 # ifndef VULKAN_HPP_NO_SMART_HANDLE
12347 template <
typename Dispatch>
12349 Instance::createViSurfaceNNUnique(
const VULKAN_HPP_NAMESPACE::ViSurfaceCreateInfoNN & createInfo,
12350 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
12351 Dispatch
const & d )
const
12354 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12355 VULKAN_HPP_ASSERT( d.vkCreateViSurfaceNN &&
"Function <vkCreateViSurfaceNN> requires <VK_NN_vi_surface>" );
12358 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
12360 d.vkCreateViSurfaceNN( m_instance,
12361 reinterpret_cast<const VkViSurfaceCreateInfoNN *
>( &createInfo ),
12362 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
12367 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
12375 template <
typename Dispatch>
12386 template <
typename Dispatch>
12388 Instance::enumeratePhysicalDeviceGroupsKHR( uint32_t * pPhysicalDeviceGroupCount,
12389 VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties * pPhysicalDeviceGroupProperties,
12393 return static_cast<Result>( d.vkEnumeratePhysicalDeviceGroupsKHR(
12397 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12398 template <
typename PhysicalDeviceGroupPropertiesAllocator,
typename Dispatch>
12401 Instance::enumeratePhysicalDeviceGroupsKHR( Dispatch
const & d )
const
12404 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12406 "Function <vkEnumeratePhysicalDeviceGroupsKHR> requires <VK_KHR_device_group_creation> or <VK_VERSION_1_1>" );
12409 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator> physicalDeviceGroupProperties;
12410 uint32_t physicalDeviceGroupCount;
12414 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkEnumeratePhysicalDeviceGroupsKHR( m_instance, &physicalDeviceGroupCount,
nullptr ) );
12415 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && physicalDeviceGroupCount )
12417 physicalDeviceGroupProperties.resize( physicalDeviceGroupCount );
12421 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
12423 VULKAN_HPP_ASSERT( physicalDeviceGroupCount <= physicalDeviceGroupProperties.size() );
12424 if ( physicalDeviceGroupCount < physicalDeviceGroupProperties.size() )
12426 physicalDeviceGroupProperties.resize( physicalDeviceGroupCount );
12431 template <
typename PhysicalDeviceGroupPropertiesAllocator,
12433 typename std::enable_if<
12434 std::is_same<typename PhysicalDeviceGroupPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties>::value,
12438 Instance::enumeratePhysicalDeviceGroupsKHR( PhysicalDeviceGroupPropertiesAllocator & physicalDeviceGroupPropertiesAllocator, Dispatch
const & d )
const
12441 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12443 "Function <vkEnumeratePhysicalDeviceGroupsKHR> requires <VK_KHR_device_group_creation> or <VK_VERSION_1_1>" );
12446 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator> physicalDeviceGroupProperties(
12447 physicalDeviceGroupPropertiesAllocator );
12448 uint32_t physicalDeviceGroupCount;
12452 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkEnumeratePhysicalDeviceGroupsKHR( m_instance, &physicalDeviceGroupCount,
nullptr ) );
12453 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && physicalDeviceGroupCount )
12455 physicalDeviceGroupProperties.resize( physicalDeviceGroupCount );
12459 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
12461 VULKAN_HPP_ASSERT( physicalDeviceGroupCount <= physicalDeviceGroupProperties.size() );
12462 if ( physicalDeviceGroupCount < physicalDeviceGroupProperties.size() )
12464 physicalDeviceGroupProperties.resize( physicalDeviceGroupCount );
12472 template <
typename Dispatch>
12473 VULKAN_HPP_INLINE void PhysicalDevice::getExternalBufferPropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalBufferInfo * pExternalBufferInfo,
12474 VULKAN_HPP_NAMESPACE::ExternalBufferProperties * pExternalBufferProperties,
12478 d.vkGetPhysicalDeviceExternalBufferPropertiesKHR( m_physicalDevice,
12483 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12484 template <
typename Dispatch>
12486 PhysicalDevice::getExternalBufferPropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalBufferInfo & externalBufferInfo,
12490 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12492 "Function <vkGetPhysicalDeviceExternalBufferPropertiesKHR> requires <VK_KHR_external_memory_capabilities> or <VK_VERSION_1_1>" );
12495 VULKAN_HPP_NAMESPACE::ExternalBufferProperties externalBufferProperties;
12496 d.vkGetPhysicalDeviceExternalBufferPropertiesKHR( m_physicalDevice,
12500 return externalBufferProperties;
12504 #if defined( VK_USE_PLATFORM_WIN32_KHR )
12507 template <
typename Dispatch>
12513 return static_cast<Result>(
12514 d.vkGetMemoryWin32HandleKHR( m_device,
reinterpret_cast<const VkMemoryGetWin32HandleInfoKHR *
>( pGetWin32HandleInfo ), pHandle ) );
12517 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12518 template <
typename Dispatch>
12520 Device::getMemoryWin32HandleKHR(
const VULKAN_HPP_NAMESPACE::MemoryGetWin32HandleInfoKHR & getWin32HandleInfo, Dispatch
const & d )
const
12523 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12524 VULKAN_HPP_ASSERT( d.vkGetMemoryWin32HandleKHR &&
"Function <vkGetMemoryWin32HandleKHR> requires <VK_KHR_external_memory_win32>" );
12529 d.vkGetMemoryWin32HandleKHR( m_device,
reinterpret_cast<const VkMemoryGetWin32HandleInfoKHR *
>( &getWin32HandleInfo ), &handle ) );
12536 template <
typename Dispatch>
12540 VULKAN_HPP_NAMESPACE::MemoryWin32HandlePropertiesKHR * pMemoryWin32HandleProperties,
12544 return static_cast<Result>( d.vkGetMemoryWin32HandlePropertiesKHR( m_device,
12547 reinterpret_cast<VkMemoryWin32HandlePropertiesKHR *
>( pMemoryWin32HandleProperties ) ) );
12550 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12551 template <
typename Dispatch>
12556 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12557 VULKAN_HPP_ASSERT( d.vkGetMemoryWin32HandlePropertiesKHR &&
"Function <vkGetMemoryWin32HandlePropertiesKHR> requires <VK_KHR_external_memory_win32>" );
12560 VULKAN_HPP_NAMESPACE::MemoryWin32HandlePropertiesKHR memoryWin32HandleProperties;
12562 d.vkGetMemoryWin32HandlePropertiesKHR( m_device,
12565 reinterpret_cast<VkMemoryWin32HandlePropertiesKHR *
>( &memoryWin32HandleProperties ) ) );
12575 template <
typename Dispatch>
12581 return static_cast<Result>( d.vkGetMemoryFdKHR( m_device,
reinterpret_cast<const VkMemoryGetFdInfoKHR *
>( pGetFdInfo ), pFd ) );
12584 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12585 template <
typename Dispatch>
12587 Dispatch
const & d )
const
12590 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12591 VULKAN_HPP_ASSERT( d.vkGetMemoryFdKHR &&
"Function <vkGetMemoryFdKHR> requires <VK_KHR_external_memory_fd>" );
12603 template <
typename Dispatch>
12606 VULKAN_HPP_NAMESPACE::MemoryFdPropertiesKHR * pMemoryFdProperties,
12610 return static_cast<Result>( d.vkGetMemoryFdPropertiesKHR(
12614 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12615 template <
typename Dispatch>
12620 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12621 VULKAN_HPP_ASSERT( d.vkGetMemoryFdPropertiesKHR &&
"Function <vkGetMemoryFdPropertiesKHR> requires <VK_KHR_external_memory_fd>" );
12624 VULKAN_HPP_NAMESPACE::MemoryFdPropertiesKHR memoryFdProperties;
12635 template <
typename Dispatch>
12637 PhysicalDevice::getExternalSemaphorePropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalSemaphoreInfo * pExternalSemaphoreInfo,
12638 VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties * pExternalSemaphoreProperties,
12642 d.vkGetPhysicalDeviceExternalSemaphorePropertiesKHR( m_physicalDevice,
12647 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12648 template <
typename Dispatch>
12650 PhysicalDevice::getExternalSemaphorePropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalSemaphoreInfo & externalSemaphoreInfo,
12654 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12656 "Function <vkGetPhysicalDeviceExternalSemaphorePropertiesKHR> requires <VK_KHR_external_semaphore_capabilities> or <VK_VERSION_1_1>" );
12659 VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties externalSemaphoreProperties;
12660 d.vkGetPhysicalDeviceExternalSemaphorePropertiesKHR( m_physicalDevice,
12664 return externalSemaphoreProperties;
12668 #if defined( VK_USE_PLATFORM_WIN32_KHR )
12671 template <
typename Dispatch>
12673 const VULKAN_HPP_NAMESPACE::ImportSemaphoreWin32HandleInfoKHR * pImportSemaphoreWin32HandleInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
12676 return static_cast<Result>(
12677 d.vkImportSemaphoreWin32HandleKHR( m_device,
reinterpret_cast<const VkImportSemaphoreWin32HandleInfoKHR *
>( pImportSemaphoreWin32HandleInfo ) ) );
12680 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12681 template <
typename Dispatch>
12683 Device::importSemaphoreWin32HandleKHR(
const VULKAN_HPP_NAMESPACE::ImportSemaphoreWin32HandleInfoKHR & importSemaphoreWin32HandleInfo,
12684 Dispatch
const & d )
const
12687 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12688 VULKAN_HPP_ASSERT( d.vkImportSemaphoreWin32HandleKHR &&
"Function <vkImportSemaphoreWin32HandleKHR> requires <VK_KHR_external_semaphore_win32>" );
12692 d.vkImportSemaphoreWin32HandleKHR( m_device,
reinterpret_cast<const VkImportSemaphoreWin32HandleInfoKHR *
>( &importSemaphoreWin32HandleInfo ) ) );
12699 template <
typename Dispatch>
12701 const VULKAN_HPP_NAMESPACE::SemaphoreGetWin32HandleInfoKHR * pGetWin32HandleInfo, HANDLE * pHandle, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
12704 return static_cast<Result>(
12705 d.vkGetSemaphoreWin32HandleKHR( m_device,
reinterpret_cast<const VkSemaphoreGetWin32HandleInfoKHR *
>( pGetWin32HandleInfo ), pHandle ) );
12708 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12709 template <
typename Dispatch>
12711 Device::getSemaphoreWin32HandleKHR(
const VULKAN_HPP_NAMESPACE::SemaphoreGetWin32HandleInfoKHR & getWin32HandleInfo, Dispatch
const & d )
const
12714 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12715 VULKAN_HPP_ASSERT( d.vkGetSemaphoreWin32HandleKHR &&
"Function <vkGetSemaphoreWin32HandleKHR> requires <VK_KHR_external_semaphore_win32>" );
12720 d.vkGetSemaphoreWin32HandleKHR( m_device,
reinterpret_cast<const VkSemaphoreGetWin32HandleInfoKHR *
>( &getWin32HandleInfo ), &handle ) );
12730 template <
typename Dispatch>
12738 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12739 template <
typename Dispatch>
12741 Device::importSemaphoreFdKHR(
const VULKAN_HPP_NAMESPACE::ImportSemaphoreFdInfoKHR & importSemaphoreFdInfo, Dispatch
const & d )
const
12744 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12745 VULKAN_HPP_ASSERT( d.vkImportSemaphoreFdKHR &&
"Function <vkImportSemaphoreFdKHR> requires <VK_KHR_external_semaphore_fd>" );
12756 template <
typename Dispatch>
12765 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12766 template <
typename Dispatch>
12768 Device::getSemaphoreFdKHR(
const VULKAN_HPP_NAMESPACE::SemaphoreGetFdInfoKHR & getFdInfo, Dispatch
const & d )
const
12771 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12772 VULKAN_HPP_ASSERT( d.vkGetSemaphoreFdKHR &&
"Function <vkGetSemaphoreFdKHR> requires <VK_KHR_external_semaphore_fd>" );
12786 template <
typename Dispatch>
12788 VULKAN_HPP_NAMESPACE::PipelineLayout layout,
12790 uint32_t descriptorWriteCount,
12791 const VULKAN_HPP_NAMESPACE::WriteDescriptorSet * pDescriptorWrites,
12795 d.vkCmdPushDescriptorSetKHR( m_commandBuffer,
12799 descriptorWriteCount,
12803 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12804 template <
typename Dispatch>
12807 VULKAN_HPP_NAMESPACE::PipelineLayout layout,
12809 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::WriteDescriptorSet>
const & descriptorWrites,
12813 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12814 VULKAN_HPP_ASSERT( d.vkCmdPushDescriptorSetKHR &&
"Function <vkCmdPushDescriptorSetKHR> requires <VK_KHR_push_descriptor>" );
12817 d.vkCmdPushDescriptorSetKHR( m_commandBuffer,
12821 descriptorWrites.size(),
12826 template <
typename Dispatch>
12827 VULKAN_HPP_INLINE void CommandBuffer::pushDescriptorSetWithTemplateKHR( VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
12828 VULKAN_HPP_NAMESPACE::PipelineLayout layout,
12830 const void * pData,
12834 d.vkCmdPushDescriptorSetWithTemplateKHR(
12838 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12839 template <
typename DataType,
typename Dispatch>
12840 VULKAN_HPP_INLINE void CommandBuffer::pushDescriptorSetWithTemplateKHR( VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
12841 VULKAN_HPP_NAMESPACE::PipelineLayout layout,
12843 DataType
const & data,
12847 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12849 "Function <vkCmdPushDescriptorSetWithTemplateKHR> requires <VK_KHR_descriptor_update_template> or <VK_KHR_push_descriptor>" );
12852 d.vkCmdPushDescriptorSetWithTemplateKHR( m_commandBuffer,
12856 reinterpret_cast<const void *
>( &data ) );
12862 template <
typename Dispatch>
12863 VULKAN_HPP_INLINE void CommandBuffer::beginConditionalRenderingEXT(
const VULKAN_HPP_NAMESPACE::ConditionalRenderingBeginInfoEXT * pConditionalRenderingBegin,
12870 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12871 template <
typename Dispatch>
12872 VULKAN_HPP_INLINE void CommandBuffer::beginConditionalRenderingEXT(
const VULKAN_HPP_NAMESPACE::ConditionalRenderingBeginInfoEXT & conditionalRenderingBegin,
12876 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12877 VULKAN_HPP_ASSERT( d.vkCmdBeginConditionalRenderingEXT &&
"Function <vkCmdBeginConditionalRenderingEXT> requires <VK_EXT_conditional_rendering>" );
12884 template <
typename Dispatch>
12888 d.vkCmdEndConditionalRenderingEXT( m_commandBuffer );
12893 template <
typename Dispatch>
12895 Device::createDescriptorUpdateTemplateKHR(
const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo * pCreateInfo,
12896 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
12897 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate * pDescriptorUpdateTemplate,
12901 return static_cast<Result>( d.vkCreateDescriptorUpdateTemplateKHR( m_device,
12907 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12908 template <
typename Dispatch>
12910 Device::createDescriptorUpdateTemplateKHR(
const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo & createInfo,
12912 Dispatch
const & d )
const
12915 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12917 "Function <vkCreateDescriptorUpdateTemplateKHR> requires <VK_KHR_descriptor_update_template> or <VK_VERSION_1_1>" );
12920 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate;
12924 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
12931 # ifndef VULKAN_HPP_NO_SMART_HANDLE
12932 template <
typename Dispatch>
12934 Device::createDescriptorUpdateTemplateKHRUnique(
const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo & createInfo,
12936 Dispatch
const & d )
const
12939 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12941 "Function <vkCreateDescriptorUpdateTemplateKHR> requires <VK_KHR_descriptor_update_template> or <VK_VERSION_1_1>" );
12944 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate;
12948 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
12959 template <
typename Dispatch>
12960 VULKAN_HPP_INLINE void Device::destroyDescriptorUpdateTemplateKHR( VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
12961 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
12965 d.vkDestroyDescriptorUpdateTemplateKHR(
12969 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
12970 template <
typename Dispatch>
12971 VULKAN_HPP_INLINE void Device::destroyDescriptorUpdateTemplateKHR( VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
12976 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
12978 "Function <vkDestroyDescriptorUpdateTemplateKHR> requires <VK_KHR_descriptor_update_template> or <VK_VERSION_1_1>" );
12981 d.vkDestroyDescriptorUpdateTemplateKHR(
12984 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
12988 template <
typename Dispatch>
12989 VULKAN_HPP_INLINE void Device::updateDescriptorSetWithTemplateKHR( VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet,
12990 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
12991 const void * pData,
12995 d.vkUpdateDescriptorSetWithTemplateKHR(
12999 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13000 template <
typename DataType,
typename Dispatch>
13001 VULKAN_HPP_INLINE void Device::updateDescriptorSetWithTemplateKHR( VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet,
13002 VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate,
13003 DataType
const & data,
13007 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13009 "Function <vkUpdateDescriptorSetWithTemplateKHR> requires <VK_KHR_descriptor_update_template> or <VK_VERSION_1_1>" );
13012 d.vkUpdateDescriptorSetWithTemplateKHR( m_device,
13015 reinterpret_cast<const void *
>( &data ) );
13021 template <
typename Dispatch>
13023 uint32_t viewportCount,
13024 const VULKAN_HPP_NAMESPACE::ViewportWScalingNV * pViewportWScalings,
13028 d.vkCmdSetViewportWScalingNV( m_commandBuffer, firstViewport, viewportCount,
reinterpret_cast<const VkViewportWScalingNV *
>( pViewportWScalings ) );
13031 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13032 template <
typename Dispatch>
13034 CommandBuffer::setViewportWScalingNV( uint32_t firstViewport,
13035 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ViewportWScalingNV>
const & viewportWScalings,
13039 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13040 VULKAN_HPP_ASSERT( d.vkCmdSetViewportWScalingNV &&
"Function <vkCmdSetViewportWScalingNV> requires <VK_NV_clip_space_w_scaling>" );
13043 d.vkCmdSetViewportWScalingNV(
13044 m_commandBuffer, firstViewport, viewportWScalings.size(),
reinterpret_cast<const VkViewportWScalingNV *
>( viewportWScalings.data() ) );
13050 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
13051 template <
typename Dispatch>
13055 return static_cast<Result>( d.vkReleaseDisplayEXT( m_physicalDevice,
static_cast<VkDisplayKHR>( display ) ) );
13058 template <
typename Dispatch>
13062 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13063 VULKAN_HPP_ASSERT( d.vkReleaseDisplayEXT &&
"Function <vkReleaseDisplayEXT> requires <VK_EXT_direct_mode_display>" );
13066 d.vkReleaseDisplayEXT( m_physicalDevice,
static_cast<VkDisplayKHR>( display ) );
13070 #if defined( VK_USE_PLATFORM_XLIB_XRANDR_EXT )
13073 template <
typename Dispatch>
13075 VULKAN_HPP_NAMESPACE::DisplayKHR display,
13079 return static_cast<Result>( d.vkAcquireXlibDisplayEXT( m_physicalDevice, dpy,
static_cast<VkDisplayKHR>( display ) ) );
13082 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13083 template <
typename Dispatch>
13085 PhysicalDevice::acquireXlibDisplayEXT( Display & dpy, VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch
const & d )
const
13088 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13089 VULKAN_HPP_ASSERT( d.vkAcquireXlibDisplayEXT &&
"Function <vkAcquireXlibDisplayEXT> requires <VK_EXT_acquire_xlib_display>" );
13100 template <
typename Dispatch>
13103 VULKAN_HPP_NAMESPACE::DisplayKHR * pDisplay,
13107 return static_cast<Result>( d.vkGetRandROutputDisplayEXT( m_physicalDevice, dpy, rrOutput,
reinterpret_cast<VkDisplayKHR *
>( pDisplay ) ) );
13110 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13111 template <
typename Dispatch>
13113 PhysicalDevice::getRandROutputDisplayEXT( Display & dpy, RROutput rrOutput, Dispatch
const & d )
const
13116 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13117 VULKAN_HPP_ASSERT( d.vkGetRandROutputDisplayEXT &&
"Function <vkGetRandROutputDisplayEXT> requires <VK_EXT_acquire_xlib_display>" );
13120 VULKAN_HPP_NAMESPACE::DisplayKHR display;
13122 d.vkGetRandROutputDisplayEXT( m_physicalDevice, &dpy, rrOutput,
reinterpret_cast<VkDisplayKHR *
>( &display ) ) );
13128 # ifndef VULKAN_HPP_NO_SMART_HANDLE
13129 template <
typename Dispatch>
13131 PhysicalDevice::getRandROutputDisplayEXTUnique( Display & dpy, RROutput rrOutput, Dispatch
const & d )
const
13134 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13135 VULKAN_HPP_ASSERT( d.vkGetRandROutputDisplayEXT &&
"Function <vkGetRandROutputDisplayEXT> requires <VK_EXT_acquire_xlib_display>" );
13138 VULKAN_HPP_NAMESPACE::DisplayKHR display;
13140 d.vkGetRandROutputDisplayEXT( m_physicalDevice, &dpy, rrOutput,
reinterpret_cast<VkDisplayKHR *
>( &display ) ) );
13144 UniqueHandle<VULKAN_HPP_NAMESPACE::DisplayKHR, Dispatch>( display, ObjectRelease<PhysicalDevice, Dispatch>( *
this, d ) ) );
13152 template <
typename Dispatch>
13154 PhysicalDevice::getSurfaceCapabilities2EXT( VULKAN_HPP_NAMESPACE::SurfaceKHR surface,
13155 VULKAN_HPP_NAMESPACE::SurfaceCapabilities2EXT * pSurfaceCapabilities,
13159 return static_cast<Result>( d.vkGetPhysicalDeviceSurfaceCapabilities2EXT(
13163 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13164 template <
typename Dispatch>
13166 PhysicalDevice::getSurfaceCapabilities2EXT( VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch
const & d )
const
13169 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13171 "Function <vkGetPhysicalDeviceSurfaceCapabilities2EXT> requires <VK_EXT_display_surface_counter>" );
13174 VULKAN_HPP_NAMESPACE::SurfaceCapabilities2EXT surfaceCapabilities;
13185 template <
typename Dispatch>
13187 const VULKAN_HPP_NAMESPACE::DisplayPowerInfoEXT * pDisplayPowerInfo,
13191 return static_cast<Result>(
13195 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13196 template <
typename Dispatch>
13198 const VULKAN_HPP_NAMESPACE::DisplayPowerInfoEXT & displayPowerInfo,
13199 Dispatch
const & d )
const
13202 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13203 VULKAN_HPP_ASSERT( d.vkDisplayPowerControlEXT &&
"Function <vkDisplayPowerControlEXT> requires <VK_EXT_display_control>" );
13214 template <
typename Dispatch>
13216 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
13217 VULKAN_HPP_NAMESPACE::Fence * pFence,
13221 return static_cast<Result>( d.vkRegisterDeviceEventEXT( m_device,
13224 reinterpret_cast<VkFence *
>( pFence ) ) );
13227 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13228 template <
typename Dispatch>
13230 Device::registerEventEXT(
const VULKAN_HPP_NAMESPACE::DeviceEventInfoEXT & deviceEventInfo,
13232 Dispatch
const & d )
const
13235 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13236 VULKAN_HPP_ASSERT( d.vkRegisterDeviceEventEXT &&
"Function <vkRegisterDeviceEventEXT> requires <VK_EXT_display_control>" );
13239 VULKAN_HPP_NAMESPACE::Fence fence;
13243 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
13244 reinterpret_cast<VkFence *
>( &fence ) ) );
13250 # ifndef VULKAN_HPP_NO_SMART_HANDLE
13251 template <
typename Dispatch>
13253 Device::registerEventEXTUnique(
const VULKAN_HPP_NAMESPACE::DeviceEventInfoEXT & deviceEventInfo,
13255 Dispatch
const & d )
const
13258 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13259 VULKAN_HPP_ASSERT( d.vkRegisterDeviceEventEXT &&
"Function <vkRegisterDeviceEventEXT> requires <VK_EXT_display_control>" );
13262 VULKAN_HPP_NAMESPACE::Fence fence;
13266 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
13267 reinterpret_cast<VkFence *
>( &fence ) ) );
13276 template <
typename Dispatch>
13278 const VULKAN_HPP_NAMESPACE::DisplayEventInfoEXT * pDisplayEventInfo,
13279 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
13280 VULKAN_HPP_NAMESPACE::Fence * pFence,
13284 return static_cast<Result>( d.vkRegisterDisplayEventEXT( m_device,
13288 reinterpret_cast<VkFence *
>( pFence ) ) );
13291 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13292 template <
typename Dispatch>
13294 Device::registerDisplayEventEXT( VULKAN_HPP_NAMESPACE::DisplayKHR display,
13295 const VULKAN_HPP_NAMESPACE::DisplayEventInfoEXT & displayEventInfo,
13297 Dispatch
const & d )
const
13300 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13301 VULKAN_HPP_ASSERT( d.vkRegisterDisplayEventEXT &&
"Function <vkRegisterDisplayEventEXT> requires <VK_EXT_display_control>" );
13304 VULKAN_HPP_NAMESPACE::Fence fence;
13309 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
13310 reinterpret_cast<VkFence *
>( &fence ) ) );
13316 # ifndef VULKAN_HPP_NO_SMART_HANDLE
13317 template <
typename Dispatch>
13319 Device::registerDisplayEventEXTUnique( VULKAN_HPP_NAMESPACE::DisplayKHR display,
13320 const VULKAN_HPP_NAMESPACE::DisplayEventInfoEXT & displayEventInfo,
13322 Dispatch
const & d )
const
13325 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13326 VULKAN_HPP_ASSERT( d.vkRegisterDisplayEventEXT &&
"Function <vkRegisterDisplayEventEXT> requires <VK_EXT_display_control>" );
13329 VULKAN_HPP_NAMESPACE::Fence fence;
13334 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
13335 reinterpret_cast<VkFence *
>( &fence ) ) );
13344 template <
typename Dispatch>
13347 uint64_t * pCounterValue,
13351 return static_cast<Result>(
13355 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13356 template <
typename Dispatch>
13361 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13362 VULKAN_HPP_ASSERT( d.vkGetSwapchainCounterEXT &&
"Function <vkGetSwapchainCounterEXT> requires <VK_EXT_display_control>" );
13365 uint64_t counterValue;
13376 template <
typename Dispatch>
13378 Device::getRefreshCycleDurationGOOGLE( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain,
13379 VULKAN_HPP_NAMESPACE::RefreshCycleDurationGOOGLE * pDisplayTimingProperties,
13383 return static_cast<Result>( d.vkGetRefreshCycleDurationGOOGLE(
13387 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13388 template <
typename Dispatch>
13390 Device::getRefreshCycleDurationGOOGLE( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch
const & d )
const
13393 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13394 VULKAN_HPP_ASSERT( d.vkGetRefreshCycleDurationGOOGLE &&
"Function <vkGetRefreshCycleDurationGOOGLE> requires <VK_GOOGLE_display_timing>" );
13397 VULKAN_HPP_NAMESPACE::RefreshCycleDurationGOOGLE displayTimingProperties;
13406 template <
typename Dispatch>
13408 Device::getPastPresentationTimingGOOGLE( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain,
13409 uint32_t * pPresentationTimingCount,
13410 VULKAN_HPP_NAMESPACE::PastPresentationTimingGOOGLE * pPresentationTimings,
13414 return static_cast<Result>( d.vkGetPastPresentationTimingGOOGLE( m_device,
13416 pPresentationTimingCount,
13420 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13421 template <
typename PastPresentationTimingGOOGLEAllocator,
typename Dispatch>
13424 Device::getPastPresentationTimingGOOGLE( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch
const & d )
const
13427 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13428 VULKAN_HPP_ASSERT( d.vkGetPastPresentationTimingGOOGLE &&
"Function <vkGetPastPresentationTimingGOOGLE> requires <VK_GOOGLE_display_timing>" );
13431 std::vector<VULKAN_HPP_NAMESPACE::PastPresentationTimingGOOGLE, PastPresentationTimingGOOGLEAllocator> presentationTimings;
13432 uint32_t presentationTimingCount;
13437 d.vkGetPastPresentationTimingGOOGLE( m_device,
static_cast<VkSwapchainKHR>( swapchain ), &presentationTimingCount,
nullptr ) );
13438 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && presentationTimingCount )
13440 presentationTimings.resize( presentationTimingCount );
13442 d.vkGetPastPresentationTimingGOOGLE( m_device,
13444 &presentationTimingCount,
13447 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
13450 if ( presentationTimingCount < presentationTimings.size() )
13452 presentationTimings.resize( presentationTimingCount );
13458 typename PastPresentationTimingGOOGLEAllocator,
13460 typename std::enable_if<std::is_same<typename PastPresentationTimingGOOGLEAllocator::value_type, VULKAN_HPP_NAMESPACE::PastPresentationTimingGOOGLE>::value,
13464 Device::getPastPresentationTimingGOOGLE( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain,
13465 PastPresentationTimingGOOGLEAllocator & pastPresentationTimingGOOGLEAllocator,
13466 Dispatch
const & d )
const
13469 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13470 VULKAN_HPP_ASSERT( d.vkGetPastPresentationTimingGOOGLE &&
"Function <vkGetPastPresentationTimingGOOGLE> requires <VK_GOOGLE_display_timing>" );
13473 std::vector<VULKAN_HPP_NAMESPACE::PastPresentationTimingGOOGLE, PastPresentationTimingGOOGLEAllocator> presentationTimings(
13474 pastPresentationTimingGOOGLEAllocator );
13475 uint32_t presentationTimingCount;
13480 d.vkGetPastPresentationTimingGOOGLE( m_device,
static_cast<VkSwapchainKHR>( swapchain ), &presentationTimingCount,
nullptr ) );
13481 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && presentationTimingCount )
13483 presentationTimings.resize( presentationTimingCount );
13485 d.vkGetPastPresentationTimingGOOGLE( m_device,
13487 &presentationTimingCount,
13490 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
13493 if ( presentationTimingCount < presentationTimings.size() )
13495 presentationTimings.resize( presentationTimingCount );
13503 template <
typename Dispatch>
13505 uint32_t discardRectangleCount,
13506 const VULKAN_HPP_NAMESPACE::Rect2D * pDiscardRectangles,
13510 d.vkCmdSetDiscardRectangleEXT( m_commandBuffer, firstDiscardRectangle, discardRectangleCount,
reinterpret_cast<const VkRect2D *
>( pDiscardRectangles ) );
13513 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13514 template <
typename Dispatch>
13516 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D>
const & discardRectangles,
13520 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13521 VULKAN_HPP_ASSERT( d.vkCmdSetDiscardRectangleEXT &&
"Function <vkCmdSetDiscardRectangleEXT> requires <VK_EXT_discard_rectangles>" );
13524 d.vkCmdSetDiscardRectangleEXT(
13525 m_commandBuffer, firstDiscardRectangle, discardRectangles.size(),
reinterpret_cast<const VkRect2D *
>( discardRectangles.data() ) );
13529 template <
typename Dispatch>
13534 d.vkCmdSetDiscardRectangleEnableEXT( m_commandBuffer,
static_cast<VkBool32>( discardRectangleEnable ) );
13537 template <
typename Dispatch>
13547 template <
typename Dispatch>
13549 const VULKAN_HPP_NAMESPACE::SwapchainKHR * pSwapchains,
13550 const VULKAN_HPP_NAMESPACE::HdrMetadataEXT * pMetadata,
13554 d.vkSetHdrMetadataEXT(
13558 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13559 template <
typename Dispatch>
13560 VULKAN_HPP_INLINE void Device::setHdrMetadataEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainKHR>
const & swapchains,
13561 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::HdrMetadataEXT>
const & metadata,
13565 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13566 VULKAN_HPP_ASSERT( d.vkSetHdrMetadataEXT &&
"Function <vkSetHdrMetadataEXT> requires <VK_EXT_hdr_metadata>" );
13568 # ifdef VULKAN_HPP_NO_EXCEPTIONS
13571 if ( swapchains.size() != metadata.size() )
13577 d.vkSetHdrMetadataEXT( m_device,
13586 template <
typename Dispatch>
13588 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
13589 VULKAN_HPP_NAMESPACE::RenderPass * pRenderPass,
13593 return static_cast<Result>( d.vkCreateRenderPass2KHR( m_device,
13599 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13600 template <
typename Dispatch>
13602 Device::createRenderPass2KHR(
const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 & createInfo,
13604 Dispatch
const & d )
const
13607 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13608 VULKAN_HPP_ASSERT( d.vkCreateRenderPass2KHR &&
"Function <vkCreateRenderPass2KHR> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
13611 VULKAN_HPP_NAMESPACE::RenderPass renderPass;
13613 d.vkCreateRenderPass2KHR( m_device,
13615 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
13622 # ifndef VULKAN_HPP_NO_SMART_HANDLE
13623 template <
typename Dispatch>
13625 Device::createRenderPass2KHRUnique(
const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 & createInfo,
13627 Dispatch
const & d )
const
13630 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13631 VULKAN_HPP_ASSERT( d.vkCreateRenderPass2KHR &&
"Function <vkCreateRenderPass2KHR> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
13634 VULKAN_HPP_NAMESPACE::RenderPass renderPass;
13636 d.vkCreateRenderPass2KHR( m_device,
13638 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
13648 template <
typename Dispatch>
13649 VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass2KHR(
const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo * pRenderPassBegin,
13650 const VULKAN_HPP_NAMESPACE::SubpassBeginInfo * pSubpassBeginInfo,
13654 d.vkCmdBeginRenderPass2KHR(
13658 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13659 template <
typename Dispatch>
13660 VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass2KHR(
const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo & renderPassBegin,
13661 const VULKAN_HPP_NAMESPACE::SubpassBeginInfo & subpassBeginInfo,
13665 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13666 VULKAN_HPP_ASSERT( d.vkCmdBeginRenderPass2KHR &&
"Function <vkCmdBeginRenderPass2KHR> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
13669 d.vkCmdBeginRenderPass2KHR(
13674 template <
typename Dispatch>
13675 VULKAN_HPP_INLINE void CommandBuffer::nextSubpass2KHR(
const VULKAN_HPP_NAMESPACE::SubpassBeginInfo * pSubpassBeginInfo,
13676 const VULKAN_HPP_NAMESPACE::SubpassEndInfo * pSubpassEndInfo,
13680 d.vkCmdNextSubpass2KHR(
13684 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13685 template <
typename Dispatch>
13686 VULKAN_HPP_INLINE void CommandBuffer::nextSubpass2KHR(
const VULKAN_HPP_NAMESPACE::SubpassBeginInfo & subpassBeginInfo,
13687 const VULKAN_HPP_NAMESPACE::SubpassEndInfo & subpassEndInfo,
13691 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13692 VULKAN_HPP_ASSERT( d.vkCmdNextSubpass2KHR &&
"Function <vkCmdNextSubpass2KHR> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
13695 d.vkCmdNextSubpass2KHR(
13700 template <
typename Dispatch>
13701 VULKAN_HPP_INLINE void CommandBuffer::endRenderPass2KHR(
const VULKAN_HPP_NAMESPACE::SubpassEndInfo * pSubpassEndInfo,
13705 d.vkCmdEndRenderPass2KHR( m_commandBuffer,
reinterpret_cast<const VkSubpassEndInfo *
>( pSubpassEndInfo ) );
13708 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13709 template <
typename Dispatch>
13710 VULKAN_HPP_INLINE void CommandBuffer::endRenderPass2KHR(
const VULKAN_HPP_NAMESPACE::SubpassEndInfo & subpassEndInfo,
13714 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13715 VULKAN_HPP_ASSERT( d.vkCmdEndRenderPass2KHR &&
"Function <vkCmdEndRenderPass2KHR> requires <VK_KHR_create_renderpass2> or <VK_VERSION_1_2>" );
13718 d.vkCmdEndRenderPass2KHR( m_commandBuffer,
reinterpret_cast<const VkSubpassEndInfo *
>( &subpassEndInfo ) );
13724 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
13725 template <
typename Dispatch>
13730 return static_cast<Result>( d.vkGetSwapchainStatusKHR( m_device,
static_cast<VkSwapchainKHR>( swapchain ) ) );
13733 template <
typename Dispatch>
13735 Dispatch
const & d )
const
13738 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13739 VULKAN_HPP_ASSERT( d.vkGetSwapchainStatusKHR &&
"Function <vkGetSwapchainStatusKHR> requires <VK_KHR_shared_presentable_image>" );
13746 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR } );
13754 template <
typename Dispatch>
13755 VULKAN_HPP_INLINE void PhysicalDevice::getExternalFencePropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalFenceInfo * pExternalFenceInfo,
13756 VULKAN_HPP_NAMESPACE::ExternalFenceProperties * pExternalFenceProperties,
13760 d.vkGetPhysicalDeviceExternalFencePropertiesKHR( m_physicalDevice,
13765 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13766 template <
typename Dispatch>
13768 PhysicalDevice::getExternalFencePropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalFenceInfo & externalFenceInfo,
13772 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13774 "Function <vkGetPhysicalDeviceExternalFencePropertiesKHR> requires <VK_KHR_external_fence_capabilities> or <VK_VERSION_1_1>" );
13777 VULKAN_HPP_NAMESPACE::ExternalFenceProperties externalFenceProperties;
13778 d.vkGetPhysicalDeviceExternalFencePropertiesKHR( m_physicalDevice,
13782 return externalFenceProperties;
13786 #if defined( VK_USE_PLATFORM_WIN32_KHR )
13789 template <
typename Dispatch>
13791 const VULKAN_HPP_NAMESPACE::ImportFenceWin32HandleInfoKHR * pImportFenceWin32HandleInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
13794 return static_cast<Result>(
13795 d.vkImportFenceWin32HandleKHR( m_device,
reinterpret_cast<const VkImportFenceWin32HandleInfoKHR *
>( pImportFenceWin32HandleInfo ) ) );
13798 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13799 template <
typename Dispatch>
13801 Device::importFenceWin32HandleKHR(
const VULKAN_HPP_NAMESPACE::ImportFenceWin32HandleInfoKHR & importFenceWin32HandleInfo, Dispatch
const & d )
const
13804 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13805 VULKAN_HPP_ASSERT( d.vkImportFenceWin32HandleKHR &&
"Function <vkImportFenceWin32HandleKHR> requires <VK_KHR_external_fence_win32>" );
13809 d.vkImportFenceWin32HandleKHR( m_device,
reinterpret_cast<const VkImportFenceWin32HandleInfoKHR *
>( &importFenceWin32HandleInfo ) ) );
13816 template <
typename Dispatch>
13822 return static_cast<Result>(
13823 d.vkGetFenceWin32HandleKHR( m_device,
reinterpret_cast<const VkFenceGetWin32HandleInfoKHR *
>( pGetWin32HandleInfo ), pHandle ) );
13826 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13827 template <
typename Dispatch>
13829 Device::getFenceWin32HandleKHR(
const VULKAN_HPP_NAMESPACE::FenceGetWin32HandleInfoKHR & getWin32HandleInfo, Dispatch
const & d )
const
13832 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13833 VULKAN_HPP_ASSERT( d.vkGetFenceWin32HandleKHR &&
"Function <vkGetFenceWin32HandleKHR> requires <VK_KHR_external_fence_win32>" );
13838 d.vkGetFenceWin32HandleKHR( m_device,
reinterpret_cast<const VkFenceGetWin32HandleInfoKHR *
>( &getWin32HandleInfo ), &handle ) );
13848 template <
typename Dispatch>
13853 return static_cast<Result>( d.vkImportFenceFdKHR( m_device,
reinterpret_cast<const VkImportFenceFdInfoKHR *
>( pImportFenceFdInfo ) ) );
13856 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13857 template <
typename Dispatch>
13859 Device::importFenceFdKHR(
const VULKAN_HPP_NAMESPACE::ImportFenceFdInfoKHR & importFenceFdInfo, Dispatch
const & d )
const
13862 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13863 VULKAN_HPP_ASSERT( d.vkImportFenceFdKHR &&
"Function <vkImportFenceFdKHR> requires <VK_KHR_external_fence_fd>" );
13874 template <
typename Dispatch>
13880 return static_cast<Result>( d.vkGetFenceFdKHR( m_device,
reinterpret_cast<const VkFenceGetFdInfoKHR *
>( pGetFdInfo ), pFd ) );
13883 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13884 template <
typename Dispatch>
13886 Dispatch
const & d )
const
13889 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13890 VULKAN_HPP_ASSERT( d.vkGetFenceFdKHR &&
"Function <vkGetFenceFdKHR> requires <VK_KHR_external_fence_fd>" );
13904 template <
typename Dispatch>
13906 PhysicalDevice::enumerateQueueFamilyPerformanceQueryCountersKHR( uint32_t queueFamilyIndex,
13907 uint32_t * pCounterCount,
13908 VULKAN_HPP_NAMESPACE::PerformanceCounterKHR * pCounters,
13909 VULKAN_HPP_NAMESPACE::PerformanceCounterDescriptionKHR * pCounterDescriptions,
13913 return static_cast<Result>(
13914 d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR( m_physicalDevice,
13921 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
13922 template <
typename PerformanceCounterKHRAllocator,
typename PerformanceCounterDescriptionKHRAllocator,
typename Dispatch>
13925 std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator>>>::type
13926 PhysicalDevice::enumerateQueueFamilyPerformanceQueryCountersKHR( uint32_t queueFamilyIndex, Dispatch
const & d )
const
13929 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13930 VULKAN_HPP_ASSERT( d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR &&
13931 "Function <vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR> requires <VK_KHR_performance_query>" );
13934 std::pair<std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterKHR, PerformanceCounterKHRAllocator>,
13935 std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator>>
13937 std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterKHR, PerformanceCounterKHRAllocator> & counters = data_.first;
13938 std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator> & counterDescriptions = data_.second;
13939 uint32_t counterCount;
13944 d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR( m_physicalDevice, queueFamilyIndex, &counterCount,
nullptr,
nullptr ) );
13945 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && counterCount )
13947 counters.resize( counterCount );
13948 counterDescriptions.resize( counterCount );
13956 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
13959 if ( counterCount < counters.size() )
13961 counters.resize( counterCount );
13962 counterDescriptions.resize( counterCount );
13967 template <
typename PerformanceCounterKHRAllocator,
13968 typename PerformanceCounterDescriptionKHRAllocator,
13970 typename std::enable_if<
13971 std::is_same<typename PerformanceCounterKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::PerformanceCounterKHR>::value &&
13972 std::is_same<typename PerformanceCounterDescriptionKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::PerformanceCounterDescriptionKHR>::value,
13976 std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator>>>::type
13977 PhysicalDevice::enumerateQueueFamilyPerformanceQueryCountersKHR( uint32_t queueFamilyIndex,
13978 PerformanceCounterKHRAllocator & performanceCounterKHRAllocator,
13979 PerformanceCounterDescriptionKHRAllocator & performanceCounterDescriptionKHRAllocator,
13980 Dispatch
const & d )
const
13983 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
13984 VULKAN_HPP_ASSERT( d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR &&
13985 "Function <vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR> requires <VK_KHR_performance_query>" );
13988 std::pair<std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterKHR, PerformanceCounterKHRAllocator>,
13989 std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator>>
13991 std::piecewise_construct, std::forward_as_tuple( performanceCounterKHRAllocator ), std::forward_as_tuple( performanceCounterDescriptionKHRAllocator ) );
13992 std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterKHR, PerformanceCounterKHRAllocator> & counters = data_.first;
13993 std::vector<VULKAN_HPP_NAMESPACE::PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator> & counterDescriptions = data_.second;
13994 uint32_t counterCount;
13999 d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR( m_physicalDevice, queueFamilyIndex, &counterCount,
nullptr,
nullptr ) );
14000 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && counterCount )
14002 counters.resize( counterCount );
14003 counterDescriptions.resize( counterCount );
14011 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14014 if ( counterCount < counters.size() )
14016 counters.resize( counterCount );
14017 counterDescriptions.resize( counterCount );
14023 template <
typename Dispatch>
14025 PhysicalDevice::getQueueFamilyPerformanceQueryPassesKHR(
const VULKAN_HPP_NAMESPACE::QueryPoolPerformanceCreateInfoKHR * pPerformanceQueryCreateInfo,
14026 uint32_t * pNumPasses,
14030 d.vkGetPhysicalDeviceQueueFamilyPerformanceQueryPassesKHR(
14034 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14035 template <
typename Dispatch>
14037 const VULKAN_HPP_NAMESPACE::QueryPoolPerformanceCreateInfoKHR & performanceQueryCreateInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
14040 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14041 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceQueueFamilyPerformanceQueryPassesKHR &&
14042 "Function <vkGetPhysicalDeviceQueueFamilyPerformanceQueryPassesKHR> requires <VK_KHR_performance_query>" );
14045 uint32_t numPasses;
14046 d.vkGetPhysicalDeviceQueueFamilyPerformanceQueryPassesKHR(
14053 template <
typename Dispatch>
14061 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14062 template <
typename Dispatch>
14064 Device::acquireProfilingLockKHR(
const VULKAN_HPP_NAMESPACE::AcquireProfilingLockInfoKHR & info, Dispatch
const & d )
const
14067 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14068 VULKAN_HPP_ASSERT( d.vkAcquireProfilingLockKHR &&
"Function <vkAcquireProfilingLockKHR> requires <VK_KHR_performance_query>" );
14079 template <
typename Dispatch>
14083 d.vkReleaseProfilingLockKHR( m_device );
14088 template <
typename Dispatch>
14090 PhysicalDevice::getSurfaceCapabilities2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR * pSurfaceInfo,
14091 VULKAN_HPP_NAMESPACE::SurfaceCapabilities2KHR * pSurfaceCapabilities,
14095 return static_cast<Result>( d.vkGetPhysicalDeviceSurfaceCapabilities2KHR( m_physicalDevice,
14100 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14101 template <
typename Dispatch>
14103 PhysicalDevice::getSurfaceCapabilities2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR & surfaceInfo, Dispatch
const & d )
const
14106 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14108 "Function <vkGetPhysicalDeviceSurfaceCapabilities2KHR> requires <VK_KHR_get_surface_capabilities2>" );
14111 VULKAN_HPP_NAMESPACE::SurfaceCapabilities2KHR surfaceCapabilities;
14113 d.vkGetPhysicalDeviceSurfaceCapabilities2KHR( m_physicalDevice,
14121 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
14123 PhysicalDevice::getSurfaceCapabilities2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR & surfaceInfo, Dispatch
const & d )
const
14126 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14128 "Function <vkGetPhysicalDeviceSurfaceCapabilities2KHR> requires <VK_KHR_get_surface_capabilities2>" );
14132 VULKAN_HPP_NAMESPACE::SurfaceCapabilities2KHR & surfaceCapabilities = structureChain.template get<VULKAN_HPP_NAMESPACE::SurfaceCapabilities2KHR>();
14134 d.vkGetPhysicalDeviceSurfaceCapabilities2KHR( m_physicalDevice,
14143 template <
typename Dispatch>
14145 uint32_t * pSurfaceFormatCount,
14146 VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR * pSurfaceFormats,
14150 return static_cast<Result>( d.vkGetPhysicalDeviceSurfaceFormats2KHR( m_physicalDevice,
14152 pSurfaceFormatCount,
14156 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14157 template <
typename SurfaceFormat2KHRAllocator,
typename Dispatch>
14159 PhysicalDevice::getSurfaceFormats2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR & surfaceInfo, Dispatch
const & d )
const
14162 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14164 "Function <vkGetPhysicalDeviceSurfaceFormats2KHR> requires <VK_KHR_get_surface_capabilities2>" );
14167 std::vector<VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR, SurfaceFormat2KHRAllocator> surfaceFormats;
14168 uint32_t surfaceFormatCount;
14174 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && surfaceFormatCount )
14176 surfaceFormats.resize( surfaceFormatCount );
14178 d.vkGetPhysicalDeviceSurfaceFormats2KHR( m_physicalDevice,
14180 &surfaceFormatCount,
14183 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14186 if ( surfaceFormatCount < surfaceFormats.size() )
14188 surfaceFormats.resize( surfaceFormatCount );
14193 template <
typename SurfaceFormat2KHRAllocator,
14195 typename std::enable_if<std::is_same<typename SurfaceFormat2KHRAllocator::value_type, VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR>::value,
int>::type>
14197 PhysicalDevice::getSurfaceFormats2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR & surfaceInfo,
14198 SurfaceFormat2KHRAllocator & surfaceFormat2KHRAllocator,
14199 Dispatch
const & d )
const
14202 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14204 "Function <vkGetPhysicalDeviceSurfaceFormats2KHR> requires <VK_KHR_get_surface_capabilities2>" );
14207 std::vector<VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR, SurfaceFormat2KHRAllocator> surfaceFormats( surfaceFormat2KHRAllocator );
14208 uint32_t surfaceFormatCount;
14214 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && surfaceFormatCount )
14216 surfaceFormats.resize( surfaceFormatCount );
14218 d.vkGetPhysicalDeviceSurfaceFormats2KHR( m_physicalDevice,
14220 &surfaceFormatCount,
14223 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14226 if ( surfaceFormatCount < surfaceFormats.size() )
14228 surfaceFormats.resize( surfaceFormatCount );
14233 template <
typename StructureChain,
typename StructureChainAllocator,
typename Dispatch>
14235 PhysicalDevice::getSurfaceFormats2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR & surfaceInfo, Dispatch
const & d )
const
14238 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14240 "Function <vkGetPhysicalDeviceSurfaceFormats2KHR> requires <VK_KHR_get_surface_capabilities2>" );
14243 std::vector<StructureChain, StructureChainAllocator> structureChains;
14244 std::vector<VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR> surfaceFormats;
14245 uint32_t surfaceFormatCount;
14251 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && surfaceFormatCount )
14253 structureChains.resize( surfaceFormatCount );
14254 surfaceFormats.resize( surfaceFormatCount );
14255 for ( uint32_t i = 0; i < surfaceFormatCount; i++ )
14257 surfaceFormats[i].pNext = structureChains[i].template get<VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR>().pNext;
14260 d.vkGetPhysicalDeviceSurfaceFormats2KHR( m_physicalDevice,
14262 &surfaceFormatCount,
14265 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14268 if ( surfaceFormatCount < surfaceFormats.size() )
14270 structureChains.resize( surfaceFormatCount );
14272 for ( uint32_t i = 0; i < surfaceFormatCount; i++ )
14274 structureChains[i].template get<VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR>() = surfaceFormats[i];
14280 typename StructureChainAllocator,
14282 typename std::enable_if<std::is_same<typename StructureChainAllocator::value_type, StructureChain>::value,
int>::type>
14284 PhysicalDevice::getSurfaceFormats2KHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR & surfaceInfo,
14285 StructureChainAllocator & structureChainAllocator,
14286 Dispatch
const & d )
const
14289 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14291 "Function <vkGetPhysicalDeviceSurfaceFormats2KHR> requires <VK_KHR_get_surface_capabilities2>" );
14294 std::vector<StructureChain, StructureChainAllocator> structureChains( structureChainAllocator );
14295 std::vector<VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR> surfaceFormats;
14296 uint32_t surfaceFormatCount;
14302 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && surfaceFormatCount )
14304 structureChains.resize( surfaceFormatCount );
14305 surfaceFormats.resize( surfaceFormatCount );
14306 for ( uint32_t i = 0; i < surfaceFormatCount; i++ )
14308 surfaceFormats[i].pNext = structureChains[i].template get<VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR>().pNext;
14311 d.vkGetPhysicalDeviceSurfaceFormats2KHR( m_physicalDevice,
14313 &surfaceFormatCount,
14316 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14319 if ( surfaceFormatCount < surfaceFormats.size() )
14321 structureChains.resize( surfaceFormatCount );
14323 for ( uint32_t i = 0; i < surfaceFormatCount; i++ )
14325 structureChains[i].template get<VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR>() = surfaceFormats[i];
14333 template <
typename Dispatch>
14335 VULKAN_HPP_NAMESPACE::DisplayProperties2KHR * pProperties,
14339 return static_cast<Result>(
14340 d.vkGetPhysicalDeviceDisplayProperties2KHR( m_physicalDevice, pPropertyCount,
reinterpret_cast<VkDisplayProperties2KHR *
>( pProperties ) ) );
14343 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14344 template <
typename DisplayProperties2KHRAllocator,
typename Dispatch>
14347 PhysicalDevice::getDisplayProperties2KHR( Dispatch
const & d )
const
14350 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14352 "Function <vkGetPhysicalDeviceDisplayProperties2KHR> requires <VK_KHR_get_display_properties2>" );
14355 std::vector<VULKAN_HPP_NAMESPACE::DisplayProperties2KHR, DisplayProperties2KHRAllocator> properties;
14356 uint32_t propertyCount;
14360 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceDisplayProperties2KHR( m_physicalDevice, &propertyCount,
nullptr ) );
14361 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
14363 properties.resize( propertyCount );
14365 d.vkGetPhysicalDeviceDisplayProperties2KHR( m_physicalDevice, &propertyCount,
reinterpret_cast<VkDisplayProperties2KHR *
>( properties.data() ) ) );
14367 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14370 if ( propertyCount < properties.size() )
14372 properties.resize( propertyCount );
14378 typename DisplayProperties2KHRAllocator,
14380 typename std::enable_if<std::is_same<typename DisplayProperties2KHRAllocator::value_type, VULKAN_HPP_NAMESPACE::DisplayProperties2KHR>::value,
int>::type>
14383 PhysicalDevice::getDisplayProperties2KHR( DisplayProperties2KHRAllocator & displayProperties2KHRAllocator, Dispatch
const & d )
const
14386 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14388 "Function <vkGetPhysicalDeviceDisplayProperties2KHR> requires <VK_KHR_get_display_properties2>" );
14391 std::vector<VULKAN_HPP_NAMESPACE::DisplayProperties2KHR, DisplayProperties2KHRAllocator> properties( displayProperties2KHRAllocator );
14392 uint32_t propertyCount;
14396 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceDisplayProperties2KHR( m_physicalDevice, &propertyCount,
nullptr ) );
14397 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
14399 properties.resize( propertyCount );
14401 d.vkGetPhysicalDeviceDisplayProperties2KHR( m_physicalDevice, &propertyCount,
reinterpret_cast<VkDisplayProperties2KHR *
>( properties.data() ) ) );
14403 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14406 if ( propertyCount < properties.size() )
14408 properties.resize( propertyCount );
14414 template <
typename Dispatch>
14416 VULKAN_HPP_NAMESPACE::DisplayPlaneProperties2KHR * pProperties,
14420 return static_cast<Result>(
14421 d.vkGetPhysicalDeviceDisplayPlaneProperties2KHR( m_physicalDevice, pPropertyCount,
reinterpret_cast<VkDisplayPlaneProperties2KHR *
>( pProperties ) ) );
14424 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14425 template <
typename DisplayPlaneProperties2KHRAllocator,
typename Dispatch>
14428 PhysicalDevice::getDisplayPlaneProperties2KHR( Dispatch
const & d )
const
14431 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14433 "Function <vkGetPhysicalDeviceDisplayPlaneProperties2KHR> requires <VK_KHR_get_display_properties2>" );
14436 std::vector<VULKAN_HPP_NAMESPACE::DisplayPlaneProperties2KHR, DisplayPlaneProperties2KHRAllocator> properties;
14437 uint32_t propertyCount;
14441 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceDisplayPlaneProperties2KHR( m_physicalDevice, &propertyCount,
nullptr ) );
14442 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
14444 properties.resize( propertyCount );
14448 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14451 if ( propertyCount < properties.size() )
14453 properties.resize( propertyCount );
14459 typename DisplayPlaneProperties2KHRAllocator,
14461 typename std::enable_if<std::is_same<typename DisplayPlaneProperties2KHRAllocator::value_type, VULKAN_HPP_NAMESPACE::DisplayPlaneProperties2KHR>::value,
14465 PhysicalDevice::getDisplayPlaneProperties2KHR( DisplayPlaneProperties2KHRAllocator & displayPlaneProperties2KHRAllocator, Dispatch
const & d )
const
14468 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14470 "Function <vkGetPhysicalDeviceDisplayPlaneProperties2KHR> requires <VK_KHR_get_display_properties2>" );
14473 std::vector<VULKAN_HPP_NAMESPACE::DisplayPlaneProperties2KHR, DisplayPlaneProperties2KHRAllocator> properties( displayPlaneProperties2KHRAllocator );
14474 uint32_t propertyCount;
14478 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceDisplayPlaneProperties2KHR( m_physicalDevice, &propertyCount,
nullptr ) );
14479 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
14481 properties.resize( propertyCount );
14485 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14488 if ( propertyCount < properties.size() )
14490 properties.resize( propertyCount );
14496 template <
typename Dispatch>
14498 uint32_t * pPropertyCount,
14499 VULKAN_HPP_NAMESPACE::DisplayModeProperties2KHR * pProperties,
14503 return static_cast<Result>( d.vkGetDisplayModeProperties2KHR(
14507 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14508 template <
typename DisplayModeProperties2KHRAllocator,
typename Dispatch>
14511 PhysicalDevice::getDisplayModeProperties2KHR( VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch
const & d )
const
14514 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14515 VULKAN_HPP_ASSERT( d.vkGetDisplayModeProperties2KHR &&
"Function <vkGetDisplayModeProperties2KHR> requires <VK_KHR_get_display_properties2>" );
14518 std::vector<VULKAN_HPP_NAMESPACE::DisplayModeProperties2KHR, DisplayModeProperties2KHRAllocator> properties;
14519 uint32_t propertyCount;
14524 d.vkGetDisplayModeProperties2KHR( m_physicalDevice,
static_cast<VkDisplayKHR>( display ), &propertyCount,
nullptr ) );
14525 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
14527 properties.resize( propertyCount );
14531 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14534 if ( propertyCount < properties.size() )
14536 properties.resize( propertyCount );
14542 typename DisplayModeProperties2KHRAllocator,
14544 typename std::enable_if<std::is_same<typename DisplayModeProperties2KHRAllocator::value_type, VULKAN_HPP_NAMESPACE::DisplayModeProperties2KHR>::value,
14548 PhysicalDevice::getDisplayModeProperties2KHR( VULKAN_HPP_NAMESPACE::DisplayKHR display,
14549 DisplayModeProperties2KHRAllocator & displayModeProperties2KHRAllocator,
14550 Dispatch
const & d )
const
14553 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14554 VULKAN_HPP_ASSERT( d.vkGetDisplayModeProperties2KHR &&
"Function <vkGetDisplayModeProperties2KHR> requires <VK_KHR_get_display_properties2>" );
14557 std::vector<VULKAN_HPP_NAMESPACE::DisplayModeProperties2KHR, DisplayModeProperties2KHRAllocator> properties( displayModeProperties2KHRAllocator );
14558 uint32_t propertyCount;
14563 d.vkGetDisplayModeProperties2KHR( m_physicalDevice,
static_cast<VkDisplayKHR>( display ), &propertyCount,
nullptr ) );
14564 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
14566 properties.resize( propertyCount );
14570 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
14573 if ( propertyCount < properties.size() )
14575 properties.resize( propertyCount );
14581 template <
typename Dispatch>
14583 PhysicalDevice::getDisplayPlaneCapabilities2KHR(
const VULKAN_HPP_NAMESPACE::DisplayPlaneInfo2KHR * pDisplayPlaneInfo,
14584 VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilities2KHR * pCapabilities,
14588 return static_cast<Result>( d.vkGetDisplayPlaneCapabilities2KHR( m_physicalDevice,
14593 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14594 template <
typename Dispatch>
14596 PhysicalDevice::getDisplayPlaneCapabilities2KHR(
const VULKAN_HPP_NAMESPACE::DisplayPlaneInfo2KHR & displayPlaneInfo, Dispatch
const & d )
const
14599 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14600 VULKAN_HPP_ASSERT( d.vkGetDisplayPlaneCapabilities2KHR &&
"Function <vkGetDisplayPlaneCapabilities2KHR> requires <VK_KHR_get_display_properties2>" );
14603 VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilities2KHR capabilities;
14614 #if defined( VK_USE_PLATFORM_IOS_MVK )
14617 template <
typename Dispatch>
14619 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
14620 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
14624 return static_cast<Result>( d.vkCreateIOSSurfaceMVK( m_instance,
14625 reinterpret_cast<const VkIOSSurfaceCreateInfoMVK *
>( pCreateInfo ),
14630 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14631 template <
typename Dispatch>
14633 Instance::createIOSSurfaceMVK(
const VULKAN_HPP_NAMESPACE::IOSSurfaceCreateInfoMVK & createInfo,
14634 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
14635 Dispatch
const & d )
const
14638 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14639 VULKAN_HPP_ASSERT( d.vkCreateIOSSurfaceMVK &&
"Function <vkCreateIOSSurfaceMVK> requires <VK_MVK_ios_surface>" );
14642 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
14644 d.vkCreateIOSSurfaceMVK( m_instance,
14645 reinterpret_cast<const VkIOSSurfaceCreateInfoMVK *
>( &createInfo ),
14646 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
14653 # ifndef VULKAN_HPP_NO_SMART_HANDLE
14654 template <
typename Dispatch>
14656 Instance::createIOSSurfaceMVKUnique(
const VULKAN_HPP_NAMESPACE::IOSSurfaceCreateInfoMVK & createInfo,
14657 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
14658 Dispatch
const & d )
const
14661 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14662 VULKAN_HPP_ASSERT( d.vkCreateIOSSurfaceMVK &&
"Function <vkCreateIOSSurfaceMVK> requires <VK_MVK_ios_surface>" );
14665 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
14667 d.vkCreateIOSSurfaceMVK( m_instance,
14668 reinterpret_cast<const VkIOSSurfaceCreateInfoMVK *
>( &createInfo ),
14669 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
14674 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
14680 #if defined( VK_USE_PLATFORM_MACOS_MVK )
14683 template <
typename Dispatch>
14685 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
14686 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
14690 return static_cast<Result>( d.vkCreateMacOSSurfaceMVK( m_instance,
14691 reinterpret_cast<const VkMacOSSurfaceCreateInfoMVK *
>( pCreateInfo ),
14696 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14697 template <
typename Dispatch>
14699 Instance::createMacOSSurfaceMVK(
const VULKAN_HPP_NAMESPACE::MacOSSurfaceCreateInfoMVK & createInfo,
14700 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
14701 Dispatch
const & d )
const
14704 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14705 VULKAN_HPP_ASSERT( d.vkCreateMacOSSurfaceMVK &&
"Function <vkCreateMacOSSurfaceMVK> requires <VK_MVK_macos_surface>" );
14708 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
14710 d.vkCreateMacOSSurfaceMVK( m_instance,
14711 reinterpret_cast<const VkMacOSSurfaceCreateInfoMVK *
>( &createInfo ),
14712 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
14719 # ifndef VULKAN_HPP_NO_SMART_HANDLE
14720 template <
typename Dispatch>
14722 Instance::createMacOSSurfaceMVKUnique(
const VULKAN_HPP_NAMESPACE::MacOSSurfaceCreateInfoMVK & createInfo,
14723 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
14724 Dispatch
const & d )
const
14727 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14728 VULKAN_HPP_ASSERT( d.vkCreateMacOSSurfaceMVK &&
"Function <vkCreateMacOSSurfaceMVK> requires <VK_MVK_macos_surface>" );
14731 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
14733 d.vkCreateMacOSSurfaceMVK( m_instance,
14734 reinterpret_cast<const VkMacOSSurfaceCreateInfoMVK *
>( &createInfo ),
14735 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
14740 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
14748 template <
typename Dispatch>
14756 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14757 template <
typename Dispatch>
14759 Device::setDebugUtilsObjectNameEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsObjectNameInfoEXT & nameInfo, Dispatch
const & d )
const
14762 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14763 VULKAN_HPP_ASSERT( d.vkSetDebugUtilsObjectNameEXT &&
"Function <vkSetDebugUtilsObjectNameEXT> requires <VK_EXT_debug_utils>" );
14774 template <
typename Dispatch>
14782 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14783 template <
typename Dispatch>
14785 Device::setDebugUtilsObjectTagEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsObjectTagInfoEXT & tagInfo, Dispatch
const & d )
const
14788 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14789 VULKAN_HPP_ASSERT( d.vkSetDebugUtilsObjectTagEXT &&
"Function <vkSetDebugUtilsObjectTagEXT> requires <VK_EXT_debug_utils>" );
14800 template <
typename Dispatch>
14801 VULKAN_HPP_INLINE void Queue::beginDebugUtilsLabelEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT * pLabelInfo,
14805 d.vkQueueBeginDebugUtilsLabelEXT( m_queue,
reinterpret_cast<const VkDebugUtilsLabelEXT *
>( pLabelInfo ) );
14808 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14809 template <
typename Dispatch>
14810 VULKAN_HPP_INLINE void Queue::beginDebugUtilsLabelEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT & labelInfo,
14814 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14815 VULKAN_HPP_ASSERT( d.vkQueueBeginDebugUtilsLabelEXT &&
"Function <vkQueueBeginDebugUtilsLabelEXT> requires <VK_EXT_debug_utils>" );
14818 d.vkQueueBeginDebugUtilsLabelEXT( m_queue,
reinterpret_cast<const VkDebugUtilsLabelEXT *
>( &labelInfo ) );
14822 template <
typename Dispatch>
14826 d.vkQueueEndDebugUtilsLabelEXT( m_queue );
14829 template <
typename Dispatch>
14830 VULKAN_HPP_INLINE void Queue::insertDebugUtilsLabelEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT * pLabelInfo,
14834 d.vkQueueInsertDebugUtilsLabelEXT( m_queue,
reinterpret_cast<const VkDebugUtilsLabelEXT *
>( pLabelInfo ) );
14837 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14838 template <
typename Dispatch>
14839 VULKAN_HPP_INLINE void Queue::insertDebugUtilsLabelEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT & labelInfo,
14843 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14844 VULKAN_HPP_ASSERT( d.vkQueueInsertDebugUtilsLabelEXT &&
"Function <vkQueueInsertDebugUtilsLabelEXT> requires <VK_EXT_debug_utils>" );
14847 d.vkQueueInsertDebugUtilsLabelEXT( m_queue,
reinterpret_cast<const VkDebugUtilsLabelEXT *
>( &labelInfo ) );
14851 template <
typename Dispatch>
14852 VULKAN_HPP_INLINE void CommandBuffer::beginDebugUtilsLabelEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT * pLabelInfo,
14856 d.vkCmdBeginDebugUtilsLabelEXT( m_commandBuffer,
reinterpret_cast<const VkDebugUtilsLabelEXT *
>( pLabelInfo ) );
14859 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14860 template <
typename Dispatch>
14861 VULKAN_HPP_INLINE void CommandBuffer::beginDebugUtilsLabelEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT & labelInfo,
14865 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14866 VULKAN_HPP_ASSERT( d.vkCmdBeginDebugUtilsLabelEXT &&
"Function <vkCmdBeginDebugUtilsLabelEXT> requires <VK_EXT_debug_utils>" );
14869 d.vkCmdBeginDebugUtilsLabelEXT( m_commandBuffer,
reinterpret_cast<const VkDebugUtilsLabelEXT *
>( &labelInfo ) );
14873 template <
typename Dispatch>
14877 d.vkCmdEndDebugUtilsLabelEXT( m_commandBuffer );
14880 template <
typename Dispatch>
14881 VULKAN_HPP_INLINE void CommandBuffer::insertDebugUtilsLabelEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT * pLabelInfo,
14885 d.vkCmdInsertDebugUtilsLabelEXT( m_commandBuffer,
reinterpret_cast<const VkDebugUtilsLabelEXT *
>( pLabelInfo ) );
14888 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14889 template <
typename Dispatch>
14890 VULKAN_HPP_INLINE void CommandBuffer::insertDebugUtilsLabelEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT & labelInfo,
14894 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14895 VULKAN_HPP_ASSERT( d.vkCmdInsertDebugUtilsLabelEXT &&
"Function <vkCmdInsertDebugUtilsLabelEXT> requires <VK_EXT_debug_utils>" );
14898 d.vkCmdInsertDebugUtilsLabelEXT( m_commandBuffer,
reinterpret_cast<const VkDebugUtilsLabelEXT *
>( &labelInfo ) );
14902 template <
typename Dispatch>
14904 Instance::createDebugUtilsMessengerEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCreateInfoEXT * pCreateInfo,
14905 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
14906 VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT * pMessenger,
14910 return static_cast<Result>( d.vkCreateDebugUtilsMessengerEXT( m_instance,
14916 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14917 template <
typename Dispatch>
14919 Instance::createDebugUtilsMessengerEXT(
const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCreateInfoEXT & createInfo,
14921 Dispatch
const & d )
const
14924 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14925 VULKAN_HPP_ASSERT( d.vkCreateDebugUtilsMessengerEXT &&
"Function <vkCreateDebugUtilsMessengerEXT> requires <VK_EXT_debug_utils>" );
14928 VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger;
14932 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
14939 # ifndef VULKAN_HPP_NO_SMART_HANDLE
14940 template <
typename Dispatch>
14942 Instance::createDebugUtilsMessengerEXTUnique(
const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCreateInfoEXT & createInfo,
14944 Dispatch
const & d )
const
14947 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14948 VULKAN_HPP_ASSERT( d.vkCreateDebugUtilsMessengerEXT &&
"Function <vkCreateDebugUtilsMessengerEXT> requires <VK_EXT_debug_utils>" );
14951 VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger;
14955 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
14965 template <
typename Dispatch>
14966 VULKAN_HPP_INLINE void Instance::destroyDebugUtilsMessengerEXT( VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger,
14967 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
14971 d.vkDestroyDebugUtilsMessengerEXT(
14975 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
14976 template <
typename Dispatch>
14977 VULKAN_HPP_INLINE void Instance::destroyDebugUtilsMessengerEXT( VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger,
14982 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
14983 VULKAN_HPP_ASSERT( d.vkDestroyDebugUtilsMessengerEXT &&
"Function <vkDestroyDebugUtilsMessengerEXT> requires <VK_EXT_debug_utils>" );
14986 d.vkDestroyDebugUtilsMessengerEXT(
14989 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
14993 template <
typename Dispatch>
14995 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
14999 d.vkDestroyDebugUtilsMessengerEXT(
15003 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15004 template <
typename Dispatch>
15010 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15011 VULKAN_HPP_ASSERT( d.vkDestroyDebugUtilsMessengerEXT &&
"Function <vkDestroyDebugUtilsMessengerEXT> requires <VK_EXT_debug_utils>" );
15014 d.vkDestroyDebugUtilsMessengerEXT(
15017 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
15021 template <
typename Dispatch>
15024 const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCallbackDataEXT * pCallbackData,
15028 d.vkSubmitDebugUtilsMessageEXT( m_instance,
15034 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15035 template <
typename Dispatch>
15038 const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCallbackDataEXT & callbackData,
15042 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15043 VULKAN_HPP_ASSERT( d.vkSubmitDebugUtilsMessageEXT &&
"Function <vkSubmitDebugUtilsMessageEXT> requires <VK_EXT_debug_utils>" );
15046 d.vkSubmitDebugUtilsMessageEXT( m_instance,
15053 #if defined( VK_USE_PLATFORM_ANDROID_KHR )
15056 template <
typename Dispatch>
15058 Device::getAndroidHardwareBufferPropertiesANDROID(
const struct AHardwareBuffer * buffer,
15059 VULKAN_HPP_NAMESPACE::AndroidHardwareBufferPropertiesANDROID * pProperties,
15063 return static_cast<Result>(
15064 d.vkGetAndroidHardwareBufferPropertiesANDROID( m_device, buffer,
reinterpret_cast<VkAndroidHardwareBufferPropertiesANDROID *
>( pProperties ) ) );
15067 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15068 template <
typename Dispatch>
15070 Device::getAndroidHardwareBufferPropertiesANDROID(
const struct AHardwareBuffer & buffer, Dispatch
const & d )
const
15073 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15075 "Function <vkGetAndroidHardwareBufferPropertiesANDROID> requires <VK_ANDROID_external_memory_android_hardware_buffer>" );
15078 VULKAN_HPP_NAMESPACE::AndroidHardwareBufferPropertiesANDROID properties;
15080 d.vkGetAndroidHardwareBufferPropertiesANDROID( m_device, &buffer,
reinterpret_cast<VkAndroidHardwareBufferPropertiesANDROID *
>( &properties ) ) );
15086 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
15088 Device::getAndroidHardwareBufferPropertiesANDROID(
const struct AHardwareBuffer & buffer, Dispatch
const & d )
const
15091 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15093 "Function <vkGetAndroidHardwareBufferPropertiesANDROID> requires <VK_ANDROID_external_memory_android_hardware_buffer>" );
15096 StructureChain<X, Y, Z...> structureChain;
15097 VULKAN_HPP_NAMESPACE::AndroidHardwareBufferPropertiesANDROID & properties =
15098 structureChain.template get<VULKAN_HPP_NAMESPACE::AndroidHardwareBufferPropertiesANDROID>();
15100 d.vkGetAndroidHardwareBufferPropertiesANDROID( m_device, &buffer,
reinterpret_cast<VkAndroidHardwareBufferPropertiesANDROID *
>( &properties ) ) );
15107 template <
typename Dispatch>
15109 Device::getMemoryAndroidHardwareBufferANDROID(
const VULKAN_HPP_NAMESPACE::MemoryGetAndroidHardwareBufferInfoANDROID * pInfo,
15110 struct AHardwareBuffer ** pBuffer,
15114 return static_cast<Result>(
15115 d.vkGetMemoryAndroidHardwareBufferANDROID( m_device,
reinterpret_cast<const VkMemoryGetAndroidHardwareBufferInfoANDROID *
>( pInfo ), pBuffer ) );
15118 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15119 template <
typename Dispatch>
15121 Device::getMemoryAndroidHardwareBufferANDROID(
const VULKAN_HPP_NAMESPACE::MemoryGetAndroidHardwareBufferInfoANDROID & info, Dispatch
const & d )
const
15124 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15126 "Function <vkGetMemoryAndroidHardwareBufferANDROID> requires <VK_ANDROID_external_memory_android_hardware_buffer>" );
15129 struct AHardwareBuffer * buffer;
15131 d.vkGetMemoryAndroidHardwareBufferANDROID( m_device,
reinterpret_cast<const VkMemoryGetAndroidHardwareBufferInfoANDROID *
>( &info ), &buffer ) );
15139 #if defined( VK_ENABLE_BETA_EXTENSIONS )
15142 template <
typename Dispatch>
15144 Device::createExecutionGraphPipelinesAMDX( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
15145 uint32_t createInfoCount,
15146 const VULKAN_HPP_NAMESPACE::ExecutionGraphPipelineCreateInfoAMDX * pCreateInfos,
15147 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
15148 VULKAN_HPP_NAMESPACE::Pipeline * pPipelines,
15152 return static_cast<Result>( d.vkCreateExecutionGraphPipelinesAMDX( m_device,
15155 reinterpret_cast<const VkExecutionGraphPipelineCreateInfoAMDX *
>( pCreateInfos ),
15157 reinterpret_cast<VkPipeline *
>( pPipelines ) ) );
15160 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15161 template <
typename PipelineAllocator,
typename Dispatch>
15163 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
15164 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ExecutionGraphPipelineCreateInfoAMDX>
const & createInfos,
15166 Dispatch
const & d )
const
15169 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15170 VULKAN_HPP_ASSERT( d.vkCreateExecutionGraphPipelinesAMDX &&
"Function <vkCreateExecutionGraphPipelinesAMDX> requires <VK_AMDX_shader_enqueue>" );
15173 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
15177 createInfos.size(),
15178 reinterpret_cast<const VkExecutionGraphPipelineCreateInfoAMDX *
>( createInfos.data() ),
15179 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
15180 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
15183 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
15188 template <
typename PipelineAllocator,
15190 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, VULKAN_HPP_NAMESPACE::Pipeline>::value,
int>::type>
15192 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
15193 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ExecutionGraphPipelineCreateInfoAMDX>
const & createInfos,
15195 PipelineAllocator & pipelineAllocator,
15196 Dispatch
const & d )
const
15199 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15200 VULKAN_HPP_ASSERT( d.vkCreateExecutionGraphPipelinesAMDX &&
"Function <vkCreateExecutionGraphPipelinesAMDX> requires <VK_AMDX_shader_enqueue>" );
15203 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
15207 createInfos.size(),
15208 reinterpret_cast<const VkExecutionGraphPipelineCreateInfoAMDX *
>( createInfos.data() ),
15209 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
15210 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
15213 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
15218 template <
typename Dispatch>
15220 Device::createExecutionGraphPipelineAMDX( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
15221 const VULKAN_HPP_NAMESPACE::ExecutionGraphPipelineCreateInfoAMDX & createInfo,
15223 Dispatch
const & d )
const
15226 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15227 VULKAN_HPP_ASSERT( d.vkCreateExecutionGraphPipelinesAMDX &&
"Function <vkCreateExecutionGraphPipelinesAMDX> requires <VK_AMDX_shader_enqueue>" );
15230 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
15235 reinterpret_cast<const VkExecutionGraphPipelineCreateInfoAMDX *
>( &createInfo ),
15236 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
15237 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
15240 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
15245 # ifndef VULKAN_HPP_NO_SMART_HANDLE
15246 template <
typename Dispatch,
typename PipelineAllocator>
15248 Device::createExecutionGraphPipelinesAMDXUnique(
15249 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
15250 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ExecutionGraphPipelineCreateInfoAMDX>
const & createInfos,
15252 Dispatch
const & d )
const
15255 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15256 VULKAN_HPP_ASSERT( d.vkCreateExecutionGraphPipelinesAMDX &&
"Function <vkCreateExecutionGraphPipelinesAMDX> requires <VK_AMDX_shader_enqueue>" );
15259 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
15263 createInfos.size(),
15264 reinterpret_cast<const VkExecutionGraphPipelineCreateInfoAMDX *
>( createInfos.data() ),
15265 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
15266 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
15269 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
15270 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines;
15271 uniquePipelines.reserve( createInfos.size() );
15273 for (
auto const & pipeline : pipelines )
15283 typename PipelineAllocator,
15284 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>>::value,
int>::type>
15286 Device::createExecutionGraphPipelinesAMDXUnique(
15287 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
15288 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ExecutionGraphPipelineCreateInfoAMDX>
const & createInfos,
15290 PipelineAllocator & pipelineAllocator,
15291 Dispatch
const & d )
const
15294 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15295 VULKAN_HPP_ASSERT( d.vkCreateExecutionGraphPipelinesAMDX &&
"Function <vkCreateExecutionGraphPipelinesAMDX> requires <VK_AMDX_shader_enqueue>" );
15298 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
15302 createInfos.size(),
15303 reinterpret_cast<const VkExecutionGraphPipelineCreateInfoAMDX *
>( createInfos.data() ),
15304 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
15305 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
15308 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
15309 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines( pipelineAllocator );
15310 uniquePipelines.reserve( createInfos.size() );
15312 for (
auto const & pipeline : pipelines )
15320 template <
typename Dispatch>
15322 Device::createExecutionGraphPipelineAMDXUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
15323 const VULKAN_HPP_NAMESPACE::ExecutionGraphPipelineCreateInfoAMDX & createInfo,
15325 Dispatch
const & d )
const
15328 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15329 VULKAN_HPP_ASSERT( d.vkCreateExecutionGraphPipelinesAMDX &&
"Function <vkCreateExecutionGraphPipelinesAMDX> requires <VK_AMDX_shader_enqueue>" );
15332 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
15337 reinterpret_cast<const VkExecutionGraphPipelineCreateInfoAMDX *
>( &createInfo ),
15338 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
15339 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
15342 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
15351 template <
typename Dispatch>
15353 Device::getExecutionGraphPipelineScratchSizeAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
15354 VULKAN_HPP_NAMESPACE::ExecutionGraphPipelineScratchSizeAMDX * pSizeInfo,
15358 return static_cast<Result>( d.vkGetExecutionGraphPipelineScratchSizeAMDX(
15359 m_device,
static_cast<VkPipeline>( executionGraph ),
reinterpret_cast<VkExecutionGraphPipelineScratchSizeAMDX *
>( pSizeInfo ) ) );
15362 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15363 template <
typename Dispatch>
15365 Device::getExecutionGraphPipelineScratchSizeAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph, Dispatch
const & d )
const
15368 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15370 "Function <vkGetExecutionGraphPipelineScratchSizeAMDX> requires <VK_AMDX_shader_enqueue>" );
15373 VULKAN_HPP_NAMESPACE::ExecutionGraphPipelineScratchSizeAMDX sizeInfo;
15375 m_device,
static_cast<VkPipeline>( executionGraph ),
reinterpret_cast<VkExecutionGraphPipelineScratchSizeAMDX *
>( &sizeInfo ) ) );
15382 template <
typename Dispatch>
15384 Device::getExecutionGraphPipelineNodeIndexAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
15385 const VULKAN_HPP_NAMESPACE::PipelineShaderStageNodeCreateInfoAMDX * pNodeInfo,
15386 uint32_t * pNodeIndex,
15390 return static_cast<Result>( d.vkGetExecutionGraphPipelineNodeIndexAMDX(
15391 m_device,
static_cast<VkPipeline>( executionGraph ),
reinterpret_cast<const VkPipelineShaderStageNodeCreateInfoAMDX *
>( pNodeInfo ), pNodeIndex ) );
15394 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15395 template <
typename Dispatch>
15397 VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
const VULKAN_HPP_NAMESPACE::PipelineShaderStageNodeCreateInfoAMDX & nodeInfo, Dispatch
const & d )
const
15400 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15401 VULKAN_HPP_ASSERT( d.vkGetExecutionGraphPipelineNodeIndexAMDX &&
"Function <vkGetExecutionGraphPipelineNodeIndexAMDX> requires <VK_AMDX_shader_enqueue>" );
15404 uint32_t nodeIndex;
15406 m_device,
static_cast<VkPipeline>( executionGraph ),
reinterpret_cast<const VkPipelineShaderStageNodeCreateInfoAMDX *
>( &nodeInfo ), &nodeIndex ) );
15413 template <
typename Dispatch>
15418 d.vkCmdInitializeGraphScratchMemoryAMDX( m_commandBuffer,
static_cast<VkDeviceAddress>( scratch ) );
15421 template <
typename Dispatch>
15423 const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
15427 d.vkCmdDispatchGraphAMDX( m_commandBuffer,
static_cast<VkDeviceAddress>( scratch ),
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *
>( pCountInfo ) );
15430 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15431 template <
typename Dispatch>
15433 const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
15437 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15438 VULKAN_HPP_ASSERT( d.vkCmdDispatchGraphAMDX &&
"Function <vkCmdDispatchGraphAMDX> requires <VK_AMDX_shader_enqueue>" );
15441 d.vkCmdDispatchGraphAMDX( m_commandBuffer,
static_cast<VkDeviceAddress>( scratch ),
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *
>( &countInfo ) );
15445 template <
typename Dispatch>
15447 const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
15451 d.vkCmdDispatchGraphIndirectAMDX(
15452 m_commandBuffer,
static_cast<VkDeviceAddress>( scratch ),
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *
>( pCountInfo ) );
15455 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15456 template <
typename Dispatch>
15458 const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
15462 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15463 VULKAN_HPP_ASSERT( d.vkCmdDispatchGraphIndirectAMDX &&
"Function <vkCmdDispatchGraphIndirectAMDX> requires <VK_AMDX_shader_enqueue>" );
15466 d.vkCmdDispatchGraphIndirectAMDX(
15467 m_commandBuffer,
static_cast<VkDeviceAddress>( scratch ),
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *
>( &countInfo ) );
15471 template <
typename Dispatch>
15483 template <
typename Dispatch>
15484 VULKAN_HPP_INLINE void CommandBuffer::setSampleLocationsEXT(
const VULKAN_HPP_NAMESPACE::SampleLocationsInfoEXT * pSampleLocationsInfo,
15488 d.vkCmdSetSampleLocationsEXT( m_commandBuffer,
reinterpret_cast<const VkSampleLocationsInfoEXT *
>( pSampleLocationsInfo ) );
15491 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15492 template <
typename Dispatch>
15493 VULKAN_HPP_INLINE void CommandBuffer::setSampleLocationsEXT(
const VULKAN_HPP_NAMESPACE::SampleLocationsInfoEXT & sampleLocationsInfo,
15497 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15498 VULKAN_HPP_ASSERT( d.vkCmdSetSampleLocationsEXT &&
"Function <vkCmdSetSampleLocationsEXT> requires <VK_EXT_sample_locations>" );
15501 d.vkCmdSetSampleLocationsEXT( m_commandBuffer,
reinterpret_cast<const VkSampleLocationsInfoEXT *
>( &sampleLocationsInfo ) );
15505 template <
typename Dispatch>
15507 VULKAN_HPP_NAMESPACE::MultisamplePropertiesEXT * pMultisampleProperties,
15511 d.vkGetPhysicalDeviceMultisamplePropertiesEXT(
15515 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15516 template <
typename Dispatch>
15521 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15523 "Function <vkGetPhysicalDeviceMultisamplePropertiesEXT> requires <VK_EXT_sample_locations>" );
15526 VULKAN_HPP_NAMESPACE::MultisamplePropertiesEXT multisampleProperties;
15527 d.vkGetPhysicalDeviceMultisamplePropertiesEXT(
15530 return multisampleProperties;
15536 template <
typename Dispatch>
15537 VULKAN_HPP_INLINE void Device::getImageMemoryRequirements2KHR(
const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 * pInfo,
15538 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
15542 d.vkGetImageMemoryRequirements2KHR(
15546 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15547 template <
typename Dispatch>
15549 Device::getImageMemoryRequirements2KHR(
const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
15552 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15554 "Function <vkGetImageMemoryRequirements2KHR> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
15557 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
15558 d.vkGetImageMemoryRequirements2KHR(
15561 return memoryRequirements;
15564 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
15566 Device::getImageMemoryRequirements2KHR(
const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
15569 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15571 "Function <vkGetImageMemoryRequirements2KHR> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
15574 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
15575 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
15576 d.vkGetImageMemoryRequirements2KHR(
15579 return structureChain;
15583 template <
typename Dispatch>
15584 VULKAN_HPP_INLINE void Device::getBufferMemoryRequirements2KHR(
const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 * pInfo,
15585 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
15589 d.vkGetBufferMemoryRequirements2KHR(
15593 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15594 template <
typename Dispatch>
15596 Device::getBufferMemoryRequirements2KHR(
const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
15599 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15601 "Function <vkGetBufferMemoryRequirements2KHR> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
15604 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
15605 d.vkGetBufferMemoryRequirements2KHR(
15608 return memoryRequirements;
15611 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
15613 Device::getBufferMemoryRequirements2KHR(
const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
15616 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15618 "Function <vkGetBufferMemoryRequirements2KHR> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
15621 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
15622 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
15623 d.vkGetBufferMemoryRequirements2KHR(
15626 return structureChain;
15630 template <
typename Dispatch>
15631 VULKAN_HPP_INLINE void Device::getImageSparseMemoryRequirements2KHR(
const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 * pInfo,
15632 uint32_t * pSparseMemoryRequirementCount,
15633 VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2 * pSparseMemoryRequirements,
15637 d.vkGetImageSparseMemoryRequirements2KHR( m_device,
15639 pSparseMemoryRequirementCount,
15643 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15644 template <
typename SparseImageMemoryRequirements2Allocator,
typename Dispatch>
15646 Device::getImageSparseMemoryRequirements2KHR(
const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 & info, Dispatch
const & d )
const
15649 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15651 "Function <vkGetImageSparseMemoryRequirements2KHR> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
15654 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements;
15655 uint32_t sparseMemoryRequirementCount;
15656 d.vkGetImageSparseMemoryRequirements2KHR(
15658 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
15659 d.vkGetImageSparseMemoryRequirements2KHR( m_device,
15661 &sparseMemoryRequirementCount,
15664 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
15665 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
15667 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
15669 return sparseMemoryRequirements;
15672 template <
typename SparseImageMemoryRequirements2Allocator,
15674 typename std::enable_if<
15675 std::is_same<typename SparseImageMemoryRequirements2Allocator::value_type, VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2>::value,
15678 Device::getImageSparseMemoryRequirements2KHR(
const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 & info,
15679 SparseImageMemoryRequirements2Allocator & sparseImageMemoryRequirements2Allocator,
15680 Dispatch
const & d )
const
15683 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15685 "Function <vkGetImageSparseMemoryRequirements2KHR> requires <VK_KHR_get_memory_requirements2> or <VK_VERSION_1_1>" );
15688 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements(
15689 sparseImageMemoryRequirements2Allocator );
15690 uint32_t sparseMemoryRequirementCount;
15691 d.vkGetImageSparseMemoryRequirements2KHR(
15693 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
15694 d.vkGetImageSparseMemoryRequirements2KHR( m_device,
15696 &sparseMemoryRequirementCount,
15699 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
15700 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
15702 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
15704 return sparseMemoryRequirements;
15710 template <
typename Dispatch>
15712 Device::createAccelerationStructureKHR(
const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoKHR * pCreateInfo,
15713 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
15714 VULKAN_HPP_NAMESPACE::AccelerationStructureKHR * pAccelerationStructure,
15718 return static_cast<Result>( d.vkCreateAccelerationStructureKHR( m_device,
15724 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15725 template <
typename Dispatch>
15727 Device::createAccelerationStructureKHR(
const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoKHR & createInfo,
15729 Dispatch
const & d )
const
15732 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15733 VULKAN_HPP_ASSERT( d.vkCreateAccelerationStructureKHR &&
"Function <vkCreateAccelerationStructureKHR> requires <VK_KHR_acceleration_structure>" );
15736 VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure;
15740 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
15747 # ifndef VULKAN_HPP_NO_SMART_HANDLE
15748 template <
typename Dispatch>
15750 Device::createAccelerationStructureKHRUnique(
const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoKHR & createInfo,
15752 Dispatch
const & d )
const
15755 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15756 VULKAN_HPP_ASSERT( d.vkCreateAccelerationStructureKHR &&
"Function <vkCreateAccelerationStructureKHR> requires <VK_KHR_acceleration_structure>" );
15759 VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure;
15763 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
15774 template <
typename Dispatch>
15775 VULKAN_HPP_INLINE void Device::destroyAccelerationStructureKHR( VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure,
15776 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
15780 d.vkDestroyAccelerationStructureKHR(
15784 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15785 template <
typename Dispatch>
15786 VULKAN_HPP_INLINE void Device::destroyAccelerationStructureKHR( VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure,
15791 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15792 VULKAN_HPP_ASSERT( d.vkDestroyAccelerationStructureKHR &&
"Function <vkDestroyAccelerationStructureKHR> requires <VK_KHR_acceleration_structure>" );
15795 d.vkDestroyAccelerationStructureKHR(
15798 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
15802 template <
typename Dispatch>
15803 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure,
15804 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
15808 d.vkDestroyAccelerationStructureKHR(
15812 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15813 template <
typename Dispatch>
15814 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure,
15819 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15820 VULKAN_HPP_ASSERT( d.vkDestroyAccelerationStructureKHR &&
"Function <vkDestroyAccelerationStructureKHR> requires <VK_KHR_acceleration_structure>" );
15823 d.vkDestroyAccelerationStructureKHR(
15826 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
15830 template <
typename Dispatch>
15832 CommandBuffer::buildAccelerationStructuresKHR( uint32_t infoCount,
15833 const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR * pInfos,
15834 const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildRangeInfoKHR *
const * ppBuildRangeInfos,
15838 d.vkCmdBuildAccelerationStructuresKHR( m_commandBuffer,
15844 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15845 template <
typename Dispatch>
15847 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR>
const & infos,
15848 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildRangeInfoKHR * const>
const & pBuildRangeInfos,
15852 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15853 VULKAN_HPP_ASSERT( d.vkCmdBuildAccelerationStructuresKHR &&
"Function <vkCmdBuildAccelerationStructuresKHR> requires <VK_KHR_acceleration_structure>" );
15855 # ifdef VULKAN_HPP_NO_EXCEPTIONS
15858 if ( infos.size() != pBuildRangeInfos.size() )
15864 d.vkCmdBuildAccelerationStructuresKHR( m_commandBuffer,
15871 template <
typename Dispatch>
15873 const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR * pInfos,
15875 const uint32_t * pIndirectStrides,
15876 const uint32_t *
const * ppMaxPrimitiveCounts,
15880 d.vkCmdBuildAccelerationStructuresIndirectKHR( m_commandBuffer,
15883 reinterpret_cast<const VkDeviceAddress *
>( pIndirectDeviceAddresses ),
15885 ppMaxPrimitiveCounts );
15888 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15889 template <
typename Dispatch>
15891 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR>
const & infos,
15892 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceAddress>
const & indirectDeviceAddresses,
15893 VULKAN_HPP_NAMESPACE::ArrayProxy<const uint32_t>
const & indirectStrides,
15894 VULKAN_HPP_NAMESPACE::ArrayProxy<const uint32_t * const>
const & pMaxPrimitiveCounts,
15898 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15900 "Function <vkCmdBuildAccelerationStructuresIndirectKHR> requires <VK_KHR_acceleration_structure>" );
15902 # ifdef VULKAN_HPP_NO_EXCEPTIONS
15907 if ( infos.size() != indirectDeviceAddresses.size() )
15911 if ( infos.size() != indirectStrides.size() )
15915 if ( infos.size() != pMaxPrimitiveCounts.size() )
15921 d.vkCmdBuildAccelerationStructuresIndirectKHR( m_commandBuffer,
15924 reinterpret_cast<const VkDeviceAddress *
>( indirectDeviceAddresses.data() ),
15925 indirectStrides.data(),
15926 pMaxPrimitiveCounts.data() );
15930 template <
typename Dispatch>
15932 Device::buildAccelerationStructuresKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
15933 uint32_t infoCount,
15934 const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR * pInfos,
15935 const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildRangeInfoKHR *
const * ppBuildRangeInfos,
15939 return static_cast<Result>(
15940 d.vkBuildAccelerationStructuresKHR( m_device,
15947 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15948 template <
typename Dispatch>
15950 VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
15951 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR>
const & infos,
15952 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildRangeInfoKHR * const>
const & pBuildRangeInfos,
15953 Dispatch
const & d )
const
15956 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
15957 VULKAN_HPP_ASSERT( d.vkBuildAccelerationStructuresKHR &&
"Function <vkBuildAccelerationStructuresKHR> requires <VK_KHR_acceleration_structure>" );
15959 # ifdef VULKAN_HPP_NO_EXCEPTIONS
15962 if ( infos.size() != pBuildRangeInfos.size() )
15969 d.vkBuildAccelerationStructuresKHR( m_device,
15977 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR } );
15983 template <
typename Dispatch>
15985 const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureInfoKHR * pInfo,
15989 return static_cast<Result>( d.vkCopyAccelerationStructureKHR(
15993 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
15994 template <
typename Dispatch>
15996 Device::copyAccelerationStructureKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
15997 const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureInfoKHR & info,
15998 Dispatch
const & d )
const
16001 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16002 VULKAN_HPP_ASSERT( d.vkCopyAccelerationStructureKHR &&
"Function <vkCopyAccelerationStructureKHR> requires <VK_KHR_acceleration_structure>" );
16010 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR } );
16016 template <
typename Dispatch>
16018 Device::copyAccelerationStructureToMemoryKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16019 const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureToMemoryInfoKHR * pInfo,
16023 return static_cast<Result>( d.vkCopyAccelerationStructureToMemoryKHR(
16027 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16028 template <
typename Dispatch>
16030 Device::copyAccelerationStructureToMemoryKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16031 const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureToMemoryInfoKHR & info,
16032 Dispatch
const & d )
const
16035 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16037 "Function <vkCopyAccelerationStructureToMemoryKHR> requires <VK_KHR_acceleration_structure>" );
16045 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR } );
16051 template <
typename Dispatch>
16053 Device::copyMemoryToAccelerationStructureKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16054 const VULKAN_HPP_NAMESPACE::CopyMemoryToAccelerationStructureInfoKHR * pInfo,
16058 return static_cast<Result>( d.vkCopyMemoryToAccelerationStructureKHR(
16062 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16063 template <
typename Dispatch>
16065 Device::copyMemoryToAccelerationStructureKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16066 const VULKAN_HPP_NAMESPACE::CopyMemoryToAccelerationStructureInfoKHR & info,
16067 Dispatch
const & d )
const
16070 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16072 "Function <vkCopyMemoryToAccelerationStructureKHR> requires <VK_KHR_acceleration_structure>" );
16080 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR } );
16086 template <
typename Dispatch>
16088 Device::writeAccelerationStructuresPropertiesKHR( uint32_t accelerationStructureCount,
16089 const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR * pAccelerationStructures,
16097 return static_cast<Result>( d.vkWriteAccelerationStructuresPropertiesKHR( m_device,
16098 accelerationStructureCount,
16106 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16107 template <
typename DataType,
typename DataTypeAllocator,
typename Dispatch>
16109 Device::writeAccelerationStructuresPropertiesKHR(
16110 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR>
const & accelerationStructures,
16114 Dispatch
const & d )
const
16117 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16119 "Function <vkWriteAccelerationStructuresPropertiesKHR> requires <VK_KHR_acceleration_structure>" );
16123 std::vector<DataType, DataTypeAllocator> data( dataSize /
sizeof( DataType ) );
16125 d.vkWriteAccelerationStructuresPropertiesKHR( m_device,
16126 accelerationStructures.size(),
16129 data.size() *
sizeof( DataType ),
16130 reinterpret_cast<void *
>( data.data() ),
16137 template <
typename DataType,
typename Dispatch>
16139 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR>
const & accelerationStructures,
16142 Dispatch
const & d )
const
16145 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16147 "Function <vkWriteAccelerationStructuresPropertiesKHR> requires <VK_KHR_acceleration_structure>" );
16152 d.vkWriteAccelerationStructuresPropertiesKHR( m_device,
16153 accelerationStructures.size(),
16156 sizeof( DataType ),
16157 reinterpret_cast<void *
>( &data ),
16165 template <
typename Dispatch>
16166 VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureKHR(
const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureInfoKHR * pInfo,
16173 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16174 template <
typename Dispatch>
16175 VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureKHR(
const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureInfoKHR & info,
16179 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16180 VULKAN_HPP_ASSERT( d.vkCmdCopyAccelerationStructureKHR &&
"Function <vkCmdCopyAccelerationStructureKHR> requires <VK_KHR_acceleration_structure>" );
16187 template <
typename Dispatch>
16188 VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureToMemoryKHR(
const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureToMemoryInfoKHR * pInfo,
16195 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16196 template <
typename Dispatch>
16197 VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureToMemoryKHR(
const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureToMemoryInfoKHR & info,
16201 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16203 "Function <vkCmdCopyAccelerationStructureToMemoryKHR> requires <VK_KHR_acceleration_structure>" );
16210 template <
typename Dispatch>
16211 VULKAN_HPP_INLINE void CommandBuffer::copyMemoryToAccelerationStructureKHR(
const VULKAN_HPP_NAMESPACE::CopyMemoryToAccelerationStructureInfoKHR * pInfo,
16218 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16219 template <
typename Dispatch>
16220 VULKAN_HPP_INLINE void CommandBuffer::copyMemoryToAccelerationStructureKHR(
const VULKAN_HPP_NAMESPACE::CopyMemoryToAccelerationStructureInfoKHR & info,
16224 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16226 "Function <vkCmdCopyMemoryToAccelerationStructureKHR> requires <VK_KHR_acceleration_structure>" );
16233 template <
typename Dispatch>
16242 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16243 template <
typename Dispatch>
16245 Device::getAccelerationStructureAddressKHR(
const VULKAN_HPP_NAMESPACE::AccelerationStructureDeviceAddressInfoKHR & info,
16249 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16251 "Function <vkGetAccelerationStructureDeviceAddressKHR> requires <VK_KHR_acceleration_structure>" );
16261 template <
typename Dispatch>
16263 CommandBuffer::writeAccelerationStructuresPropertiesKHR( uint32_t accelerationStructureCount,
16264 const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR * pAccelerationStructures,
16266 VULKAN_HPP_NAMESPACE::QueryPool queryPool,
16267 uint32_t firstQuery,
16271 d.vkCmdWriteAccelerationStructuresPropertiesKHR( m_commandBuffer,
16272 accelerationStructureCount,
16279 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16280 template <
typename Dispatch>
16282 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR>
const & accelerationStructures,
16284 VULKAN_HPP_NAMESPACE::QueryPool queryPool,
16285 uint32_t firstQuery,
16289 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16291 "Function <vkCmdWriteAccelerationStructuresPropertiesKHR> requires <VK_KHR_acceleration_structure>" );
16294 d.vkCmdWriteAccelerationStructuresPropertiesKHR( m_commandBuffer,
16295 accelerationStructures.size(),
16303 template <
typename Dispatch>
16304 VULKAN_HPP_INLINE void Device::getAccelerationStructureCompatibilityKHR(
const VULKAN_HPP_NAMESPACE::AccelerationStructureVersionInfoKHR * pVersionInfo,
16309 d.vkGetDeviceAccelerationStructureCompatibilityKHR( m_device,
16314 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16315 template <
typename Dispatch>
16317 Device::getAccelerationStructureCompatibilityKHR(
const VULKAN_HPP_NAMESPACE::AccelerationStructureVersionInfoKHR & versionInfo,
16321 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16323 "Function <vkGetDeviceAccelerationStructureCompatibilityKHR> requires <VK_KHR_acceleration_structure>" );
16327 d.vkGetDeviceAccelerationStructureCompatibilityKHR( m_device,
16331 return compatibility;
16335 template <
typename Dispatch>
16337 const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR * pBuildInfo,
16338 const uint32_t * pMaxPrimitiveCounts,
16339 VULKAN_HPP_NAMESPACE::AccelerationStructureBuildSizesInfoKHR * pSizeInfo,
16343 d.vkGetAccelerationStructureBuildSizesKHR( m_device,
16346 pMaxPrimitiveCounts,
16350 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16351 template <
typename Dispatch>
16354 const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR & buildInfo,
16355 VULKAN_HPP_NAMESPACE::ArrayProxy<const uint32_t>
const & maxPrimitiveCounts,
16359 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16361 "Function <vkGetAccelerationStructureBuildSizesKHR> requires <VK_KHR_acceleration_structure>" );
16363 # ifdef VULKAN_HPP_NO_EXCEPTIONS
16366 if ( maxPrimitiveCounts.size() != buildInfo.geometryCount )
16372 VULKAN_HPP_NAMESPACE::AccelerationStructureBuildSizesInfoKHR sizeInfo;
16373 d.vkGetAccelerationStructureBuildSizesKHR( m_device,
16376 maxPrimitiveCounts.data(),
16385 template <
typename Dispatch>
16386 VULKAN_HPP_INLINE void CommandBuffer::traceRaysKHR(
const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR * pRaygenShaderBindingTable,
16387 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR * pMissShaderBindingTable,
16388 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR * pHitShaderBindingTable,
16389 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR * pCallableShaderBindingTable,
16396 d.vkCmdTraceRaysKHR( m_commandBuffer,
16406 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16407 template <
typename Dispatch>
16408 VULKAN_HPP_INLINE void CommandBuffer::traceRaysKHR(
const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR & raygenShaderBindingTable,
16409 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR & missShaderBindingTable,
16410 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR & hitShaderBindingTable,
16411 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR & callableShaderBindingTable,
16418 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16419 VULKAN_HPP_ASSERT( d.vkCmdTraceRaysKHR &&
"Function <vkCmdTraceRaysKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16422 d.vkCmdTraceRaysKHR( m_commandBuffer,
16433 template <
typename Dispatch>
16435 Device::createRayTracingPipelinesKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16436 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
16437 uint32_t createInfoCount,
16438 const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR * pCreateInfos,
16439 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
16440 VULKAN_HPP_NAMESPACE::Pipeline * pPipelines,
16444 return static_cast<Result>( d.vkCreateRayTracingPipelinesKHR( m_device,
16450 reinterpret_cast<VkPipeline *
>( pPipelines ) ) );
16453 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16454 template <
typename PipelineAllocator,
typename Dispatch>
16456 Device::createRayTracingPipelinesKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16457 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
16458 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR>
const & createInfos,
16460 Dispatch
const & d )
const
16463 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16464 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesKHR &&
"Function <vkCreateRayTracingPipelinesKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16467 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
16472 createInfos.size(),
16474 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
16475 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
16478 { VULKAN_HPP_NAMESPACE::Result::eSuccess,
16479 VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR,
16480 VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR,
16481 VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
16486 template <
typename PipelineAllocator,
16488 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, VULKAN_HPP_NAMESPACE::Pipeline>::value,
int>::type>
16490 Device::createRayTracingPipelinesKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16491 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
16492 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR>
const & createInfos,
16494 PipelineAllocator & pipelineAllocator,
16495 Dispatch
const & d )
const
16498 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16499 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesKHR &&
"Function <vkCreateRayTracingPipelinesKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16502 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
16507 createInfos.size(),
16509 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
16510 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
16513 { VULKAN_HPP_NAMESPACE::Result::eSuccess,
16514 VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR,
16515 VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR,
16516 VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
16521 template <
typename Dispatch>
16523 Device::createRayTracingPipelineKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16524 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
16525 const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR & createInfo,
16527 Dispatch
const & d )
const
16530 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16531 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesKHR &&
"Function <vkCreateRayTracingPipelinesKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16534 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
16541 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
16542 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
16545 { VULKAN_HPP_NAMESPACE::Result::eSuccess,
16546 VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR,
16547 VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR,
16548 VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
16553 # ifndef VULKAN_HPP_NO_SMART_HANDLE
16554 template <
typename Dispatch,
typename PipelineAllocator>
16556 Device::createRayTracingPipelinesKHRUnique(
16557 VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16558 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
16559 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR>
const & createInfos,
16561 Dispatch
const & d )
const
16564 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16565 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesKHR &&
"Function <vkCreateRayTracingPipelinesKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16568 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
16573 createInfos.size(),
16575 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
16576 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
16579 { VULKAN_HPP_NAMESPACE::Result::eSuccess,
16580 VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR,
16581 VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR,
16582 VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
16583 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines;
16584 uniquePipelines.reserve( createInfos.size() );
16586 for (
auto const & pipeline : pipelines )
16596 typename PipelineAllocator,
16597 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>>::value,
int>::type>
16599 Device::createRayTracingPipelinesKHRUnique(
16600 VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16601 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
16602 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR>
const & createInfos,
16604 PipelineAllocator & pipelineAllocator,
16605 Dispatch
const & d )
const
16608 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16609 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesKHR &&
"Function <vkCreateRayTracingPipelinesKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16612 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
16617 createInfos.size(),
16619 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
16620 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
16623 { VULKAN_HPP_NAMESPACE::Result::eSuccess,
16624 VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR,
16625 VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR,
16626 VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
16627 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines( pipelineAllocator );
16628 uniquePipelines.reserve( createInfos.size() );
16630 for (
auto const & pipeline : pipelines )
16638 template <
typename Dispatch>
16640 Device::createRayTracingPipelineKHRUnique( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
16641 VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
16642 const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR & createInfo,
16644 Dispatch
const & d )
const
16647 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16648 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesKHR &&
"Function <vkCreateRayTracingPipelinesKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16651 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
16658 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
16659 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
16662 { VULKAN_HPP_NAMESPACE::Result::eSuccess,
16663 VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR,
16664 VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR,
16665 VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
16674 template <
typename Dispatch>
16676 uint32_t firstGroup,
16677 uint32_t groupCount,
16683 return static_cast<Result>(
16684 d.vkGetRayTracingShaderGroupHandlesKHR( m_device,
static_cast<VkPipeline>( pipeline ), firstGroup, groupCount, dataSize, pData ) );
16687 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16688 template <
typename DataType,
typename DataTypeAllocator,
typename Dispatch>
16690 VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount,
size_t dataSize, Dispatch
const & d )
const
16693 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16695 "Function <vkGetRayTracingShaderGroupHandlesKHR> requires <VK_KHR_ray_tracing_pipeline> or <VK_NV_ray_tracing>" );
16699 std::vector<DataType, DataTypeAllocator> data( dataSize /
sizeof( DataType ) );
16701 m_device,
static_cast<VkPipeline>( pipeline ), firstGroup, groupCount, data.size() *
sizeof( DataType ),
reinterpret_cast<void *
>( data.data() ) ) );
16707 template <
typename DataType,
typename Dispatch>
16709 Device::getRayTracingShaderGroupHandleKHR( VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount, Dispatch
const & d )
const
16712 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16714 "Function <vkGetRayTracingShaderGroupHandlesKHR> requires <VK_KHR_ray_tracing_pipeline> or <VK_NV_ray_tracing>" );
16719 m_device,
static_cast<VkPipeline>( pipeline ), firstGroup, groupCount,
sizeof( DataType ),
reinterpret_cast<void *
>( &data ) ) );
16726 template <
typename Dispatch>
16728 uint32_t firstGroup,
16729 uint32_t groupCount,
16735 return static_cast<Result>(
16736 d.vkGetRayTracingCaptureReplayShaderGroupHandlesKHR( m_device,
static_cast<VkPipeline>( pipeline ), firstGroup, groupCount, dataSize, pData ) );
16739 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16740 template <
typename DataType,
typename DataTypeAllocator,
typename Dispatch>
16742 Device::getRayTracingCaptureReplayShaderGroupHandlesKHR(
16743 VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount,
size_t dataSize, Dispatch
const & d )
const
16746 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16748 "Function <vkGetRayTracingCaptureReplayShaderGroupHandlesKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16752 std::vector<DataType, DataTypeAllocator> data( dataSize /
sizeof( DataType ) );
16754 m_device,
static_cast<VkPipeline>( pipeline ), firstGroup, groupCount, data.size() *
sizeof( DataType ),
reinterpret_cast<void *
>( data.data() ) ) );
16760 template <
typename DataType,
typename Dispatch>
16762 VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount, Dispatch
const & d )
const
16765 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16767 "Function <vkGetRayTracingCaptureReplayShaderGroupHandlesKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16772 m_device,
static_cast<VkPipeline>( pipeline ), firstGroup, groupCount,
sizeof( DataType ),
reinterpret_cast<void *
>( &data ) ) );
16779 template <
typename Dispatch>
16780 VULKAN_HPP_INLINE void CommandBuffer::traceRaysIndirectKHR(
const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR * pRaygenShaderBindingTable,
16781 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR * pMissShaderBindingTable,
16782 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR * pHitShaderBindingTable,
16783 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR * pCallableShaderBindingTable,
16788 d.vkCmdTraceRaysIndirectKHR( m_commandBuffer,
16796 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16797 template <
typename Dispatch>
16798 VULKAN_HPP_INLINE void CommandBuffer::traceRaysIndirectKHR(
const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR & raygenShaderBindingTable,
16799 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR & missShaderBindingTable,
16800 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR & hitShaderBindingTable,
16801 const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR & callableShaderBindingTable,
16806 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16807 VULKAN_HPP_ASSERT( d.vkCmdTraceRaysIndirectKHR &&
"Function <vkCmdTraceRaysIndirectKHR> requires <VK_KHR_ray_tracing_pipeline>" );
16810 d.vkCmdTraceRaysIndirectKHR( m_commandBuffer,
16819 template <
typename Dispatch>
16830 template <
typename Dispatch>
16834 d.vkCmdSetRayTracingPipelineStackSizeKHR( m_commandBuffer, pipelineStackSize );
16839 template <
typename Dispatch>
16841 Device::createSamplerYcbcrConversionKHR(
const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo * pCreateInfo,
16842 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
16843 VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion * pYcbcrConversion,
16847 return static_cast<Result>( d.vkCreateSamplerYcbcrConversionKHR( m_device,
16853 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16854 template <
typename Dispatch>
16856 Device::createSamplerYcbcrConversionKHR(
const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo & createInfo,
16858 Dispatch
const & d )
const
16861 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16863 "Function <vkCreateSamplerYcbcrConversionKHR> requires <VK_KHR_sampler_ycbcr_conversion> or <VK_VERSION_1_1>" );
16866 VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion;
16870 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
16877 # ifndef VULKAN_HPP_NO_SMART_HANDLE
16878 template <
typename Dispatch>
16880 Device::createSamplerYcbcrConversionKHRUnique(
const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo & createInfo,
16882 Dispatch
const & d )
const
16885 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16887 "Function <vkCreateSamplerYcbcrConversionKHR> requires <VK_KHR_sampler_ycbcr_conversion> or <VK_VERSION_1_1>" );
16890 VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion;
16894 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
16904 template <
typename Dispatch>
16905 VULKAN_HPP_INLINE void Device::destroySamplerYcbcrConversionKHR( VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion,
16906 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
16910 d.vkDestroySamplerYcbcrConversionKHR(
16914 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16915 template <
typename Dispatch>
16916 VULKAN_HPP_INLINE void Device::destroySamplerYcbcrConversionKHR( VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion,
16921 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16923 "Function <vkDestroySamplerYcbcrConversionKHR> requires <VK_KHR_sampler_ycbcr_conversion> or <VK_VERSION_1_1>" );
16926 d.vkDestroySamplerYcbcrConversionKHR(
16929 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
16935 template <
typename Dispatch>
16937 const VULKAN_HPP_NAMESPACE::BindBufferMemoryInfo * pBindInfos,
16941 return static_cast<Result>( d.vkBindBufferMemory2KHR( m_device, bindInfoCount,
reinterpret_cast<const VkBindBufferMemoryInfo *
>( pBindInfos ) ) );
16944 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16945 template <
typename Dispatch>
16947 Device::bindBufferMemory2KHR( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BindBufferMemoryInfo>
const & bindInfos,
16948 Dispatch
const & d )
const
16951 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16952 VULKAN_HPP_ASSERT( d.vkBindBufferMemory2KHR &&
"Function <vkBindBufferMemory2KHR> requires <VK_KHR_bind_memory2> or <VK_VERSION_1_1>" );
16956 d.vkBindBufferMemory2KHR( m_device, bindInfos.size(),
reinterpret_cast<const VkBindBufferMemoryInfo *
>( bindInfos.data() ) ) );
16963 template <
typename Dispatch>
16965 const VULKAN_HPP_NAMESPACE::BindImageMemoryInfo * pBindInfos,
16969 return static_cast<Result>( d.vkBindImageMemory2KHR( m_device, bindInfoCount,
reinterpret_cast<const VkBindImageMemoryInfo *
>( pBindInfos ) ) );
16972 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
16973 template <
typename Dispatch>
16975 Device::bindImageMemory2KHR( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BindImageMemoryInfo>
const & bindInfos, Dispatch
const & d )
const
16978 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
16979 VULKAN_HPP_ASSERT( d.vkBindImageMemory2KHR &&
"Function <vkBindImageMemory2KHR> requires <VK_KHR_bind_memory2> or <VK_VERSION_1_1>" );
16983 d.vkBindImageMemory2KHR( m_device, bindInfos.size(),
reinterpret_cast<const VkBindImageMemoryInfo *
>( bindInfos.data() ) ) );
16992 template <
typename Dispatch>
16994 VULKAN_HPP_NAMESPACE::Image image, VULKAN_HPP_NAMESPACE::ImageDrmFormatModifierPropertiesEXT * pProperties, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
16997 return static_cast<Result>( d.vkGetImageDrmFormatModifierPropertiesEXT(
17001 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17002 template <
typename Dispatch>
17004 Device::getImageDrmFormatModifierPropertiesEXT( VULKAN_HPP_NAMESPACE::Image image, Dispatch
const & d )
const
17007 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17009 "Function <vkGetImageDrmFormatModifierPropertiesEXT> requires <VK_EXT_image_drm_format_modifier>" );
17012 VULKAN_HPP_NAMESPACE::ImageDrmFormatModifierPropertiesEXT properties;
17023 template <
typename Dispatch>
17025 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
17026 VULKAN_HPP_NAMESPACE::ValidationCacheEXT * pValidationCache,
17030 return static_cast<Result>( d.vkCreateValidationCacheEXT( m_device,
17036 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17037 template <
typename Dispatch>
17039 Device::createValidationCacheEXT(
const VULKAN_HPP_NAMESPACE::ValidationCacheCreateInfoEXT & createInfo,
17041 Dispatch
const & d )
const
17044 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17045 VULKAN_HPP_ASSERT( d.vkCreateValidationCacheEXT &&
"Function <vkCreateValidationCacheEXT> requires <VK_EXT_validation_cache>" );
17048 VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache;
17052 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17059 # ifndef VULKAN_HPP_NO_SMART_HANDLE
17060 template <
typename Dispatch>
17062 Device::createValidationCacheEXTUnique(
const VULKAN_HPP_NAMESPACE::ValidationCacheCreateInfoEXT & createInfo,
17064 Dispatch
const & d )
const
17067 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17068 VULKAN_HPP_ASSERT( d.vkCreateValidationCacheEXT &&
"Function <vkCreateValidationCacheEXT> requires <VK_EXT_validation_cache>" );
17071 VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache;
17075 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17085 template <
typename Dispatch>
17086 VULKAN_HPP_INLINE void Device::destroyValidationCacheEXT( VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache,
17087 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
17091 d.vkDestroyValidationCacheEXT(
17095 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17096 template <
typename Dispatch>
17097 VULKAN_HPP_INLINE void Device::destroyValidationCacheEXT( VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache,
17102 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17103 VULKAN_HPP_ASSERT( d.vkDestroyValidationCacheEXT &&
"Function <vkDestroyValidationCacheEXT> requires <VK_EXT_validation_cache>" );
17106 d.vkDestroyValidationCacheEXT(
17109 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
17113 template <
typename Dispatch>
17115 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
17119 d.vkDestroyValidationCacheEXT(
17123 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17124 template <
typename Dispatch>
17130 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17131 VULKAN_HPP_ASSERT( d.vkDestroyValidationCacheEXT &&
"Function <vkDestroyValidationCacheEXT> requires <VK_EXT_validation_cache>" );
17134 d.vkDestroyValidationCacheEXT(
17137 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
17141 template <
typename Dispatch>
17143 uint32_t srcCacheCount,
17144 const VULKAN_HPP_NAMESPACE::ValidationCacheEXT * pSrcCaches,
17148 return static_cast<Result>( d.vkMergeValidationCachesEXT(
17152 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17153 template <
typename Dispatch>
17155 Device::mergeValidationCachesEXT( VULKAN_HPP_NAMESPACE::ValidationCacheEXT dstCache,
17156 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ValidationCacheEXT>
const & srcCaches,
17157 Dispatch
const & d )
const
17160 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17161 VULKAN_HPP_ASSERT( d.vkMergeValidationCachesEXT &&
"Function <vkMergeValidationCachesEXT> requires <VK_EXT_validation_cache>" );
17172 template <
typename Dispatch>
17174 size_t * pDataSize,
17179 return static_cast<Result>( d.vkGetValidationCacheDataEXT( m_device,
static_cast<VkValidationCacheEXT>( validationCache ), pDataSize, pData ) );
17182 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17183 template <
typename U
int8_tAllocator,
typename Dispatch>
17185 Device::getValidationCacheDataEXT( VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache, Dispatch
const & d )
const
17188 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17189 VULKAN_HPP_ASSERT( d.vkGetValidationCacheDataEXT &&
"Function <vkGetValidationCacheDataEXT> requires <VK_EXT_validation_cache>" );
17192 std::vector<uint8_t, Uint8_tAllocator> data;
17198 d.vkGetValidationCacheDataEXT( m_device,
static_cast<VkValidationCacheEXT>( validationCache ), &dataSize,
nullptr ) );
17199 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
17201 data.resize( dataSize );
17203 d.vkGetValidationCacheDataEXT( m_device,
static_cast<VkValidationCacheEXT>( validationCache ), &dataSize,
reinterpret_cast<void *
>( data.data() ) ) );
17205 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
17208 if ( dataSize < data.size() )
17210 data.resize( dataSize );
17215 template <
typename Uint8_tAllocator,
17217 typename std::enable_if<std::is_same<typename Uint8_tAllocator::value_type, uint8_t>::value,
int>::type>
17219 Device::getValidationCacheDataEXT( VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache, Uint8_tAllocator & uint8_tAllocator, Dispatch
const & d )
const
17222 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17223 VULKAN_HPP_ASSERT( d.vkGetValidationCacheDataEXT &&
"Function <vkGetValidationCacheDataEXT> requires <VK_EXT_validation_cache>" );
17226 std::vector<uint8_t, Uint8_tAllocator> data( uint8_tAllocator );
17232 d.vkGetValidationCacheDataEXT( m_device,
static_cast<VkValidationCacheEXT>( validationCache ), &dataSize,
nullptr ) );
17233 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
17235 data.resize( dataSize );
17237 d.vkGetValidationCacheDataEXT( m_device,
static_cast<VkValidationCacheEXT>( validationCache ), &dataSize,
reinterpret_cast<void *
>( data.data() ) ) );
17239 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
17242 if ( dataSize < data.size() )
17244 data.resize( dataSize );
17252 template <
typename Dispatch>
17253 VULKAN_HPP_INLINE void CommandBuffer::bindShadingRateImageNV( VULKAN_HPP_NAMESPACE::ImageView imageView,
17258 d.vkCmdBindShadingRateImageNV( m_commandBuffer,
static_cast<VkImageView>( imageView ),
static_cast<VkImageLayout>( imageLayout ) );
17261 template <
typename Dispatch>
17263 uint32_t viewportCount,
17264 const VULKAN_HPP_NAMESPACE::ShadingRatePaletteNV * pShadingRatePalettes,
17268 d.vkCmdSetViewportShadingRatePaletteNV(
17269 m_commandBuffer, firstViewport, viewportCount,
reinterpret_cast<const VkShadingRatePaletteNV *
>( pShadingRatePalettes ) );
17272 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17273 template <
typename Dispatch>
17275 uint32_t firstViewport,
17276 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ShadingRatePaletteNV>
const & shadingRatePalettes,
17280 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17281 VULKAN_HPP_ASSERT( d.vkCmdSetViewportShadingRatePaletteNV &&
"Function <vkCmdSetViewportShadingRatePaletteNV> requires <VK_NV_shading_rate_image>" );
17284 d.vkCmdSetViewportShadingRatePaletteNV(
17285 m_commandBuffer, firstViewport, shadingRatePalettes.size(),
reinterpret_cast<const VkShadingRatePaletteNV *
>( shadingRatePalettes.data() ) );
17289 template <
typename Dispatch>
17291 uint32_t customSampleOrderCount,
17292 const VULKAN_HPP_NAMESPACE::CoarseSampleOrderCustomNV * pCustomSampleOrders,
17296 d.vkCmdSetCoarseSampleOrderNV( m_commandBuffer,
17298 customSampleOrderCount,
17302 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17303 template <
typename Dispatch>
17306 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::CoarseSampleOrderCustomNV>
const & customSampleOrders,
17310 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17311 VULKAN_HPP_ASSERT( d.vkCmdSetCoarseSampleOrderNV &&
"Function <vkCmdSetCoarseSampleOrderNV> requires <VK_NV_shading_rate_image>" );
17314 d.vkCmdSetCoarseSampleOrderNV( m_commandBuffer,
17316 customSampleOrders.size(),
17323 template <
typename Dispatch>
17325 Device::createAccelerationStructureNV(
const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoNV * pCreateInfo,
17326 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
17327 VULKAN_HPP_NAMESPACE::AccelerationStructureNV * pAccelerationStructure,
17331 return static_cast<Result>( d.vkCreateAccelerationStructureNV( m_device,
17337 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17338 template <
typename Dispatch>
17340 Device::createAccelerationStructureNV(
const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoNV & createInfo,
17342 Dispatch
const & d )
const
17345 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17346 VULKAN_HPP_ASSERT( d.vkCreateAccelerationStructureNV &&
"Function <vkCreateAccelerationStructureNV> requires <VK_NV_ray_tracing>" );
17349 VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure;
17353 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17360 # ifndef VULKAN_HPP_NO_SMART_HANDLE
17361 template <
typename Dispatch>
17363 Device::createAccelerationStructureNVUnique(
const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoNV & createInfo,
17365 Dispatch
const & d )
const
17368 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17369 VULKAN_HPP_ASSERT( d.vkCreateAccelerationStructureNV &&
"Function <vkCreateAccelerationStructureNV> requires <VK_NV_ray_tracing>" );
17372 VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure;
17376 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17387 template <
typename Dispatch>
17388 VULKAN_HPP_INLINE void Device::destroyAccelerationStructureNV( VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure,
17389 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
17393 d.vkDestroyAccelerationStructureNV(
17397 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17398 template <
typename Dispatch>
17399 VULKAN_HPP_INLINE void Device::destroyAccelerationStructureNV( VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure,
17404 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17405 VULKAN_HPP_ASSERT( d.vkDestroyAccelerationStructureNV &&
"Function <vkDestroyAccelerationStructureNV> requires <VK_NV_ray_tracing>" );
17408 d.vkDestroyAccelerationStructureNV(
17411 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
17415 template <
typename Dispatch>
17416 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure,
17417 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
17421 d.vkDestroyAccelerationStructureNV(
17425 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17426 template <
typename Dispatch>
17427 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure,
17432 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17433 VULKAN_HPP_ASSERT( d.vkDestroyAccelerationStructureNV &&
"Function <vkDestroyAccelerationStructureNV> requires <VK_NV_ray_tracing>" );
17436 d.vkDestroyAccelerationStructureNV(
17439 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
17443 template <
typename Dispatch>
17445 Device::getAccelerationStructureMemoryRequirementsNV(
const VULKAN_HPP_NAMESPACE::AccelerationStructureMemoryRequirementsInfoNV * pInfo,
17450 d.vkGetAccelerationStructureMemoryRequirementsNV( m_device,
17455 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17456 template <
typename Dispatch>
17458 Device::getAccelerationStructureMemoryRequirementsNV(
const VULKAN_HPP_NAMESPACE::AccelerationStructureMemoryRequirementsInfoNV & info,
17462 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17464 "Function <vkGetAccelerationStructureMemoryRequirementsNV> requires <VK_NV_ray_tracing>" );
17468 d.vkGetAccelerationStructureMemoryRequirementsNV( m_device,
17472 return memoryRequirements;
17475 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
17477 Device::getAccelerationStructureMemoryRequirementsNV(
const VULKAN_HPP_NAMESPACE::AccelerationStructureMemoryRequirementsInfoNV & info,
17481 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17483 "Function <vkGetAccelerationStructureMemoryRequirementsNV> requires <VK_NV_ray_tracing>" );
17486 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
17488 d.vkGetAccelerationStructureMemoryRequirementsNV( m_device,
17492 return structureChain;
17496 template <
typename Dispatch>
17498 uint32_t bindInfoCount,
const VULKAN_HPP_NAMESPACE::BindAccelerationStructureMemoryInfoNV * pBindInfos, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
17501 return static_cast<Result>(
17505 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17506 template <
typename Dispatch>
17508 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::BindAccelerationStructureMemoryInfoNV>
const & bindInfos, Dispatch
const & d )
const
17511 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17512 VULKAN_HPP_ASSERT( d.vkBindAccelerationStructureMemoryNV &&
"Function <vkBindAccelerationStructureMemoryNV> requires <VK_NV_ray_tracing>" );
17523 template <
typename Dispatch>
17524 VULKAN_HPP_INLINE void CommandBuffer::buildAccelerationStructureNV(
const VULKAN_HPP_NAMESPACE::AccelerationStructureInfoNV * pInfo,
17525 VULKAN_HPP_NAMESPACE::Buffer instanceData,
17528 VULKAN_HPP_NAMESPACE::AccelerationStructureNV dst,
17529 VULKAN_HPP_NAMESPACE::AccelerationStructureNV src,
17530 VULKAN_HPP_NAMESPACE::Buffer scratch,
17535 d.vkCmdBuildAccelerationStructureNV( m_commandBuffer,
17537 static_cast<VkBuffer>( instanceData ),
17546 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17547 template <
typename Dispatch>
17548 VULKAN_HPP_INLINE void CommandBuffer::buildAccelerationStructureNV(
const VULKAN_HPP_NAMESPACE::AccelerationStructureInfoNV & info,
17549 VULKAN_HPP_NAMESPACE::Buffer instanceData,
17552 VULKAN_HPP_NAMESPACE::AccelerationStructureNV dst,
17553 VULKAN_HPP_NAMESPACE::AccelerationStructureNV src,
17554 VULKAN_HPP_NAMESPACE::Buffer scratch,
17559 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17560 VULKAN_HPP_ASSERT( d.vkCmdBuildAccelerationStructureNV &&
"Function <vkCmdBuildAccelerationStructureNV> requires <VK_NV_ray_tracing>" );
17563 d.vkCmdBuildAccelerationStructureNV( m_commandBuffer,
17565 static_cast<VkBuffer>( instanceData ),
17575 template <
typename Dispatch>
17576 VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureNV( VULKAN_HPP_NAMESPACE::AccelerationStructureNV dst,
17577 VULKAN_HPP_NAMESPACE::AccelerationStructureNV src,
17582 d.vkCmdCopyAccelerationStructureNV( m_commandBuffer,
17588 template <
typename Dispatch>
17589 VULKAN_HPP_INLINE void CommandBuffer::traceRaysNV( VULKAN_HPP_NAMESPACE::Buffer raygenShaderBindingTableBuffer,
17591 VULKAN_HPP_NAMESPACE::Buffer missShaderBindingTableBuffer,
17594 VULKAN_HPP_NAMESPACE::Buffer hitShaderBindingTableBuffer,
17597 VULKAN_HPP_NAMESPACE::Buffer callableShaderBindingTableBuffer,
17606 d.vkCmdTraceRaysNV( m_commandBuffer,
17607 static_cast<VkBuffer>( raygenShaderBindingTableBuffer ),
17608 static_cast<VkDeviceSize>( raygenShaderBindingOffset ),
17609 static_cast<VkBuffer>( missShaderBindingTableBuffer ),
17612 static_cast<VkBuffer>( hitShaderBindingTableBuffer ),
17615 static_cast<VkBuffer>( callableShaderBindingTableBuffer ),
17616 static_cast<VkDeviceSize>( callableShaderBindingOffset ),
17617 static_cast<VkDeviceSize>( callableShaderBindingStride ),
17623 template <
typename Dispatch>
17625 uint32_t createInfoCount,
17626 const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV * pCreateInfos,
17627 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
17628 VULKAN_HPP_NAMESPACE::Pipeline * pPipelines,
17632 return static_cast<Result>( d.vkCreateRayTracingPipelinesNV( m_device,
17637 reinterpret_cast<VkPipeline *
>( pPipelines ) ) );
17640 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17641 template <
typename PipelineAllocator,
typename Dispatch>
17643 Device::createRayTracingPipelinesNV( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
17644 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV>
const & createInfos,
17646 Dispatch
const & d )
const
17649 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17650 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesNV &&
"Function <vkCreateRayTracingPipelinesNV> requires <VK_NV_ray_tracing>" );
17653 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
17657 createInfos.size(),
17659 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17660 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
17663 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
17668 template <
typename PipelineAllocator,
17670 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, VULKAN_HPP_NAMESPACE::Pipeline>::value,
int>::type>
17672 Device::createRayTracingPipelinesNV( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
17673 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV>
const & createInfos,
17675 PipelineAllocator & pipelineAllocator,
17676 Dispatch
const & d )
const
17679 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17680 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesNV &&
"Function <vkCreateRayTracingPipelinesNV> requires <VK_NV_ray_tracing>" );
17683 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
17687 createInfos.size(),
17689 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17690 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
17693 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
17698 template <
typename Dispatch>
17700 Device::createRayTracingPipelineNV( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
17701 const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV & createInfo,
17703 Dispatch
const & d )
const
17706 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17707 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesNV &&
"Function <vkCreateRayTracingPipelinesNV> requires <VK_NV_ray_tracing>" );
17710 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
17716 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17717 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
17720 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
17725 # ifndef VULKAN_HPP_NO_SMART_HANDLE
17726 template <
typename Dispatch,
typename PipelineAllocator>
17728 Device::createRayTracingPipelinesNVUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
17729 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV>
const & createInfos,
17731 Dispatch
const & d )
const
17734 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17735 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesNV &&
"Function <vkCreateRayTracingPipelinesNV> requires <VK_NV_ray_tracing>" );
17738 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
17742 createInfos.size(),
17744 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17745 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
17748 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
17749 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines;
17750 uniquePipelines.reserve( createInfos.size() );
17752 for (
auto const & pipeline : pipelines )
17762 typename PipelineAllocator,
17763 typename std::enable_if<std::is_same<typename PipelineAllocator::value_type, UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>>::value,
int>::type>
17765 Device::createRayTracingPipelinesNVUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
17766 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV>
const & createInfos,
17768 PipelineAllocator & pipelineAllocator,
17769 Dispatch
const & d )
const
17772 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17773 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesNV &&
"Function <vkCreateRayTracingPipelinesNV> requires <VK_NV_ray_tracing>" );
17776 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
17780 createInfos.size(),
17782 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17783 reinterpret_cast<VkPipeline *
>( pipelines.data() ) ) );
17786 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
17787 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::Pipeline, Dispatch>, PipelineAllocator> uniquePipelines( pipelineAllocator );
17788 uniquePipelines.reserve( createInfos.size() );
17790 for (
auto const & pipeline : pipelines )
17798 template <
typename Dispatch>
17800 Device::createRayTracingPipelineNVUnique( VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache,
17801 const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV & createInfo,
17803 Dispatch
const & d )
const
17806 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17807 VULKAN_HPP_ASSERT( d.vkCreateRayTracingPipelinesNV &&
"Function <vkCreateRayTracingPipelinesNV> requires <VK_NV_ray_tracing>" );
17810 VULKAN_HPP_NAMESPACE::Pipeline pipeline;
17816 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
17817 reinterpret_cast<VkPipeline *
>( &pipeline ) ) );
17820 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT } );
17829 template <
typename Dispatch>
17831 uint32_t firstGroup,
17832 uint32_t groupCount,
17838 return static_cast<Result>(
17839 d.vkGetRayTracingShaderGroupHandlesNV( m_device,
static_cast<VkPipeline>( pipeline ), firstGroup, groupCount, dataSize, pData ) );
17842 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17843 template <
typename DataType,
typename DataTypeAllocator,
typename Dispatch>
17845 VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount,
size_t dataSize, Dispatch
const & d )
const
17848 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17850 "Function <vkGetRayTracingShaderGroupHandlesNV> requires <VK_KHR_ray_tracing_pipeline> or <VK_NV_ray_tracing>" );
17854 std::vector<DataType, DataTypeAllocator> data( dataSize /
sizeof( DataType ) );
17856 m_device,
static_cast<VkPipeline>( pipeline ), firstGroup, groupCount, data.size() *
sizeof( DataType ),
reinterpret_cast<void *
>( data.data() ) ) );
17862 template <
typename DataType,
typename Dispatch>
17864 Device::getRayTracingShaderGroupHandleNV( VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount, Dispatch
const & d )
const
17867 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17869 "Function <vkGetRayTracingShaderGroupHandlesNV> requires <VK_KHR_ray_tracing_pipeline> or <VK_NV_ray_tracing>" );
17874 m_device,
static_cast<VkPipeline>( pipeline ), firstGroup, groupCount,
sizeof( DataType ),
reinterpret_cast<void *
>( &data ) ) );
17881 template <
typename Dispatch>
17888 return static_cast<Result>(
17889 d.vkGetAccelerationStructureHandleNV( m_device,
static_cast<VkAccelerationStructureNV>( accelerationStructure ), dataSize, pData ) );
17892 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17893 template <
typename DataType,
typename DataTypeAllocator,
typename Dispatch>
17895 Device::getAccelerationStructureHandleNV( VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure,
size_t dataSize, Dispatch
const & d )
const
17898 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17899 VULKAN_HPP_ASSERT( d.vkGetAccelerationStructureHandleNV &&
"Function <vkGetAccelerationStructureHandleNV> requires <VK_NV_ray_tracing>" );
17903 std::vector<DataType, DataTypeAllocator> data( dataSize /
sizeof( DataType ) );
17905 m_device,
static_cast<VkAccelerationStructureNV>( accelerationStructure ), data.size() *
sizeof( DataType ),
reinterpret_cast<void *
>( data.data() ) ) );
17911 template <
typename DataType,
typename Dispatch>
17913 Device::getAccelerationStructureHandleNV( VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure, Dispatch
const & d )
const
17916 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17917 VULKAN_HPP_ASSERT( d.vkGetAccelerationStructureHandleNV &&
"Function <vkGetAccelerationStructureHandleNV> requires <VK_NV_ray_tracing>" );
17922 m_device,
static_cast<VkAccelerationStructureNV>( accelerationStructure ),
sizeof( DataType ),
reinterpret_cast<void *
>( &data ) ) );
17929 template <
typename Dispatch>
17930 VULKAN_HPP_INLINE void CommandBuffer::writeAccelerationStructuresPropertiesNV( uint32_t accelerationStructureCount,
17931 const VULKAN_HPP_NAMESPACE::AccelerationStructureNV * pAccelerationStructures,
17933 VULKAN_HPP_NAMESPACE::QueryPool queryPool,
17934 uint32_t firstQuery,
17938 d.vkCmdWriteAccelerationStructuresPropertiesNV( m_commandBuffer,
17939 accelerationStructureCount,
17946 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
17947 template <
typename Dispatch>
17949 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureNV>
const & accelerationStructures,
17951 VULKAN_HPP_NAMESPACE::QueryPool queryPool,
17952 uint32_t firstQuery,
17956 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17958 "Function <vkCmdWriteAccelerationStructuresPropertiesNV> requires <VK_NV_ray_tracing>" );
17961 d.vkCmdWriteAccelerationStructuresPropertiesNV( m_commandBuffer,
17962 accelerationStructures.size(),
17970 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
17971 template <
typename Dispatch>
17977 return static_cast<Result>( d.vkCompileDeferredNV( m_device,
static_cast<VkPipeline>( pipeline ), shader ) );
17980 template <
typename Dispatch>
17982 Device::compileDeferredNV( VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t shader, Dispatch
const & d )
const
17985 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
17986 VULKAN_HPP_ASSERT( d.vkCompileDeferredNV &&
"Function <vkCompileDeferredNV> requires <VK_NV_ray_tracing>" );
17999 template <
typename Dispatch>
18000 VULKAN_HPP_INLINE void Device::getDescriptorSetLayoutSupportKHR(
const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo * pCreateInfo,
18001 VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport * pSupport,
18005 d.vkGetDescriptorSetLayoutSupportKHR(
18009 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18010 template <
typename Dispatch>
18012 Device::getDescriptorSetLayoutSupportKHR(
const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo & createInfo,
18016 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18018 "Function <vkGetDescriptorSetLayoutSupportKHR> requires <VK_KHR_maintenance3> or <VK_VERSION_1_1>" );
18021 VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport support;
18022 d.vkGetDescriptorSetLayoutSupportKHR(
18028 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
18030 Device::getDescriptorSetLayoutSupportKHR(
const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo & createInfo,
18034 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18036 "Function <vkGetDescriptorSetLayoutSupportKHR> requires <VK_KHR_maintenance3> or <VK_VERSION_1_1>" );
18039 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
18040 VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport & support = structureChain.template get<VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport>();
18041 d.vkGetDescriptorSetLayoutSupportKHR(
18044 return structureChain;
18050 template <
typename Dispatch>
18053 VULKAN_HPP_NAMESPACE::Buffer countBuffer,
18055 uint32_t maxDrawCount,
18060 d.vkCmdDrawIndirectCountKHR( m_commandBuffer,
18063 static_cast<VkBuffer>( countBuffer ),
18069 template <
typename Dispatch>
18072 VULKAN_HPP_NAMESPACE::Buffer countBuffer,
18074 uint32_t maxDrawCount,
18079 d.vkCmdDrawIndexedIndirectCountKHR( m_commandBuffer,
18082 static_cast<VkBuffer>( countBuffer ),
18090 template <
typename Dispatch>
18093 const void * pHostPointer,
18094 VULKAN_HPP_NAMESPACE::MemoryHostPointerPropertiesEXT * pMemoryHostPointerProperties,
18098 return static_cast<Result>( d.vkGetMemoryHostPointerPropertiesEXT( m_device,
18104 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18105 template <
typename Dispatch>
18108 const void * pHostPointer,
18109 Dispatch
const & d )
const
18112 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18113 VULKAN_HPP_ASSERT( d.vkGetMemoryHostPointerPropertiesEXT &&
"Function <vkGetMemoryHostPointerPropertiesEXT> requires <VK_EXT_external_memory_host>" );
18116 VULKAN_HPP_NAMESPACE::MemoryHostPointerPropertiesEXT memoryHostPointerProperties;
18118 d.vkGetMemoryHostPointerPropertiesEXT( m_device,
18130 template <
typename Dispatch>
18132 VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
18138 d.vkCmdWriteBufferMarkerAMD( m_commandBuffer,
18140 static_cast<VkBuffer>( dstBuffer ),
18147 template <
typename Dispatch>
18153 return static_cast<Result>(
18154 d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT( m_physicalDevice, pTimeDomainCount,
reinterpret_cast<VkTimeDomainKHR *
>( pTimeDomains ) ) );
18157 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18158 template <
typename TimeDomainKHRAllocator,
typename Dispatch>
18160 PhysicalDevice::getCalibrateableTimeDomainsEXT( Dispatch
const & d )
const
18163 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18165 "Function <vkGetPhysicalDeviceCalibrateableTimeDomainsEXT> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
18168 std::vector<VULKAN_HPP_NAMESPACE::TimeDomainKHR, TimeDomainKHRAllocator> timeDomains;
18169 uint32_t timeDomainCount;
18173 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT( m_physicalDevice, &timeDomainCount,
nullptr ) );
18174 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && timeDomainCount )
18176 timeDomains.resize( timeDomainCount );
18178 d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT( m_physicalDevice, &timeDomainCount,
reinterpret_cast<VkTimeDomainKHR *
>( timeDomains.data() ) ) );
18180 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
18183 if ( timeDomainCount < timeDomains.size() )
18185 timeDomains.resize( timeDomainCount );
18190 template <
typename TimeDomainKHRAllocator,
18192 typename std::enable_if<std::is_same<typename TimeDomainKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::TimeDomainKHR>::value,
int>::type>
18194 PhysicalDevice::getCalibrateableTimeDomainsEXT( TimeDomainKHRAllocator & timeDomainKHRAllocator, Dispatch
const & d )
const
18197 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18199 "Function <vkGetPhysicalDeviceCalibrateableTimeDomainsEXT> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
18202 std::vector<VULKAN_HPP_NAMESPACE::TimeDomainKHR, TimeDomainKHRAllocator> timeDomains( timeDomainKHRAllocator );
18203 uint32_t timeDomainCount;
18207 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT( m_physicalDevice, &timeDomainCount,
nullptr ) );
18208 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && timeDomainCount )
18210 timeDomains.resize( timeDomainCount );
18212 d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT( m_physicalDevice, &timeDomainCount,
reinterpret_cast<VkTimeDomainKHR *
>( timeDomains.data() ) ) );
18214 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
18217 if ( timeDomainCount < timeDomains.size() )
18219 timeDomains.resize( timeDomainCount );
18225 template <
typename Dispatch>
18227 const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoKHR * pTimestampInfos,
18228 uint64_t * pTimestamps,
18229 uint64_t * pMaxDeviation,
18233 return static_cast<Result>( d.vkGetCalibratedTimestampsEXT(
18237 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18238 template <
typename U
int64_tAllocator,
typename Dispatch>
18240 Device::getCalibratedTimestampsEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoKHR>
const & timestampInfos,
18241 Dispatch
const & d )
const
18244 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18246 "Function <vkGetCalibratedTimestampsEXT> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
18249 std::pair<std::vector<uint64_t, Uint64_tAllocator>, uint64_t> data_(
18250 std::piecewise_construct, std::forward_as_tuple( timestampInfos.size() ), std::forward_as_tuple( 0 ) );
18251 std::vector<uint64_t, Uint64_tAllocator> & timestamps = data_.first;
18252 uint64_t & maxDeviation = data_.second;
18254 m_device, timestampInfos.size(),
reinterpret_cast<const VkCalibratedTimestampInfoKHR *
>( timestampInfos.data() ), timestamps.data(), &maxDeviation ) );
18260 template <
typename Uint64_tAllocator,
18262 typename std::enable_if<std::is_same<typename Uint64_tAllocator::value_type, uint64_t>::value,
int>::type>
18264 Device::getCalibratedTimestampsEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoKHR>
const & timestampInfos,
18265 Uint64_tAllocator & uint64_tAllocator,
18266 Dispatch
const & d )
const
18269 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18271 "Function <vkGetCalibratedTimestampsEXT> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
18274 std::pair<std::vector<uint64_t, Uint64_tAllocator>, uint64_t> data_(
18275 std::piecewise_construct, std::forward_as_tuple( timestampInfos.size(), uint64_tAllocator ), std::forward_as_tuple( 0 ) );
18276 std::vector<uint64_t, Uint64_tAllocator> & timestamps = data_.first;
18277 uint64_t & maxDeviation = data_.second;
18279 m_device, timestampInfos.size(),
reinterpret_cast<const VkCalibratedTimestampInfoKHR *
>( timestampInfos.data() ), timestamps.data(), &maxDeviation ) );
18285 template <
typename Dispatch>
18287 Device::getCalibratedTimestampEXT(
const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoKHR & timestampInfo, Dispatch
const & d )
const
18290 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18292 "Function <vkGetCalibratedTimestampsEXT> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
18295 std::pair<uint64_t, uint64_t> data_;
18296 uint64_t & timestamp = data_.first;
18297 uint64_t & maxDeviation = data_.second;
18299 d.vkGetCalibratedTimestampsEXT( m_device, 1,
reinterpret_cast<const VkCalibratedTimestampInfoKHR *
>( ×tampInfo ), ×tamp, &maxDeviation ) );
18308 template <
typename Dispatch>
18312 d.vkCmdDrawMeshTasksNV( m_commandBuffer, taskCount, firstTask );
18315 template <
typename Dispatch>
18318 uint32_t drawCount,
18323 d.vkCmdDrawMeshTasksIndirectNV( m_commandBuffer,
static_cast<VkBuffer>( buffer ),
static_cast<VkDeviceSize>( offset ), drawCount, stride );
18326 template <
typename Dispatch>
18327 VULKAN_HPP_INLINE void CommandBuffer::drawMeshTasksIndirectCountNV( VULKAN_HPP_NAMESPACE::Buffer buffer,
18329 VULKAN_HPP_NAMESPACE::Buffer countBuffer,
18331 uint32_t maxDrawCount,
18336 d.vkCmdDrawMeshTasksIndirectCountNV( m_commandBuffer,
18339 static_cast<VkBuffer>( countBuffer ),
18347 template <
typename Dispatch>
18349 uint32_t exclusiveScissorCount,
18354 d.vkCmdSetExclusiveScissorEnableNV(
18355 m_commandBuffer, firstExclusiveScissor, exclusiveScissorCount,
reinterpret_cast<const VkBool32 *
>( pExclusiveScissorEnables ) );
18358 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18359 template <
typename Dispatch>
18361 CommandBuffer::setExclusiveScissorEnableNV( uint32_t firstExclusiveScissor,
18362 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Bool32>
const & exclusiveScissorEnables,
18366 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18367 VULKAN_HPP_ASSERT( d.vkCmdSetExclusiveScissorEnableNV &&
"Function <vkCmdSetExclusiveScissorEnableNV> requires <VK_NV_scissor_exclusive>" );
18370 d.vkCmdSetExclusiveScissorEnableNV(
18371 m_commandBuffer, firstExclusiveScissor, exclusiveScissorEnables.size(),
reinterpret_cast<const VkBool32 *
>( exclusiveScissorEnables.data() ) );
18375 template <
typename Dispatch>
18377 uint32_t exclusiveScissorCount,
18378 const VULKAN_HPP_NAMESPACE::Rect2D * pExclusiveScissors,
18382 d.vkCmdSetExclusiveScissorNV( m_commandBuffer, firstExclusiveScissor, exclusiveScissorCount,
reinterpret_cast<const VkRect2D *
>( pExclusiveScissors ) );
18385 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18386 template <
typename Dispatch>
18388 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D>
const & exclusiveScissors,
18392 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18393 VULKAN_HPP_ASSERT( d.vkCmdSetExclusiveScissorNV &&
"Function <vkCmdSetExclusiveScissorNV> requires <VK_NV_scissor_exclusive>" );
18396 d.vkCmdSetExclusiveScissorNV(
18397 m_commandBuffer, firstExclusiveScissor, exclusiveScissors.size(),
reinterpret_cast<const VkRect2D *
>( exclusiveScissors.data() ) );
18403 template <
typename Dispatch>
18407 d.vkCmdSetCheckpointNV( m_commandBuffer, pCheckpointMarker );
18410 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18411 template <
typename Checkpo
intMarkerType,
typename Dispatch>
18415 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18416 VULKAN_HPP_ASSERT( d.vkCmdSetCheckpointNV &&
"Function <vkCmdSetCheckpointNV> requires <VK_NV_device_diagnostic_checkpoints>" );
18419 d.vkCmdSetCheckpointNV( m_commandBuffer,
reinterpret_cast<const void *
>( &checkpointMarker ) );
18423 template <
typename Dispatch>
18425 VULKAN_HPP_NAMESPACE::CheckpointDataNV * pCheckpointData,
18429 d.vkGetQueueCheckpointDataNV( m_queue, pCheckpointDataCount,
reinterpret_cast<VkCheckpointDataNV *
>( pCheckpointData ) );
18432 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18433 template <
typename Checkpo
intDataNVAllocator,
typename Dispatch>
18435 Queue::getCheckpointDataNV( Dispatch
const & d )
const
18438 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18439 VULKAN_HPP_ASSERT( d.vkGetQueueCheckpointDataNV &&
"Function <vkGetQueueCheckpointDataNV> requires <VK_NV_device_diagnostic_checkpoints>" );
18442 std::vector<VULKAN_HPP_NAMESPACE::CheckpointDataNV, CheckpointDataNVAllocator> checkpointData;
18443 uint32_t checkpointDataCount;
18444 d.vkGetQueueCheckpointDataNV( m_queue, &checkpointDataCount,
nullptr );
18445 checkpointData.resize( checkpointDataCount );
18446 d.vkGetQueueCheckpointDataNV( m_queue, &checkpointDataCount,
reinterpret_cast<VkCheckpointDataNV *
>( checkpointData.data() ) );
18449 if ( checkpointDataCount < checkpointData.size() )
18451 checkpointData.resize( checkpointDataCount );
18453 return checkpointData;
18456 template <
typename CheckpointDataNVAllocator,
18458 typename std::enable_if<std::is_same<typename CheckpointDataNVAllocator::value_type, VULKAN_HPP_NAMESPACE::CheckpointDataNV>::value,
int>::type>
18460 Queue::getCheckpointDataNV( CheckpointDataNVAllocator & checkpointDataNVAllocator, Dispatch
const & d )
const
18463 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18464 VULKAN_HPP_ASSERT( d.vkGetQueueCheckpointDataNV &&
"Function <vkGetQueueCheckpointDataNV> requires <VK_NV_device_diagnostic_checkpoints>" );
18467 std::vector<VULKAN_HPP_NAMESPACE::CheckpointDataNV, CheckpointDataNVAllocator> checkpointData( checkpointDataNVAllocator );
18468 uint32_t checkpointDataCount;
18469 d.vkGetQueueCheckpointDataNV( m_queue, &checkpointDataCount,
nullptr );
18470 checkpointData.resize( checkpointDataCount );
18471 d.vkGetQueueCheckpointDataNV( m_queue, &checkpointDataCount,
reinterpret_cast<VkCheckpointDataNV *
>( checkpointData.data() ) );
18474 if ( checkpointDataCount < checkpointData.size() )
18476 checkpointData.resize( checkpointDataCount );
18478 return checkpointData;
18484 template <
typename Dispatch>
18490 return static_cast<Result>( d.vkGetSemaphoreCounterValueKHR( m_device,
static_cast<VkSemaphore>( semaphore ), pValue ) );
18493 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18494 template <
typename Dispatch>
18496 Device::getSemaphoreCounterValueKHR( VULKAN_HPP_NAMESPACE::Semaphore semaphore, Dispatch
const & d )
const
18499 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18500 VULKAN_HPP_ASSERT( d.vkGetSemaphoreCounterValueKHR &&
"Function <vkGetSemaphoreCounterValueKHR> requires <VK_KHR_timeline_semaphore> or <VK_VERSION_1_2>" );
18512 template <
typename Dispatch>
18518 return static_cast<Result>( d.vkWaitSemaphoresKHR( m_device,
reinterpret_cast<const VkSemaphoreWaitInfo *
>( pWaitInfo ), timeout ) );
18521 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18522 template <
typename Dispatch>
18524 Device::waitSemaphoresKHR(
const VULKAN_HPP_NAMESPACE::SemaphoreWaitInfo & waitInfo, uint64_t timeout, Dispatch
const & d )
const
18527 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18528 VULKAN_HPP_ASSERT( d.vkWaitSemaphoresKHR &&
"Function <vkWaitSemaphoresKHR> requires <VK_KHR_timeline_semaphore> or <VK_VERSION_1_2>" );
18534 result,
VULKAN_HPP_NAMESPACE_STRING "::Device::waitSemaphoresKHR", { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eTimeout } );
18540 template <
typename Dispatch>
18545 return static_cast<Result>( d.vkSignalSemaphoreKHR( m_device,
reinterpret_cast<const VkSemaphoreSignalInfo *
>( pSignalInfo ) ) );
18548 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18549 template <
typename Dispatch>
18551 Device::signalSemaphoreKHR(
const VULKAN_HPP_NAMESPACE::SemaphoreSignalInfo & signalInfo, Dispatch
const & d )
const
18554 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18555 VULKAN_HPP_ASSERT( d.vkSignalSemaphoreKHR &&
"Function <vkSignalSemaphoreKHR> requires <VK_KHR_timeline_semaphore> or <VK_VERSION_1_2>" );
18568 template <
typename Dispatch>
18570 const VULKAN_HPP_NAMESPACE::InitializePerformanceApiInfoINTEL * pInitializeInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
18573 return static_cast<Result>(
18577 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18578 template <
typename Dispatch>
18580 Device::initializePerformanceApiINTEL(
const VULKAN_HPP_NAMESPACE::InitializePerformanceApiInfoINTEL & initializeInfo, Dispatch
const & d )
const
18583 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18584 VULKAN_HPP_ASSERT( d.vkInitializePerformanceApiINTEL &&
"Function <vkInitializePerformanceApiINTEL> requires <VK_INTEL_performance_query>" );
18595 template <
typename Dispatch>
18599 d.vkUninitializePerformanceApiINTEL( m_device );
18602 template <
typename Dispatch>
18610 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18611 template <
typename Dispatch>
18613 CommandBuffer::setPerformanceMarkerINTEL(
const VULKAN_HPP_NAMESPACE::PerformanceMarkerInfoINTEL & markerInfo, Dispatch
const & d )
const
18616 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18617 VULKAN_HPP_ASSERT( d.vkCmdSetPerformanceMarkerINTEL &&
"Function <vkCmdSetPerformanceMarkerINTEL> requires <VK_INTEL_performance_query>" );
18628 template <
typename Dispatch>
18630 const VULKAN_HPP_NAMESPACE::PerformanceStreamMarkerInfoINTEL * pMarkerInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
18633 return static_cast<Result>(
18637 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18638 template <
typename Dispatch>
18640 CommandBuffer::setPerformanceStreamMarkerINTEL(
const VULKAN_HPP_NAMESPACE::PerformanceStreamMarkerInfoINTEL & markerInfo, Dispatch
const & d )
const
18643 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18644 VULKAN_HPP_ASSERT( d.vkCmdSetPerformanceStreamMarkerINTEL &&
"Function <vkCmdSetPerformanceStreamMarkerINTEL> requires <VK_INTEL_performance_query>" );
18655 template <
typename Dispatch>
18657 const VULKAN_HPP_NAMESPACE::PerformanceOverrideInfoINTEL * pOverrideInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
18660 return static_cast<Result>(
18664 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18665 template <
typename Dispatch>
18667 CommandBuffer::setPerformanceOverrideINTEL(
const VULKAN_HPP_NAMESPACE::PerformanceOverrideInfoINTEL & overrideInfo, Dispatch
const & d )
const
18670 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18671 VULKAN_HPP_ASSERT( d.vkCmdSetPerformanceOverrideINTEL &&
"Function <vkCmdSetPerformanceOverrideINTEL> requires <VK_INTEL_performance_query>" );
18682 template <
typename Dispatch>
18684 Device::acquirePerformanceConfigurationINTEL(
const VULKAN_HPP_NAMESPACE::PerformanceConfigurationAcquireInfoINTEL * pAcquireInfo,
18685 VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL * pConfiguration,
18689 return static_cast<Result>( d.vkAcquirePerformanceConfigurationINTEL( m_device,
18694 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18695 template <
typename Dispatch>
18697 Device::acquirePerformanceConfigurationINTEL(
const VULKAN_HPP_NAMESPACE::PerformanceConfigurationAcquireInfoINTEL & acquireInfo, Dispatch
const & d )
const
18700 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18701 VULKAN_HPP_ASSERT( d.vkAcquirePerformanceConfigurationINTEL &&
"Function <vkAcquirePerformanceConfigurationINTEL> requires <VK_INTEL_performance_query>" );
18704 VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration;
18706 d.vkAcquirePerformanceConfigurationINTEL( m_device,
18714 # ifndef VULKAN_HPP_NO_SMART_HANDLE
18715 template <
typename Dispatch>
18717 Device::acquirePerformanceConfigurationINTELUnique(
const VULKAN_HPP_NAMESPACE::PerformanceConfigurationAcquireInfoINTEL & acquireInfo,
18718 Dispatch
const & d )
const
18721 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18722 VULKAN_HPP_ASSERT( d.vkAcquirePerformanceConfigurationINTEL &&
"Function <vkAcquirePerformanceConfigurationINTEL> requires <VK_INTEL_performance_query>" );
18725 VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration;
18727 d.vkAcquirePerformanceConfigurationINTEL( m_device,
18738 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
18739 template <
typename Dispatch>
18747 template <
typename Dispatch>
18749 Device::releasePerformanceConfigurationINTEL( VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration, Dispatch
const & d )
const
18752 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18753 VULKAN_HPP_ASSERT( d.vkReleasePerformanceConfigurationINTEL &&
"Function <vkReleasePerformanceConfigurationINTEL> requires <VK_INTEL_performance_query>" );
18764 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
18765 template <
typename Dispatch>
18773 template <
typename Dispatch>
18775 Device::release( VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration, Dispatch
const & d )
const
18778 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18779 VULKAN_HPP_ASSERT( d.vkReleasePerformanceConfigurationINTEL &&
"Function <vkReleasePerformanceConfigurationINTEL> requires <VK_INTEL_performance_query>" );
18790 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
18791 template <
typename Dispatch>
18799 template <
typename Dispatch>
18801 Queue::setPerformanceConfigurationINTEL( VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration, Dispatch
const & d )
const
18804 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18806 "Function <vkQueueSetPerformanceConfigurationINTEL> requires <VK_INTEL_performance_query>" );
18817 template <
typename Dispatch>
18819 VULKAN_HPP_NAMESPACE::PerformanceValueINTEL * pValue,
18823 return static_cast<Result>( d.vkGetPerformanceParameterINTEL(
18827 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18828 template <
typename Dispatch>
18833 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18834 VULKAN_HPP_ASSERT( d.vkGetPerformanceParameterINTEL &&
"Function <vkGetPerformanceParameterINTEL> requires <VK_INTEL_performance_query>" );
18837 VULKAN_HPP_NAMESPACE::PerformanceValueINTEL value;
18848 template <
typename Dispatch>
18854 d.vkSetLocalDimmingAMD( m_device,
static_cast<VkSwapchainKHR>( swapChain ),
static_cast<VkBool32>( localDimmingEnable ) );
18857 #if defined( VK_USE_PLATFORM_FUCHSIA )
18860 template <
typename Dispatch>
18862 Instance::createImagePipeSurfaceFUCHSIA(
const VULKAN_HPP_NAMESPACE::ImagePipeSurfaceCreateInfoFUCHSIA * pCreateInfo,
18863 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
18864 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
18868 return static_cast<Result>( d.vkCreateImagePipeSurfaceFUCHSIA( m_instance,
18869 reinterpret_cast<const VkImagePipeSurfaceCreateInfoFUCHSIA *
>( pCreateInfo ),
18874 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18875 template <
typename Dispatch>
18877 Instance::createImagePipeSurfaceFUCHSIA(
const VULKAN_HPP_NAMESPACE::ImagePipeSurfaceCreateInfoFUCHSIA & createInfo,
18878 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
18879 Dispatch
const & d )
const
18882 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18883 VULKAN_HPP_ASSERT( d.vkCreateImagePipeSurfaceFUCHSIA &&
"Function <vkCreateImagePipeSurfaceFUCHSIA> requires <VK_FUCHSIA_imagepipe_surface>" );
18886 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
18889 reinterpret_cast<const VkImagePipeSurfaceCreateInfoFUCHSIA *
>( &createInfo ),
18890 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
18897 # ifndef VULKAN_HPP_NO_SMART_HANDLE
18898 template <
typename Dispatch>
18900 Instance::createImagePipeSurfaceFUCHSIAUnique(
const VULKAN_HPP_NAMESPACE::ImagePipeSurfaceCreateInfoFUCHSIA & createInfo,
18901 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
18902 Dispatch
const & d )
const
18905 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18906 VULKAN_HPP_ASSERT( d.vkCreateImagePipeSurfaceFUCHSIA &&
"Function <vkCreateImagePipeSurfaceFUCHSIA> requires <VK_FUCHSIA_imagepipe_surface>" );
18909 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
18912 reinterpret_cast<const VkImagePipeSurfaceCreateInfoFUCHSIA *
>( &createInfo ),
18913 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
18918 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
18924 #if defined( VK_USE_PLATFORM_METAL_EXT )
18927 template <
typename Dispatch>
18929 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
18930 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
18934 return static_cast<Result>( d.vkCreateMetalSurfaceEXT( m_instance,
18935 reinterpret_cast<const VkMetalSurfaceCreateInfoEXT *
>( pCreateInfo ),
18940 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
18941 template <
typename Dispatch>
18943 Instance::createMetalSurfaceEXT(
const VULKAN_HPP_NAMESPACE::MetalSurfaceCreateInfoEXT & createInfo,
18944 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
18945 Dispatch
const & d )
const
18948 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18949 VULKAN_HPP_ASSERT( d.vkCreateMetalSurfaceEXT &&
"Function <vkCreateMetalSurfaceEXT> requires <VK_EXT_metal_surface>" );
18952 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
18954 d.vkCreateMetalSurfaceEXT( m_instance,
18955 reinterpret_cast<const VkMetalSurfaceCreateInfoEXT *
>( &createInfo ),
18956 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
18963 # ifndef VULKAN_HPP_NO_SMART_HANDLE
18964 template <
typename Dispatch>
18966 Instance::createMetalSurfaceEXTUnique(
const VULKAN_HPP_NAMESPACE::MetalSurfaceCreateInfoEXT & createInfo,
18967 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
18968 Dispatch
const & d )
const
18971 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
18972 VULKAN_HPP_ASSERT( d.vkCreateMetalSurfaceEXT &&
"Function <vkCreateMetalSurfaceEXT> requires <VK_EXT_metal_surface>" );
18975 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
18977 d.vkCreateMetalSurfaceEXT( m_instance,
18978 reinterpret_cast<const VkMetalSurfaceCreateInfoEXT *
>( &createInfo ),
18979 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
18984 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
18992 template <
typename Dispatch>
18994 PhysicalDevice::getFragmentShadingRatesKHR( uint32_t * pFragmentShadingRateCount,
18995 VULKAN_HPP_NAMESPACE::PhysicalDeviceFragmentShadingRateKHR * pFragmentShadingRates,
18999 return static_cast<Result>( d.vkGetPhysicalDeviceFragmentShadingRatesKHR(
19003 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19004 template <
typename PhysicalDeviceFragmentShadingRateKHRAllocator,
typename Dispatch>
19007 PhysicalDevice::getFragmentShadingRatesKHR( Dispatch
const & d )
const
19010 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19012 "Function <vkGetPhysicalDeviceFragmentShadingRatesKHR> requires <VK_KHR_fragment_shading_rate>" );
19015 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceFragmentShadingRateKHR, PhysicalDeviceFragmentShadingRateKHRAllocator> fragmentShadingRates;
19016 uint32_t fragmentShadingRateCount;
19021 static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceFragmentShadingRatesKHR( m_physicalDevice, &fragmentShadingRateCount,
nullptr ) );
19022 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && fragmentShadingRateCount )
19024 fragmentShadingRates.resize( fragmentShadingRateCount );
19028 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19031 if ( fragmentShadingRateCount < fragmentShadingRates.size() )
19033 fragmentShadingRates.resize( fragmentShadingRateCount );
19038 template <
typename PhysicalDeviceFragmentShadingRateKHRAllocator,
19040 typename std::enable_if<std::is_same<
typename PhysicalDeviceFragmentShadingRateKHRAllocator::value_type,
19041 VULKAN_HPP_NAMESPACE::PhysicalDeviceFragmentShadingRateKHR>::value,
19045 PhysicalDevice::getFragmentShadingRatesKHR( PhysicalDeviceFragmentShadingRateKHRAllocator & physicalDeviceFragmentShadingRateKHRAllocator,
19046 Dispatch
const & d )
const
19049 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19051 "Function <vkGetPhysicalDeviceFragmentShadingRatesKHR> requires <VK_KHR_fragment_shading_rate>" );
19054 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceFragmentShadingRateKHR, PhysicalDeviceFragmentShadingRateKHRAllocator> fragmentShadingRates(
19055 physicalDeviceFragmentShadingRateKHRAllocator );
19056 uint32_t fragmentShadingRateCount;
19061 static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceFragmentShadingRatesKHR( m_physicalDevice, &fragmentShadingRateCount,
nullptr ) );
19062 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && fragmentShadingRateCount )
19064 fragmentShadingRates.resize( fragmentShadingRateCount );
19068 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19071 if ( fragmentShadingRateCount < fragmentShadingRates.size() )
19073 fragmentShadingRates.resize( fragmentShadingRateCount );
19079 template <
typename Dispatch>
19080 VULKAN_HPP_INLINE void CommandBuffer::setFragmentShadingRateKHR(
const VULKAN_HPP_NAMESPACE::Extent2D * pFragmentSize,
19085 d.vkCmdSetFragmentShadingRateKHR(
19089 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19090 template <
typename Dispatch>
19091 VULKAN_HPP_INLINE void CommandBuffer::setFragmentShadingRateKHR(
const VULKAN_HPP_NAMESPACE::Extent2D & fragmentSize,
19096 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19097 VULKAN_HPP_ASSERT( d.vkCmdSetFragmentShadingRateKHR &&
"Function <vkCmdSetFragmentShadingRateKHR> requires <VK_KHR_fragment_shading_rate>" );
19100 d.vkCmdSetFragmentShadingRateKHR(
19107 template <
typename Dispatch>
19115 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19116 template <
typename Dispatch>
19121 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19123 "Function <vkGetBufferDeviceAddressEXT> requires <VK_EXT_buffer_device_address> or <VK_KHR_buffer_device_address> or <VK_VERSION_1_2>" );
19134 template <
typename Dispatch>
19136 VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties * pToolProperties,
19140 return static_cast<Result>(
19141 d.vkGetPhysicalDeviceToolPropertiesEXT( m_physicalDevice, pToolCount,
reinterpret_cast<VkPhysicalDeviceToolProperties *
>( pToolProperties ) ) );
19144 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19145 template <
typename PhysicalDeviceToolPropertiesAllocator,
typename Dispatch>
19148 PhysicalDevice::getToolPropertiesEXT( Dispatch
const & d )
const
19151 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19153 "Function <vkGetPhysicalDeviceToolPropertiesEXT> requires <VK_EXT_tooling_info> or <VK_VERSION_1_3>" );
19156 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator> toolProperties;
19157 uint32_t toolCount;
19161 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceToolPropertiesEXT( m_physicalDevice, &toolCount,
nullptr ) );
19162 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && toolCount )
19164 toolProperties.resize( toolCount );
19166 d.vkGetPhysicalDeviceToolPropertiesEXT( m_physicalDevice, &toolCount,
reinterpret_cast<VkPhysicalDeviceToolProperties *
>( toolProperties.data() ) ) );
19168 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19171 if ( toolCount < toolProperties.size() )
19173 toolProperties.resize( toolCount );
19179 typename PhysicalDeviceToolPropertiesAllocator,
19181 typename std::enable_if<std::is_same<typename PhysicalDeviceToolPropertiesAllocator::value_type, VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties>::value,
19185 PhysicalDevice::getToolPropertiesEXT( PhysicalDeviceToolPropertiesAllocator & physicalDeviceToolPropertiesAllocator, Dispatch
const & d )
const
19188 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19190 "Function <vkGetPhysicalDeviceToolPropertiesEXT> requires <VK_EXT_tooling_info> or <VK_VERSION_1_3>" );
19193 std::vector<VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator> toolProperties(
19194 physicalDeviceToolPropertiesAllocator );
19195 uint32_t toolCount;
19199 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceToolPropertiesEXT( m_physicalDevice, &toolCount,
nullptr ) );
19200 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && toolCount )
19202 toolProperties.resize( toolCount );
19204 d.vkGetPhysicalDeviceToolPropertiesEXT( m_physicalDevice, &toolCount,
reinterpret_cast<VkPhysicalDeviceToolProperties *
>( toolProperties.data() ) ) );
19206 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19209 if ( toolCount < toolProperties.size() )
19211 toolProperties.resize( toolCount );
19219 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
19220 template <
typename Dispatch>
19222 uint64_t presentId,
19227 return static_cast<Result>( d.vkWaitForPresentKHR( m_device,
static_cast<VkSwapchainKHR>( swapchain ), presentId, timeout ) );
19230 template <
typename Dispatch>
19232 Device::waitForPresentKHR( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, uint64_t presentId, uint64_t timeout, Dispatch
const & d )
const
19235 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19236 VULKAN_HPP_ASSERT( d.vkWaitForPresentKHR &&
"Function <vkWaitForPresentKHR> requires <VK_KHR_present_wait>" );
19243 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eTimeout, VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR } );
19251 template <
typename Dispatch>
19253 uint32_t * pPropertyCount, VULKAN_HPP_NAMESPACE::CooperativeMatrixPropertiesNV * pProperties, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
19256 return static_cast<Result>( d.vkGetPhysicalDeviceCooperativeMatrixPropertiesNV(
19260 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19261 template <
typename CooperativeMatrixPropertiesNVAllocator,
typename Dispatch>
19264 PhysicalDevice::getCooperativeMatrixPropertiesNV( Dispatch
const & d )
const
19267 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19269 "Function <vkGetPhysicalDeviceCooperativeMatrixPropertiesNV> requires <VK_NV_cooperative_matrix>" );
19272 std::vector<VULKAN_HPP_NAMESPACE::CooperativeMatrixPropertiesNV, CooperativeMatrixPropertiesNVAllocator> properties;
19273 uint32_t propertyCount;
19277 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceCooperativeMatrixPropertiesNV( m_physicalDevice, &propertyCount,
nullptr ) );
19278 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
19280 properties.resize( propertyCount );
19284 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19287 if ( propertyCount < properties.size() )
19289 properties.resize( propertyCount );
19294 template <
typename CooperativeMatrixPropertiesNVAllocator,
19296 typename std::enable_if<
19297 std::is_same<typename CooperativeMatrixPropertiesNVAllocator::value_type, VULKAN_HPP_NAMESPACE::CooperativeMatrixPropertiesNV>::value,
19301 PhysicalDevice::getCooperativeMatrixPropertiesNV( CooperativeMatrixPropertiesNVAllocator & cooperativeMatrixPropertiesNVAllocator,
19302 Dispatch
const & d )
const
19305 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19307 "Function <vkGetPhysicalDeviceCooperativeMatrixPropertiesNV> requires <VK_NV_cooperative_matrix>" );
19310 std::vector<VULKAN_HPP_NAMESPACE::CooperativeMatrixPropertiesNV, CooperativeMatrixPropertiesNVAllocator> properties(
19311 cooperativeMatrixPropertiesNVAllocator );
19312 uint32_t propertyCount;
19316 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceCooperativeMatrixPropertiesNV( m_physicalDevice, &propertyCount,
nullptr ) );
19317 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
19319 properties.resize( propertyCount );
19323 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19326 if ( propertyCount < properties.size() )
19328 properties.resize( propertyCount );
19336 template <
typename Dispatch>
19338 uint32_t * pCombinationCount, VULKAN_HPP_NAMESPACE::FramebufferMixedSamplesCombinationNV * pCombinations, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
19341 return static_cast<Result>( d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV(
19345 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19346 template <
typename FramebufferMixedSamplesCombinationNVAllocator,
typename Dispatch>
19349 PhysicalDevice::getSupportedFramebufferMixedSamplesCombinationsNV( Dispatch
const & d )
const
19352 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19353 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV &&
19354 "Function <vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV> requires <VK_NV_coverage_reduction_mode>" );
19357 std::vector<VULKAN_HPP_NAMESPACE::FramebufferMixedSamplesCombinationNV, FramebufferMixedSamplesCombinationNVAllocator> combinations;
19358 uint32_t combinationCount;
19363 d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV( m_physicalDevice, &combinationCount,
nullptr ) );
19364 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && combinationCount )
19366 combinations.resize( combinationCount );
19370 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19373 if ( combinationCount < combinations.size() )
19375 combinations.resize( combinationCount );
19380 template <
typename FramebufferMixedSamplesCombinationNVAllocator,
19382 typename std::enable_if<std::is_same<
typename FramebufferMixedSamplesCombinationNVAllocator::value_type,
19383 VULKAN_HPP_NAMESPACE::FramebufferMixedSamplesCombinationNV>::value,
19387 PhysicalDevice::getSupportedFramebufferMixedSamplesCombinationsNV(
19388 FramebufferMixedSamplesCombinationNVAllocator & framebufferMixedSamplesCombinationNVAllocator, Dispatch
const & d )
const
19391 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19392 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV &&
19393 "Function <vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV> requires <VK_NV_coverage_reduction_mode>" );
19396 std::vector<VULKAN_HPP_NAMESPACE::FramebufferMixedSamplesCombinationNV, FramebufferMixedSamplesCombinationNVAllocator> combinations(
19397 framebufferMixedSamplesCombinationNVAllocator );
19398 uint32_t combinationCount;
19403 d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV( m_physicalDevice, &combinationCount,
nullptr ) );
19404 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && combinationCount )
19406 combinations.resize( combinationCount );
19410 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19413 if ( combinationCount < combinations.size() )
19415 combinations.resize( combinationCount );
19421 #if defined( VK_USE_PLATFORM_WIN32_KHR )
19424 template <
typename Dispatch>
19426 PhysicalDevice::getSurfacePresentModes2EXT(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR * pSurfaceInfo,
19427 uint32_t * pPresentModeCount,
19432 return static_cast<Result>( d.vkGetPhysicalDeviceSurfacePresentModes2EXT( m_physicalDevice,
19438 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19439 template <
typename PresentModeKHRAllocator,
typename Dispatch>
19441 PhysicalDevice::getSurfacePresentModes2EXT(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR & surfaceInfo, Dispatch
const & d )
const
19444 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19446 "Function <vkGetPhysicalDeviceSurfacePresentModes2EXT> requires <VK_EXT_full_screen_exclusive>" );
19449 std::vector<VULKAN_HPP_NAMESPACE::PresentModeKHR, PresentModeKHRAllocator> presentModes;
19450 uint32_t presentModeCount;
19456 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && presentModeCount )
19458 presentModes.resize( presentModeCount );
19460 d.vkGetPhysicalDeviceSurfacePresentModes2EXT( m_physicalDevice,
19465 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19468 if ( presentModeCount < presentModes.size() )
19470 presentModes.resize( presentModeCount );
19475 template <
typename PresentModeKHRAllocator,
19477 typename std::enable_if<std::is_same<typename PresentModeKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::PresentModeKHR>::value,
int>::type>
19479 PhysicalDevice::getSurfacePresentModes2EXT(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR & surfaceInfo,
19480 PresentModeKHRAllocator & presentModeKHRAllocator,
19481 Dispatch
const & d )
const
19484 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19486 "Function <vkGetPhysicalDeviceSurfacePresentModes2EXT> requires <VK_EXT_full_screen_exclusive>" );
19489 std::vector<VULKAN_HPP_NAMESPACE::PresentModeKHR, PresentModeKHRAllocator> presentModes( presentModeKHRAllocator );
19490 uint32_t presentModeCount;
19496 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && presentModeCount )
19498 presentModes.resize( presentModeCount );
19500 d.vkGetPhysicalDeviceSurfacePresentModes2EXT( m_physicalDevice,
19505 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
19508 if ( presentModeCount < presentModes.size() )
19510 presentModes.resize( presentModeCount );
19516 # ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
19517 template <
typename Dispatch>
19522 return static_cast<Result>( d.vkAcquireFullScreenExclusiveModeEXT( m_device,
static_cast<VkSwapchainKHR>( swapchain ) ) );
19525 template <
typename Dispatch>
19527 Device::acquireFullScreenExclusiveModeEXT( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch
const & d )
const
19530 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19531 VULKAN_HPP_ASSERT( d.vkAcquireFullScreenExclusiveModeEXT &&
"Function <vkAcquireFullScreenExclusiveModeEXT> requires <VK_EXT_full_screen_exclusive>" );
19542 # ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
19543 template <
typename Dispatch>
19548 return static_cast<Result>( d.vkReleaseFullScreenExclusiveModeEXT( m_device,
static_cast<VkSwapchainKHR>( swapchain ) ) );
19551 template <
typename Dispatch>
19553 Device::releaseFullScreenExclusiveModeEXT( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch
const & d )
const
19556 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19557 VULKAN_HPP_ASSERT( d.vkReleaseFullScreenExclusiveModeEXT &&
"Function <vkReleaseFullScreenExclusiveModeEXT> requires <VK_EXT_full_screen_exclusive>" );
19568 template <
typename Dispatch>
19570 Device::getGroupSurfacePresentModes2EXT(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR * pSurfaceInfo,
19575 return static_cast<Result>( d.vkGetDeviceGroupSurfacePresentModes2EXT(
19579 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19580 template <
typename Dispatch>
19582 Device::getGroupSurfacePresentModes2EXT(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR & surfaceInfo, Dispatch
const & d )
const
19585 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19587 "Function <vkGetDeviceGroupSurfacePresentModes2EXT> requires <VK_EXT_full_screen_exclusive>" );
19602 template <
typename Dispatch>
19604 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
19605 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
19609 return static_cast<Result>( d.vkCreateHeadlessSurfaceEXT( m_instance,
19615 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19616 template <
typename Dispatch>
19618 Instance::createHeadlessSurfaceEXT(
const VULKAN_HPP_NAMESPACE::HeadlessSurfaceCreateInfoEXT & createInfo,
19620 Dispatch
const & d )
const
19623 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19624 VULKAN_HPP_ASSERT( d.vkCreateHeadlessSurfaceEXT &&
"Function <vkCreateHeadlessSurfaceEXT> requires <VK_EXT_headless_surface>" );
19627 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
19631 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
19638 # ifndef VULKAN_HPP_NO_SMART_HANDLE
19639 template <
typename Dispatch>
19641 Instance::createHeadlessSurfaceEXTUnique(
const VULKAN_HPP_NAMESPACE::HeadlessSurfaceCreateInfoEXT & createInfo,
19643 Dispatch
const & d )
const
19646 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19647 VULKAN_HPP_ASSERT( d.vkCreateHeadlessSurfaceEXT &&
"Function <vkCreateHeadlessSurfaceEXT> requires <VK_EXT_headless_surface>" );
19650 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
19654 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
19666 template <
typename Dispatch>
19674 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19675 template <
typename Dispatch>
19680 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19682 "Function <vkGetBufferDeviceAddressKHR> requires <VK_EXT_buffer_device_address> or <VK_KHR_buffer_device_address> or <VK_VERSION_1_2>" );
19691 template <
typename Dispatch>
19692 VULKAN_HPP_INLINE uint64_t Device::getBufferOpaqueCaptureAddressKHR(
const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo * pInfo,
19699 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19700 template <
typename Dispatch>
19701 VULKAN_HPP_INLINE uint64_t Device::getBufferOpaqueCaptureAddressKHR(
const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo & info,
19705 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19707 "Function <vkGetBufferOpaqueCaptureAddressKHR> requires <VK_KHR_buffer_device_address> or <VK_VERSION_1_2>" );
19710 uint64_t result = d.vkGetBufferOpaqueCaptureAddressKHR( m_device,
reinterpret_cast<const VkBufferDeviceAddressInfo *
>( &info ) );
19716 template <
typename Dispatch>
19717 VULKAN_HPP_INLINE uint64_t Device::getMemoryOpaqueCaptureAddressKHR(
const VULKAN_HPP_NAMESPACE::DeviceMemoryOpaqueCaptureAddressInfo * pInfo,
19724 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19725 template <
typename Dispatch>
19726 VULKAN_HPP_INLINE uint64_t Device::getMemoryOpaqueCaptureAddressKHR(
const VULKAN_HPP_NAMESPACE::DeviceMemoryOpaqueCaptureAddressInfo & info,
19730 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19732 "Function <vkGetDeviceMemoryOpaqueCaptureAddressKHR> requires <VK_KHR_buffer_device_address> or <VK_VERSION_1_2>" );
19743 template <
typename Dispatch>
19745 CommandBuffer::setLineStippleEXT( uint32_t lineStippleFactor, uint16_t lineStipplePattern, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
19748 d.vkCmdSetLineStippleEXT( m_commandBuffer, lineStippleFactor, lineStipplePattern );
19753 template <
typename Dispatch>
19755 uint32_t firstQuery,
19756 uint32_t queryCount,
19760 d.vkResetQueryPoolEXT( m_device,
static_cast<VkQueryPool>( queryPool ), firstQuery, queryCount );
19765 template <
typename Dispatch>
19769 d.vkCmdSetCullModeEXT( m_commandBuffer,
static_cast<VkCullModeFlags>( cullMode ) );
19772 template <
typename Dispatch>
19776 d.vkCmdSetFrontFaceEXT( m_commandBuffer,
static_cast<VkFrontFace>( frontFace ) );
19779 template <
typename Dispatch>
19784 d.vkCmdSetPrimitiveTopologyEXT( m_commandBuffer,
static_cast<VkPrimitiveTopology>( primitiveTopology ) );
19787 template <
typename Dispatch>
19789 const VULKAN_HPP_NAMESPACE::Viewport * pViewports,
19793 d.vkCmdSetViewportWithCountEXT( m_commandBuffer, viewportCount,
reinterpret_cast<const VkViewport *
>( pViewports ) );
19796 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19797 template <
typename Dispatch>
19798 VULKAN_HPP_INLINE void CommandBuffer::setViewportWithCountEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Viewport>
const & viewports,
19802 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19804 "Function <vkCmdSetViewportWithCountEXT> requires <VK_EXT_extended_dynamic_state> or <VK_EXT_shader_object> or <VK_VERSION_1_3>" );
19807 d.vkCmdSetViewportWithCountEXT( m_commandBuffer, viewports.size(),
reinterpret_cast<const VkViewport *
>( viewports.data() ) );
19811 template <
typename Dispatch>
19813 CommandBuffer::setScissorWithCountEXT( uint32_t scissorCount,
const VULKAN_HPP_NAMESPACE::Rect2D * pScissors, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
19816 d.vkCmdSetScissorWithCountEXT( m_commandBuffer, scissorCount,
reinterpret_cast<const VkRect2D *
>( pScissors ) );
19819 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19820 template <
typename Dispatch>
19821 VULKAN_HPP_INLINE void CommandBuffer::setScissorWithCountEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D>
const & scissors,
19825 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19827 "Function <vkCmdSetScissorWithCountEXT> requires <VK_EXT_extended_dynamic_state> or <VK_EXT_shader_object> or <VK_VERSION_1_3>" );
19830 d.vkCmdSetScissorWithCountEXT( m_commandBuffer, scissors.size(),
reinterpret_cast<const VkRect2D *
>( scissors.data() ) );
19834 template <
typename Dispatch>
19836 uint32_t bindingCount,
19837 const VULKAN_HPP_NAMESPACE::Buffer * pBuffers,
19844 d.vkCmdBindVertexBuffers2EXT( m_commandBuffer,
19847 reinterpret_cast<const VkBuffer *
>( pBuffers ),
19853 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19854 template <
typename Dispatch>
19856 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer>
const & buffers,
19857 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & offsets,
19858 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & sizes,
19859 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & strides,
19863 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19865 "Function <vkCmdBindVertexBuffers2EXT> requires <VK_EXT_extended_dynamic_state> or <VK_EXT_shader_object> or <VK_VERSION_1_3>" );
19867 # ifdef VULKAN_HPP_NO_EXCEPTIONS
19872 if ( buffers.size() != offsets.size() )
19876 if ( !sizes.empty() && buffers.size() != sizes.size() )
19880 if ( !strides.empty() && buffers.size() != strides.size() )
19886 d.vkCmdBindVertexBuffers2EXT( m_commandBuffer,
19889 reinterpret_cast<const VkBuffer *
>( buffers.data() ),
19890 reinterpret_cast<const VkDeviceSize *
>( offsets.data() ),
19891 reinterpret_cast<const VkDeviceSize *
>( sizes.data() ),
19892 reinterpret_cast<const VkDeviceSize *
>( strides.data() ) );
19896 template <
typename Dispatch>
19900 d.vkCmdSetDepthTestEnableEXT( m_commandBuffer,
static_cast<VkBool32>( depthTestEnable ) );
19903 template <
typename Dispatch>
19907 d.vkCmdSetDepthWriteEnableEXT( m_commandBuffer,
static_cast<VkBool32>( depthWriteEnable ) );
19910 template <
typename Dispatch>
19914 d.vkCmdSetDepthCompareOpEXT( m_commandBuffer,
static_cast<VkCompareOp>( depthCompareOp ) );
19917 template <
typename Dispatch>
19922 d.vkCmdSetDepthBoundsTestEnableEXT( m_commandBuffer,
static_cast<VkBool32>( depthBoundsTestEnable ) );
19925 template <
typename Dispatch>
19929 d.vkCmdSetStencilTestEnableEXT( m_commandBuffer,
static_cast<VkBool32>( stencilTestEnable ) );
19932 template <
typename Dispatch>
19941 d.vkCmdSetStencilOpEXT( m_commandBuffer,
19951 template <
typename Dispatch>
19953 VULKAN_HPP_NAMESPACE::DeferredOperationKHR * pDeferredOperation,
19957 return static_cast<Result>( d.vkCreateDeferredOperationKHR(
19961 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
19962 template <
typename Dispatch>
19967 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19968 VULKAN_HPP_ASSERT( d.vkCreateDeferredOperationKHR &&
"Function <vkCreateDeferredOperationKHR> requires <VK_KHR_deferred_host_operations>" );
19971 VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation;
19974 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
19981 # ifndef VULKAN_HPP_NO_SMART_HANDLE
19982 template <
typename Dispatch>
19987 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
19988 VULKAN_HPP_ASSERT( d.vkCreateDeferredOperationKHR &&
"Function <vkCreateDeferredOperationKHR> requires <VK_KHR_deferred_host_operations>" );
19991 VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation;
19994 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
20004 template <
typename Dispatch>
20005 VULKAN_HPP_INLINE void Device::destroyDeferredOperationKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation,
20006 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
20010 d.vkDestroyDeferredOperationKHR(
20014 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20015 template <
typename Dispatch>
20016 VULKAN_HPP_INLINE void Device::destroyDeferredOperationKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation,
20021 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20022 VULKAN_HPP_ASSERT( d.vkDestroyDeferredOperationKHR &&
"Function <vkDestroyDeferredOperationKHR> requires <VK_KHR_deferred_host_operations>" );
20025 d.vkDestroyDeferredOperationKHR(
20028 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
20032 template <
typename Dispatch>
20034 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
20038 d.vkDestroyDeferredOperationKHR(
20042 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20043 template <
typename Dispatch>
20049 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20050 VULKAN_HPP_ASSERT( d.vkDestroyDeferredOperationKHR &&
"Function <vkDestroyDeferredOperationKHR> requires <VK_KHR_deferred_host_operations>" );
20053 d.vkDestroyDeferredOperationKHR(
20056 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
20060 template <
typename Dispatch>
20061 VULKAN_HPP_INLINE uint32_t Device::getDeferredOperationMaxConcurrencyKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation,
20065 return d.vkGetDeferredOperationMaxConcurrencyKHR( m_device,
static_cast<VkDeferredOperationKHR>( operation ) );
20068 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
20069 template <
typename Dispatch>
20077 template <
typename Dispatch>
20079 Device::getDeferredOperationResultKHR( VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
20082 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20083 VULKAN_HPP_ASSERT( d.vkGetDeferredOperationResultKHR &&
"Function <vkGetDeferredOperationResultKHR> requires <VK_KHR_deferred_host_operations>" );
20093 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
20094 template <
typename Dispatch>
20102 template <
typename Dispatch>
20104 Dispatch
const & d )
const
20107 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20108 VULKAN_HPP_ASSERT( d.vkDeferredOperationJoinKHR &&
"Function <vkDeferredOperationJoinKHR> requires <VK_KHR_deferred_host_operations>" );
20115 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eThreadDoneKHR, VULKAN_HPP_NAMESPACE::Result::eThreadIdleKHR } );
20123 template <
typename Dispatch>
20125 uint32_t * pExecutableCount,
20126 VULKAN_HPP_NAMESPACE::PipelineExecutablePropertiesKHR * pProperties,
20130 return static_cast<Result>( d.vkGetPipelineExecutablePropertiesKHR( m_device,
20136 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20137 template <
typename PipelineExecutablePropertiesKHRAllocator,
typename Dispatch>
20140 Device::getPipelineExecutablePropertiesKHR(
const VULKAN_HPP_NAMESPACE::PipelineInfoKHR & pipelineInfo, Dispatch
const & d )
const
20143 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20145 "Function <vkGetPipelineExecutablePropertiesKHR> requires <VK_KHR_pipeline_executable_properties>" );
20148 std::vector<VULKAN_HPP_NAMESPACE::PipelineExecutablePropertiesKHR, PipelineExecutablePropertiesKHRAllocator> properties;
20149 uint32_t executableCount;
20154 d.vkGetPipelineExecutablePropertiesKHR( m_device,
reinterpret_cast<const VkPipelineInfoKHR *
>( &pipelineInfo ), &executableCount,
nullptr ) );
20155 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && executableCount )
20157 properties.resize( executableCount );
20159 d.vkGetPipelineExecutablePropertiesKHR( m_device,
20164 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
20167 if ( executableCount < properties.size() )
20169 properties.resize( executableCount );
20174 template <
typename PipelineExecutablePropertiesKHRAllocator,
20176 typename std::enable_if<
20177 std::is_same<typename PipelineExecutablePropertiesKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::PipelineExecutablePropertiesKHR>::value,
20181 Device::getPipelineExecutablePropertiesKHR(
const VULKAN_HPP_NAMESPACE::PipelineInfoKHR & pipelineInfo,
20182 PipelineExecutablePropertiesKHRAllocator & pipelineExecutablePropertiesKHRAllocator,
20183 Dispatch
const & d )
const
20186 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20188 "Function <vkGetPipelineExecutablePropertiesKHR> requires <VK_KHR_pipeline_executable_properties>" );
20191 std::vector<VULKAN_HPP_NAMESPACE::PipelineExecutablePropertiesKHR, PipelineExecutablePropertiesKHRAllocator> properties(
20192 pipelineExecutablePropertiesKHRAllocator );
20193 uint32_t executableCount;
20198 d.vkGetPipelineExecutablePropertiesKHR( m_device,
reinterpret_cast<const VkPipelineInfoKHR *
>( &pipelineInfo ), &executableCount,
nullptr ) );
20199 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && executableCount )
20201 properties.resize( executableCount );
20203 d.vkGetPipelineExecutablePropertiesKHR( m_device,
20208 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
20211 if ( executableCount < properties.size() )
20213 properties.resize( executableCount );
20219 template <
typename Dispatch>
20221 Device::getPipelineExecutableStatisticsKHR(
const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR * pExecutableInfo,
20222 uint32_t * pStatisticCount,
20223 VULKAN_HPP_NAMESPACE::PipelineExecutableStatisticKHR * pStatistics,
20227 return static_cast<Result>( d.vkGetPipelineExecutableStatisticsKHR( m_device,
20233 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20234 template <
typename PipelineExecutableStatisticKHRAllocator,
typename Dispatch>
20237 Device::getPipelineExecutableStatisticsKHR(
const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR & executableInfo, Dispatch
const & d )
const
20240 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20242 "Function <vkGetPipelineExecutableStatisticsKHR> requires <VK_KHR_pipeline_executable_properties>" );
20245 std::vector<VULKAN_HPP_NAMESPACE::PipelineExecutableStatisticKHR, PipelineExecutableStatisticKHRAllocator> statistics;
20246 uint32_t statisticCount;
20252 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && statisticCount )
20254 statistics.resize( statisticCount );
20256 d.vkGetPipelineExecutableStatisticsKHR( m_device,
20261 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
20264 if ( statisticCount < statistics.size() )
20266 statistics.resize( statisticCount );
20271 template <
typename PipelineExecutableStatisticKHRAllocator,
20273 typename std::enable_if<
20274 std::is_same<typename PipelineExecutableStatisticKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::PipelineExecutableStatisticKHR>::value,
20278 Device::getPipelineExecutableStatisticsKHR(
const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR & executableInfo,
20279 PipelineExecutableStatisticKHRAllocator & pipelineExecutableStatisticKHRAllocator,
20280 Dispatch
const & d )
const
20283 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20285 "Function <vkGetPipelineExecutableStatisticsKHR> requires <VK_KHR_pipeline_executable_properties>" );
20288 std::vector<VULKAN_HPP_NAMESPACE::PipelineExecutableStatisticKHR, PipelineExecutableStatisticKHRAllocator> statistics(
20289 pipelineExecutableStatisticKHRAllocator );
20290 uint32_t statisticCount;
20296 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && statisticCount )
20298 statistics.resize( statisticCount );
20300 d.vkGetPipelineExecutableStatisticsKHR( m_device,
20305 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
20308 if ( statisticCount < statistics.size() )
20310 statistics.resize( statisticCount );
20316 template <
typename Dispatch>
20318 Device::getPipelineExecutableInternalRepresentationsKHR(
const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR * pExecutableInfo,
20319 uint32_t * pInternalRepresentationCount,
20320 VULKAN_HPP_NAMESPACE::PipelineExecutableInternalRepresentationKHR * pInternalRepresentations,
20324 return static_cast<Result>(
20325 d.vkGetPipelineExecutableInternalRepresentationsKHR( m_device,
20327 pInternalRepresentationCount,
20331 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20332 template <
typename PipelineExecutableInternalRepresentationKHRAllocator,
typename Dispatch>
20334 std::vector<VULKAN_HPP_NAMESPACE::PipelineExecutableInternalRepresentationKHR, PipelineExecutableInternalRepresentationKHRAllocator>>::type
20335 Device::getPipelineExecutableInternalRepresentationsKHR(
const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR & executableInfo, Dispatch
const & d )
const
20338 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20340 "Function <vkGetPipelineExecutableInternalRepresentationsKHR> requires <VK_KHR_pipeline_executable_properties>" );
20343 std::vector<VULKAN_HPP_NAMESPACE::PipelineExecutableInternalRepresentationKHR, PipelineExecutableInternalRepresentationKHRAllocator>
20344 internalRepresentations;
20345 uint32_t internalRepresentationCount;
20351 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && internalRepresentationCount )
20353 internalRepresentations.resize( internalRepresentationCount );
20357 &internalRepresentationCount,
20360 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
20362 VULKAN_HPP_ASSERT( internalRepresentationCount <= internalRepresentations.size() );
20363 if ( internalRepresentationCount < internalRepresentations.size() )
20365 internalRepresentations.resize( internalRepresentationCount );
20370 template <
typename PipelineExecutableInternalRepresentationKHRAllocator,
20372 typename std::enable_if<std::is_same<
typename PipelineExecutableInternalRepresentationKHRAllocator::value_type,
20373 VULKAN_HPP_NAMESPACE::PipelineExecutableInternalRepresentationKHR>::value,
20376 std::vector<VULKAN_HPP_NAMESPACE::PipelineExecutableInternalRepresentationKHR, PipelineExecutableInternalRepresentationKHRAllocator>>::type
20377 Device::getPipelineExecutableInternalRepresentationsKHR(
20378 const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR & executableInfo,
20379 PipelineExecutableInternalRepresentationKHRAllocator & pipelineExecutableInternalRepresentationKHRAllocator,
20380 Dispatch
const & d )
const
20383 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20385 "Function <vkGetPipelineExecutableInternalRepresentationsKHR> requires <VK_KHR_pipeline_executable_properties>" );
20388 std::vector<VULKAN_HPP_NAMESPACE::PipelineExecutableInternalRepresentationKHR, PipelineExecutableInternalRepresentationKHRAllocator>
20389 internalRepresentations( pipelineExecutableInternalRepresentationKHRAllocator );
20390 uint32_t internalRepresentationCount;
20396 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && internalRepresentationCount )
20398 internalRepresentations.resize( internalRepresentationCount );
20402 &internalRepresentationCount,
20405 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
20407 VULKAN_HPP_ASSERT( internalRepresentationCount <= internalRepresentations.size() );
20408 if ( internalRepresentationCount < internalRepresentations.size() )
20410 internalRepresentations.resize( internalRepresentationCount );
20418 template <
typename Dispatch>
20426 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20427 template <
typename Dispatch>
20429 Device::copyMemoryToImageEXT(
const VULKAN_HPP_NAMESPACE::CopyMemoryToImageInfoEXT & copyMemoryToImageInfo, Dispatch
const & d )
const
20432 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20433 VULKAN_HPP_ASSERT( d.vkCopyMemoryToImageEXT &&
"Function <vkCopyMemoryToImageEXT> requires <VK_EXT_host_image_copy>" );
20444 template <
typename Dispatch>
20452 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20453 template <
typename Dispatch>
20455 Device::copyImageToMemoryEXT(
const VULKAN_HPP_NAMESPACE::CopyImageToMemoryInfoEXT & copyImageToMemoryInfo, Dispatch
const & d )
const
20458 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20459 VULKAN_HPP_ASSERT( d.vkCopyImageToMemoryEXT &&
"Function <vkCopyImageToMemoryEXT> requires <VK_EXT_host_image_copy>" );
20470 template <
typename Dispatch>
20478 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20479 template <
typename Dispatch>
20481 Device::copyImageToImageEXT(
const VULKAN_HPP_NAMESPACE::CopyImageToImageInfoEXT & copyImageToImageInfo, Dispatch
const & d )
const
20484 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20485 VULKAN_HPP_ASSERT( d.vkCopyImageToImageEXT &&
"Function <vkCopyImageToImageEXT> requires <VK_EXT_host_image_copy>" );
20496 template <
typename Dispatch>
20498 const VULKAN_HPP_NAMESPACE::HostImageLayoutTransitionInfoEXT * pTransitions,
20502 return static_cast<Result>(
20506 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20507 template <
typename Dispatch>
20509 Device::transitionImageLayoutEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::HostImageLayoutTransitionInfoEXT>
const & transitions,
20510 Dispatch
const & d )
const
20513 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20514 VULKAN_HPP_ASSERT( d.vkTransitionImageLayoutEXT &&
"Function <vkTransitionImageLayoutEXT> requires <VK_EXT_host_image_copy>" );
20525 template <
typename Dispatch>
20527 const VULKAN_HPP_NAMESPACE::ImageSubresource2KHR * pSubresource,
20528 VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR * pLayout,
20532 d.vkGetImageSubresourceLayout2EXT( m_device,
20533 static_cast<VkImage>( image ),
20538 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20539 template <
typename Dispatch>
20541 VULKAN_HPP_NAMESPACE::Image image,
const VULKAN_HPP_NAMESPACE::ImageSubresource2KHR & subresource, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
20544 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20546 d.vkGetImageSubresourceLayout2EXT &&
20547 "Function <vkGetImageSubresourceLayout2EXT> requires <VK_EXT_host_image_copy> or <VK_EXT_image_compression_control> or <VK_KHR_maintenance5>" );
20550 VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR layout;
20551 d.vkGetImageSubresourceLayout2EXT( m_device,
20552 static_cast<VkImage>( image ),
20559 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
20561 VULKAN_HPP_NAMESPACE::Image image,
const VULKAN_HPP_NAMESPACE::ImageSubresource2KHR & subresource, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
20564 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20566 d.vkGetImageSubresourceLayout2EXT &&
20567 "Function <vkGetImageSubresourceLayout2EXT> requires <VK_EXT_host_image_copy> or <VK_EXT_image_compression_control> or <VK_KHR_maintenance5>" );
20570 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
20571 VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR & layout = structureChain.template get<VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR>();
20572 d.vkGetImageSubresourceLayout2EXT( m_device,
20573 static_cast<VkImage>( image ),
20577 return structureChain;
20583 template <
typename Dispatch>
20589 return static_cast<Result>( d.vkMapMemory2KHR( m_device,
reinterpret_cast<const VkMemoryMapInfoKHR *
>( pMemoryMapInfo ), ppData ) );
20592 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20593 template <
typename Dispatch>
20595 Device::mapMemory2KHR(
const VULKAN_HPP_NAMESPACE::MemoryMapInfoKHR & memoryMapInfo, Dispatch
const & d )
const
20598 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20599 VULKAN_HPP_ASSERT( d.vkMapMemory2KHR &&
"Function <vkMapMemory2KHR> requires <VK_KHR_map_memory2>" );
20611 template <
typename Dispatch>
20616 return static_cast<Result>( d.vkUnmapMemory2KHR( m_device,
reinterpret_cast<const VkMemoryUnmapInfoKHR *
>( pMemoryUnmapInfo ) ) );
20619 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20620 template <
typename Dispatch>
20621 VULKAN_HPP_INLINE void Device::unmapMemory2KHR(
const VULKAN_HPP_NAMESPACE::MemoryUnmapInfoKHR & memoryUnmapInfo,
20625 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20626 VULKAN_HPP_ASSERT( d.vkUnmapMemory2KHR &&
"Function <vkUnmapMemory2KHR> requires <VK_KHR_map_memory2>" );
20629 d.vkUnmapMemory2KHR( m_device,
reinterpret_cast<const VkMemoryUnmapInfoKHR *
>( &memoryUnmapInfo ) );
20635 template <
typename Dispatch>
20643 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20644 template <
typename Dispatch>
20646 Device::releaseSwapchainImagesEXT(
const VULKAN_HPP_NAMESPACE::ReleaseSwapchainImagesInfoEXT & releaseInfo, Dispatch
const & d )
const
20649 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20650 VULKAN_HPP_ASSERT( d.vkReleaseSwapchainImagesEXT &&
"Function <vkReleaseSwapchainImagesEXT> requires <VK_EXT_swapchain_maintenance1>" );
20663 template <
typename Dispatch>
20664 VULKAN_HPP_INLINE void Device::getGeneratedCommandsMemoryRequirementsNV(
const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoNV * pInfo,
20665 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
20669 d.vkGetGeneratedCommandsMemoryRequirementsNV( m_device,
20674 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20675 template <
typename Dispatch>
20677 Device::getGeneratedCommandsMemoryRequirementsNV(
const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoNV & info,
20681 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20683 "Function <vkGetGeneratedCommandsMemoryRequirementsNV> requires <VK_NV_device_generated_commands>" );
20686 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
20687 d.vkGetGeneratedCommandsMemoryRequirementsNV( m_device,
20691 return memoryRequirements;
20694 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
20696 Device::getGeneratedCommandsMemoryRequirementsNV(
const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoNV & info,
20700 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20702 "Function <vkGetGeneratedCommandsMemoryRequirementsNV> requires <VK_NV_device_generated_commands>" );
20705 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
20706 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
20707 d.vkGetGeneratedCommandsMemoryRequirementsNV( m_device,
20711 return structureChain;
20715 template <
typename Dispatch>
20716 VULKAN_HPP_INLINE void CommandBuffer::preprocessGeneratedCommandsNV(
const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoNV * pGeneratedCommandsInfo,
20720 d.vkCmdPreprocessGeneratedCommandsNV( m_commandBuffer,
reinterpret_cast<const VkGeneratedCommandsInfoNV *
>( pGeneratedCommandsInfo ) );
20723 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20724 template <
typename Dispatch>
20725 VULKAN_HPP_INLINE void CommandBuffer::preprocessGeneratedCommandsNV(
const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoNV & generatedCommandsInfo,
20729 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20730 VULKAN_HPP_ASSERT( d.vkCmdPreprocessGeneratedCommandsNV &&
"Function <vkCmdPreprocessGeneratedCommandsNV> requires <VK_NV_device_generated_commands>" );
20733 d.vkCmdPreprocessGeneratedCommandsNV( m_commandBuffer,
reinterpret_cast<const VkGeneratedCommandsInfoNV *
>( &generatedCommandsInfo ) );
20737 template <
typename Dispatch>
20739 const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoNV * pGeneratedCommandsInfo,
20743 d.vkCmdExecuteGeneratedCommandsNV(
20747 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20748 template <
typename Dispatch>
20750 const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoNV & generatedCommandsInfo,
20754 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20755 VULKAN_HPP_ASSERT( d.vkCmdExecuteGeneratedCommandsNV &&
"Function <vkCmdExecuteGeneratedCommandsNV> requires <VK_NV_device_generated_commands>" );
20758 d.vkCmdExecuteGeneratedCommandsNV(
20763 template <
typename Dispatch>
20765 VULKAN_HPP_NAMESPACE::Pipeline pipeline,
20766 uint32_t groupIndex,
20770 d.vkCmdBindPipelineShaderGroupNV( m_commandBuffer,
static_cast<VkPipelineBindPoint>( pipelineBindPoint ),
static_cast<VkPipeline>( pipeline ), groupIndex );
20773 template <
typename Dispatch>
20775 Device::createIndirectCommandsLayoutNV(
const VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutCreateInfoNV * pCreateInfo,
20776 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
20777 VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV * pIndirectCommandsLayout,
20781 return static_cast<Result>( d.vkCreateIndirectCommandsLayoutNV( m_device,
20787 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20788 template <
typename Dispatch>
20790 Device::createIndirectCommandsLayoutNV(
const VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutCreateInfoNV & createInfo,
20792 Dispatch
const & d )
const
20795 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20796 VULKAN_HPP_ASSERT( d.vkCreateIndirectCommandsLayoutNV &&
"Function <vkCreateIndirectCommandsLayoutNV> requires <VK_NV_device_generated_commands>" );
20799 VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout;
20803 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
20810 # ifndef VULKAN_HPP_NO_SMART_HANDLE
20811 template <
typename Dispatch>
20813 Device::createIndirectCommandsLayoutNVUnique(
const VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutCreateInfoNV & createInfo,
20815 Dispatch
const & d )
const
20818 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20819 VULKAN_HPP_ASSERT( d.vkCreateIndirectCommandsLayoutNV &&
"Function <vkCreateIndirectCommandsLayoutNV> requires <VK_NV_device_generated_commands>" );
20822 VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout;
20826 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
20837 template <
typename Dispatch>
20838 VULKAN_HPP_INLINE void Device::destroyIndirectCommandsLayoutNV( VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout,
20839 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
20843 d.vkDestroyIndirectCommandsLayoutNV(
20847 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20848 template <
typename Dispatch>
20849 VULKAN_HPP_INLINE void Device::destroyIndirectCommandsLayoutNV( VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout,
20854 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20855 VULKAN_HPP_ASSERT( d.vkDestroyIndirectCommandsLayoutNV &&
"Function <vkDestroyIndirectCommandsLayoutNV> requires <VK_NV_device_generated_commands>" );
20858 d.vkDestroyIndirectCommandsLayoutNV(
20861 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
20865 template <
typename Dispatch>
20866 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout,
20867 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
20871 d.vkDestroyIndirectCommandsLayoutNV(
20875 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20876 template <
typename Dispatch>
20877 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout,
20882 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20883 VULKAN_HPP_ASSERT( d.vkDestroyIndirectCommandsLayoutNV &&
"Function <vkDestroyIndirectCommandsLayoutNV> requires <VK_NV_device_generated_commands>" );
20886 d.vkDestroyIndirectCommandsLayoutNV(
20889 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
20895 template <
typename Dispatch>
20896 VULKAN_HPP_INLINE void CommandBuffer::setDepthBias2EXT(
const VULKAN_HPP_NAMESPACE::DepthBiasInfoEXT * pDepthBiasInfo,
20900 d.vkCmdSetDepthBias2EXT( m_commandBuffer,
reinterpret_cast<const VkDepthBiasInfoEXT *
>( pDepthBiasInfo ) );
20903 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20904 template <
typename Dispatch>
20905 VULKAN_HPP_INLINE void CommandBuffer::setDepthBias2EXT(
const VULKAN_HPP_NAMESPACE::DepthBiasInfoEXT & depthBiasInfo,
20909 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20910 VULKAN_HPP_ASSERT( d.vkCmdSetDepthBias2EXT &&
"Function <vkCmdSetDepthBias2EXT> requires <VK_EXT_depth_bias_control>" );
20913 d.vkCmdSetDepthBias2EXT( m_commandBuffer,
reinterpret_cast<const VkDepthBiasInfoEXT *
>( &depthBiasInfo ) );
20919 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
20920 template <
typename Dispatch>
20922 VULKAN_HPP_NAMESPACE::DisplayKHR display,
20926 return static_cast<Result>( d.vkAcquireDrmDisplayEXT( m_physicalDevice, drmFd,
static_cast<VkDisplayKHR>( display ) ) );
20929 template <
typename Dispatch>
20931 PhysicalDevice::acquireDrmDisplayEXT( int32_t drmFd, VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch
const & d )
const
20934 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20935 VULKAN_HPP_ASSERT( d.vkAcquireDrmDisplayEXT &&
"Function <vkAcquireDrmDisplayEXT> requires <VK_EXT_acquire_drm_display>" );
20946 template <
typename Dispatch>
20948 uint32_t connectorId,
20949 VULKAN_HPP_NAMESPACE::DisplayKHR * display,
20953 return static_cast<Result>( d.vkGetDrmDisplayEXT( m_physicalDevice, drmFd, connectorId,
reinterpret_cast<VkDisplayKHR *
>( display ) ) );
20956 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
20957 template <
typename Dispatch>
20959 PhysicalDevice::getDrmDisplayEXT( int32_t drmFd, uint32_t connectorId, Dispatch
const & d )
const
20962 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20963 VULKAN_HPP_ASSERT( d.vkGetDrmDisplayEXT &&
"Function <vkGetDrmDisplayEXT> requires <VK_EXT_acquire_drm_display>" );
20966 VULKAN_HPP_NAMESPACE::DisplayKHR display;
20974 # ifndef VULKAN_HPP_NO_SMART_HANDLE
20975 template <
typename Dispatch>
20977 PhysicalDevice::getDrmDisplayEXTUnique( int32_t drmFd, uint32_t connectorId, Dispatch
const & d )
const
20980 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
20981 VULKAN_HPP_ASSERT( d.vkGetDrmDisplayEXT &&
"Function <vkGetDrmDisplayEXT> requires <VK_EXT_acquire_drm_display>" );
20984 VULKAN_HPP_NAMESPACE::DisplayKHR display;
20997 template <
typename Dispatch>
20999 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
21000 VULKAN_HPP_NAMESPACE::PrivateDataSlot * pPrivateDataSlot,
21004 return static_cast<Result>( d.vkCreatePrivateDataSlotEXT( m_device,
21010 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21011 template <
typename Dispatch>
21013 Device::createPrivateDataSlotEXT(
const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo & createInfo,
21015 Dispatch
const & d )
const
21018 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21019 VULKAN_HPP_ASSERT( d.vkCreatePrivateDataSlotEXT &&
"Function <vkCreatePrivateDataSlotEXT> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
21022 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot;
21026 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
21033 # ifndef VULKAN_HPP_NO_SMART_HANDLE
21034 template <
typename Dispatch>
21036 Device::createPrivateDataSlotEXTUnique(
const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo & createInfo,
21038 Dispatch
const & d )
const
21041 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21042 VULKAN_HPP_ASSERT( d.vkCreatePrivateDataSlotEXT &&
"Function <vkCreatePrivateDataSlotEXT> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
21045 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot;
21049 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
21059 template <
typename Dispatch>
21060 VULKAN_HPP_INLINE void Device::destroyPrivateDataSlotEXT( VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
21061 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
21068 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21069 template <
typename Dispatch>
21070 VULKAN_HPP_INLINE void Device::destroyPrivateDataSlotEXT( VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
21075 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21076 VULKAN_HPP_ASSERT( d.vkDestroyPrivateDataSlotEXT &&
"Function <vkDestroyPrivateDataSlotEXT> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
21079 d.vkDestroyPrivateDataSlotEXT(
21082 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
21086 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
21087 template <
typename Dispatch>
21089 uint64_t objectHandle,
21090 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
21095 return static_cast<Result>(
21096 d.vkSetPrivateDataEXT( m_device,
static_cast<VkObjectType>( objectType_ ), objectHandle,
static_cast<VkPrivateDataSlot>( privateDataSlot ), data ) );
21099 template <
typename Dispatch>
21101 uint64_t objectHandle,
21102 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
21104 Dispatch
const & d )
const
21107 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21108 VULKAN_HPP_ASSERT( d.vkSetPrivateDataEXT &&
"Function <vkSetPrivateDataEXT> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
21112 d.vkSetPrivateDataEXT( m_device,
static_cast<VkObjectType>( objectType_ ), objectHandle,
static_cast<VkPrivateDataSlot>( privateDataSlot ), data ) );
21119 template <
typename Dispatch>
21121 uint64_t objectHandle,
21122 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
21127 d.vkGetPrivateDataEXT( m_device,
static_cast<VkObjectType>( objectType_ ), objectHandle,
static_cast<VkPrivateDataSlot>( privateDataSlot ), pData );
21130 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21131 template <
typename Dispatch>
21133 uint64_t objectHandle,
21134 VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot,
21138 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21139 VULKAN_HPP_ASSERT( d.vkGetPrivateDataEXT &&
"Function <vkGetPrivateDataEXT> requires <VK_EXT_private_data> or <VK_VERSION_1_3>" );
21143 d.vkGetPrivateDataEXT( m_device,
static_cast<VkObjectType>( objectType_ ), objectHandle,
static_cast<VkPrivateDataSlot>( privateDataSlot ), &data );
21151 template <
typename Dispatch>
21153 PhysicalDevice::getVideoEncodeQualityLevelPropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceVideoEncodeQualityLevelInfoKHR * pQualityLevelInfo,
21154 VULKAN_HPP_NAMESPACE::VideoEncodeQualityLevelPropertiesKHR * pQualityLevelProperties,
21158 return static_cast<Result>(
21159 d.vkGetPhysicalDeviceVideoEncodeQualityLevelPropertiesKHR( m_physicalDevice,
21164 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21165 template <
typename Dispatch>
21167 PhysicalDevice::getVideoEncodeQualityLevelPropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceVideoEncodeQualityLevelInfoKHR & qualityLevelInfo,
21168 Dispatch
const & d )
const
21171 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21172 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceVideoEncodeQualityLevelPropertiesKHR &&
21173 "Function <vkGetPhysicalDeviceVideoEncodeQualityLevelPropertiesKHR> requires <VK_KHR_video_encode_queue>" );
21176 VULKAN_HPP_NAMESPACE::VideoEncodeQualityLevelPropertiesKHR qualityLevelProperties;
21178 d.vkGetPhysicalDeviceVideoEncodeQualityLevelPropertiesKHR( m_physicalDevice,
21186 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
21188 PhysicalDevice::getVideoEncodeQualityLevelPropertiesKHR(
const VULKAN_HPP_NAMESPACE::PhysicalDeviceVideoEncodeQualityLevelInfoKHR & qualityLevelInfo,
21189 Dispatch
const & d )
const
21192 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21193 VULKAN_HPP_ASSERT( d.vkGetPhysicalDeviceVideoEncodeQualityLevelPropertiesKHR &&
21194 "Function <vkGetPhysicalDeviceVideoEncodeQualityLevelPropertiesKHR> requires <VK_KHR_video_encode_queue>" );
21198 VULKAN_HPP_NAMESPACE::VideoEncodeQualityLevelPropertiesKHR & qualityLevelProperties =
21199 structureChain.template get<VULKAN_HPP_NAMESPACE::VideoEncodeQualityLevelPropertiesKHR>();
21201 d.vkGetPhysicalDeviceVideoEncodeQualityLevelPropertiesKHR( m_physicalDevice,
21210 template <
typename Dispatch>
21212 Device::getEncodedVideoSessionParametersKHR(
const VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersGetInfoKHR * pVideoSessionParametersInfo,
21213 VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersFeedbackInfoKHR * pFeedbackInfo,
21214 size_t * pDataSize,
21219 return static_cast<Result>(
21220 d.vkGetEncodedVideoSessionParametersKHR( m_device,
21227 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21228 template <
typename U
int8_tAllocator,
typename Dispatch>
21231 Device::getEncodedVideoSessionParametersKHR(
const VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersGetInfoKHR & videoSessionParametersInfo,
21232 Dispatch
const & d )
const
21235 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21236 VULKAN_HPP_ASSERT( d.vkGetEncodedVideoSessionParametersKHR &&
"Function <vkGetEncodedVideoSessionParametersKHR> requires <VK_KHR_video_encode_queue>" );
21239 std::pair<VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersFeedbackInfoKHR, std::vector<uint8_t, Uint8_tAllocator>> data_;
21240 VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersFeedbackInfoKHR & feedbackInfo = data_.first;
21241 std::vector<uint8_t, Uint8_tAllocator> & data = data_.second;
21247 d.vkGetEncodedVideoSessionParametersKHR( m_device,
21252 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
21254 data.resize( dataSize );
21256 d.vkGetEncodedVideoSessionParametersKHR( m_device,
21260 reinterpret_cast<void *
>( data.data() ) ) );
21262 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
21268 template <
typename Uint8_tAllocator,
21270 typename std::enable_if<std::is_same<typename Uint8_tAllocator::value_type, uint8_t>::value,
int>::type>
21273 Device::getEncodedVideoSessionParametersKHR(
const VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersGetInfoKHR & videoSessionParametersInfo,
21274 Uint8_tAllocator & uint8_tAllocator,
21275 Dispatch
const & d )
const
21278 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21279 VULKAN_HPP_ASSERT( d.vkGetEncodedVideoSessionParametersKHR &&
"Function <vkGetEncodedVideoSessionParametersKHR> requires <VK_KHR_video_encode_queue>" );
21282 std::pair<VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersFeedbackInfoKHR, std::vector<uint8_t, Uint8_tAllocator>> data_(
21283 std::piecewise_construct, std::forward_as_tuple(), std::forward_as_tuple( uint8_tAllocator ) );
21284 VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersFeedbackInfoKHR & feedbackInfo = data_.first;
21285 std::vector<uint8_t, Uint8_tAllocator> & data = data_.second;
21291 d.vkGetEncodedVideoSessionParametersKHR( m_device,
21296 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
21298 data.resize( dataSize );
21300 d.vkGetEncodedVideoSessionParametersKHR( m_device,
21304 reinterpret_cast<void *
>( data.data() ) ) );
21306 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
21312 template <
typename X,
typename Y,
typename... Z,
typename Uint8_tAllocator,
typename Dispatch>
21314 typename ResultValueType<std::pair<VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...>, std::vector<uint8_t, Uint8_tAllocator>>>::type
21315 Device::getEncodedVideoSessionParametersKHR(
const VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersGetInfoKHR & videoSessionParametersInfo,
21316 Dispatch
const & d )
const
21319 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21320 VULKAN_HPP_ASSERT( d.vkGetEncodedVideoSessionParametersKHR &&
"Function <vkGetEncodedVideoSessionParametersKHR> requires <VK_KHR_video_encode_queue>" );
21323 std::pair<VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...>, std::vector<uint8_t, Uint8_tAllocator>> data_;
21324 VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersFeedbackInfoKHR & feedbackInfo =
21325 data_.first.template get<VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersFeedbackInfoKHR>();
21326 std::vector<uint8_t, Uint8_tAllocator> & data = data_.second;
21332 d.vkGetEncodedVideoSessionParametersKHR( m_device,
21337 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
21339 data.resize( dataSize );
21341 d.vkGetEncodedVideoSessionParametersKHR( m_device,
21345 reinterpret_cast<void *
>( data.data() ) ) );
21347 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
21353 template <
typename X,
21356 typename Uint8_tAllocator,
21358 typename std::enable_if<std::is_same<typename Uint8_tAllocator::value_type, uint8_t>::value,
int>::type>
21360 typename ResultValueType<std::pair<VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...>, std::vector<uint8_t, Uint8_tAllocator>>>::type
21361 Device::getEncodedVideoSessionParametersKHR(
const VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersGetInfoKHR & videoSessionParametersInfo,
21362 Uint8_tAllocator & uint8_tAllocator,
21363 Dispatch
const & d )
const
21366 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21367 VULKAN_HPP_ASSERT( d.vkGetEncodedVideoSessionParametersKHR &&
"Function <vkGetEncodedVideoSessionParametersKHR> requires <VK_KHR_video_encode_queue>" );
21370 std::pair<VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...>, std::vector<uint8_t, Uint8_tAllocator>> data_(
21371 std::piecewise_construct, std::forward_as_tuple(), std::forward_as_tuple( uint8_tAllocator ) );
21372 VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersFeedbackInfoKHR & feedbackInfo =
21373 data_.first.template get<VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersFeedbackInfoKHR>();
21374 std::vector<uint8_t, Uint8_tAllocator> & data = data_.second;
21380 d.vkGetEncodedVideoSessionParametersKHR( m_device,
21385 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
21387 data.resize( dataSize );
21389 d.vkGetEncodedVideoSessionParametersKHR( m_device,
21393 reinterpret_cast<void *
>( data.data() ) ) );
21395 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
21402 template <
typename Dispatch>
21403 VULKAN_HPP_INLINE void CommandBuffer::encodeVideoKHR(
const VULKAN_HPP_NAMESPACE::VideoEncodeInfoKHR * pEncodeInfo,
21407 d.vkCmdEncodeVideoKHR( m_commandBuffer,
reinterpret_cast<const VkVideoEncodeInfoKHR *
>( pEncodeInfo ) );
21410 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21411 template <
typename Dispatch>
21412 VULKAN_HPP_INLINE void CommandBuffer::encodeVideoKHR(
const VULKAN_HPP_NAMESPACE::VideoEncodeInfoKHR & encodeInfo,
21416 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21417 VULKAN_HPP_ASSERT( d.vkCmdEncodeVideoKHR &&
"Function <vkCmdEncodeVideoKHR> requires <VK_KHR_video_encode_queue>" );
21420 d.vkCmdEncodeVideoKHR( m_commandBuffer,
reinterpret_cast<const VkVideoEncodeInfoKHR *
>( &encodeInfo ) );
21424 #if defined( VK_ENABLE_BETA_EXTENSIONS )
21427 template <
typename Dispatch>
21429 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
21430 VULKAN_HPP_NAMESPACE::CudaModuleNV * pModule,
21434 return static_cast<Result>( d.vkCreateCudaModuleNV( m_device,
21440 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21441 template <
typename Dispatch>
21443 Device::createCudaModuleNV(
const VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV & createInfo,
21445 Dispatch
const & d )
const
21448 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21449 VULKAN_HPP_ASSERT( d.vkCreateCudaModuleNV &&
"Function <vkCreateCudaModuleNV> requires <VK_NV_cuda_kernel_launch>" );
21452 VULKAN_HPP_NAMESPACE::CudaModuleNV
module;
21454 d.vkCreateCudaModuleNV( m_device,
21456 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
21463 # ifndef VULKAN_HPP_NO_SMART_HANDLE
21464 template <
typename Dispatch>
21466 Device::createCudaModuleNVUnique(
const VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV & createInfo,
21468 Dispatch
const & d )
const
21471 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21472 VULKAN_HPP_ASSERT( d.vkCreateCudaModuleNV &&
"Function <vkCreateCudaModuleNV> requires <VK_NV_cuda_kernel_launch>" );
21475 VULKAN_HPP_NAMESPACE::CudaModuleNV
module;
21477 d.vkCreateCudaModuleNV( m_device,
21479 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
21489 template <
typename Dispatch>
21491 size_t * pCacheSize,
21496 return static_cast<Result>( d.vkGetCudaModuleCacheNV( m_device,
static_cast<VkCudaModuleNV>(
module ), pCacheSize, pCacheData ) );
21499 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21500 template <
typename U
int8_tAllocator,
typename Dispatch>
21502 Device::getCudaModuleCacheNV( VULKAN_HPP_NAMESPACE::CudaModuleNV
module, Dispatch
const & d )
const
21505 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21506 VULKAN_HPP_ASSERT( d.vkGetCudaModuleCacheNV &&
"Function <vkGetCudaModuleCacheNV> requires <VK_NV_cuda_kernel_launch>" );
21509 std::vector<uint8_t, Uint8_tAllocator> cacheData;
21515 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && cacheSize )
21517 cacheData.resize( cacheSize );
21519 d.vkGetCudaModuleCacheNV( m_device,
static_cast<VkCudaModuleNV>(
module ), &cacheSize,
reinterpret_cast<void *
>( cacheData.data() ) ) );
21521 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
21524 if ( cacheSize < cacheData.size() )
21526 cacheData.resize( cacheSize );
21531 template <
typename Uint8_tAllocator,
21533 typename std::enable_if<std::is_same<typename Uint8_tAllocator::value_type, uint8_t>::value,
int>::type>
21535 Device::getCudaModuleCacheNV( VULKAN_HPP_NAMESPACE::CudaModuleNV
module, Uint8_tAllocator & uint8_tAllocator, Dispatch
const & d )
const
21538 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21539 VULKAN_HPP_ASSERT( d.vkGetCudaModuleCacheNV &&
"Function <vkGetCudaModuleCacheNV> requires <VK_NV_cuda_kernel_launch>" );
21542 std::vector<uint8_t, Uint8_tAllocator> cacheData( uint8_tAllocator );
21548 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && cacheSize )
21550 cacheData.resize( cacheSize );
21552 d.vkGetCudaModuleCacheNV( m_device,
static_cast<VkCudaModuleNV>(
module ), &cacheSize,
reinterpret_cast<void *
>( cacheData.data() ) ) );
21554 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
21557 if ( cacheSize < cacheData.size() )
21559 cacheData.resize( cacheSize );
21565 template <
typename Dispatch>
21567 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
21568 VULKAN_HPP_NAMESPACE::CudaFunctionNV * pFunction,
21572 return static_cast<Result>( d.vkCreateCudaFunctionNV( m_device,
21578 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21579 template <
typename Dispatch>
21581 Device::createCudaFunctionNV(
const VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV & createInfo,
21583 Dispatch
const & d )
const
21586 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21587 VULKAN_HPP_ASSERT( d.vkCreateCudaFunctionNV &&
"Function <vkCreateCudaFunctionNV> requires <VK_NV_cuda_kernel_launch>" );
21590 VULKAN_HPP_NAMESPACE::CudaFunctionNV
function;
21592 d.vkCreateCudaFunctionNV( m_device,
21594 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
21601 # ifndef VULKAN_HPP_NO_SMART_HANDLE
21602 template <
typename Dispatch>
21604 Device::createCudaFunctionNVUnique(
const VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV & createInfo,
21606 Dispatch
const & d )
const
21609 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21610 VULKAN_HPP_ASSERT( d.vkCreateCudaFunctionNV &&
"Function <vkCreateCudaFunctionNV> requires <VK_NV_cuda_kernel_launch>" );
21613 VULKAN_HPP_NAMESPACE::CudaFunctionNV
function;
21615 d.vkCreateCudaFunctionNV( m_device,
21617 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
21627 template <
typename Dispatch>
21629 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
21636 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21637 template <
typename Dispatch>
21643 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21644 VULKAN_HPP_ASSERT( d.vkDestroyCudaModuleNV &&
"Function <vkDestroyCudaModuleNV> requires <VK_NV_cuda_kernel_launch>" );
21647 d.vkDestroyCudaModuleNV( m_device,
21649 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
21653 template <
typename Dispatch>
21655 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
21662 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21663 template <
typename Dispatch>
21669 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21670 VULKAN_HPP_ASSERT( d.vkDestroyCudaModuleNV &&
"Function <vkDestroyCudaModuleNV> requires <VK_NV_cuda_kernel_launch>" );
21673 d.vkDestroyCudaModuleNV( m_device,
21675 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
21679 template <
typename Dispatch>
21681 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
21688 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21689 template <
typename Dispatch>
21695 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21696 VULKAN_HPP_ASSERT( d.vkDestroyCudaFunctionNV &&
"Function <vkDestroyCudaFunctionNV> requires <VK_NV_cuda_kernel_launch>" );
21699 d.vkDestroyCudaFunctionNV( m_device,
21701 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
21705 template <
typename Dispatch>
21707 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
21714 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21715 template <
typename Dispatch>
21721 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21722 VULKAN_HPP_ASSERT( d.vkDestroyCudaFunctionNV &&
"Function <vkDestroyCudaFunctionNV> requires <VK_NV_cuda_kernel_launch>" );
21725 d.vkDestroyCudaFunctionNV( m_device,
21727 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
21731 template <
typename Dispatch>
21732 VULKAN_HPP_INLINE void CommandBuffer::cudaLaunchKernelNV(
const VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV * pLaunchInfo,
21736 d.vkCmdCudaLaunchKernelNV( m_commandBuffer,
reinterpret_cast<const VkCudaLaunchInfoNV *
>( pLaunchInfo ) );
21739 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21740 template <
typename Dispatch>
21741 VULKAN_HPP_INLINE void CommandBuffer::cudaLaunchKernelNV(
const VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV & launchInfo,
21745 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21746 VULKAN_HPP_ASSERT( d.vkCmdCudaLaunchKernelNV &&
"Function <vkCmdCudaLaunchKernelNV> requires <VK_NV_cuda_kernel_launch>" );
21749 d.vkCmdCudaLaunchKernelNV( m_commandBuffer,
reinterpret_cast<const VkCudaLaunchInfoNV *
>( &launchInfo ) );
21754 #if defined( VK_USE_PLATFORM_METAL_EXT )
21757 template <
typename Dispatch>
21758 VULKAN_HPP_INLINE void Device::exportMetalObjectsEXT( VULKAN_HPP_NAMESPACE::ExportMetalObjectsInfoEXT * pMetalObjectsInfo,
21762 d.vkExportMetalObjectsEXT( m_device,
reinterpret_cast<VkExportMetalObjectsInfoEXT *
>( pMetalObjectsInfo ) );
21765 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21766 template <
typename Dispatch>
21771 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21772 VULKAN_HPP_ASSERT( d.vkExportMetalObjectsEXT &&
"Function <vkExportMetalObjectsEXT> requires <VK_EXT_metal_objects>" );
21775 VULKAN_HPP_NAMESPACE::ExportMetalObjectsInfoEXT metalObjectsInfo;
21776 d.vkExportMetalObjectsEXT( m_device,
reinterpret_cast<VkExportMetalObjectsInfoEXT *
>( &metalObjectsInfo ) );
21778 return metalObjectsInfo;
21781 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
21786 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21787 VULKAN_HPP_ASSERT( d.vkExportMetalObjectsEXT &&
"Function <vkExportMetalObjectsEXT> requires <VK_EXT_metal_objects>" );
21790 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
21791 VULKAN_HPP_NAMESPACE::ExportMetalObjectsInfoEXT & metalObjectsInfo = structureChain.template get<VULKAN_HPP_NAMESPACE::ExportMetalObjectsInfoEXT>();
21792 d.vkExportMetalObjectsEXT( m_device,
reinterpret_cast<VkExportMetalObjectsInfoEXT *
>( &metalObjectsInfo ) );
21794 return structureChain;
21801 template <
typename Dispatch>
21803 const VULKAN_HPP_NAMESPACE::DependencyInfo * pDependencyInfo,
21807 d.vkCmdSetEvent2KHR( m_commandBuffer,
static_cast<VkEvent>( event ),
reinterpret_cast<const VkDependencyInfo *
>( pDependencyInfo ) );
21810 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21811 template <
typename Dispatch>
21813 const VULKAN_HPP_NAMESPACE::DependencyInfo & dependencyInfo,
21817 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21818 VULKAN_HPP_ASSERT( d.vkCmdSetEvent2KHR &&
"Function <vkCmdSetEvent2KHR> requires <VK_KHR_synchronization2> or <VK_VERSION_1_3>" );
21821 d.vkCmdSetEvent2KHR( m_commandBuffer,
static_cast<VkEvent>( event ),
reinterpret_cast<const VkDependencyInfo *
>( &dependencyInfo ) );
21825 template <
typename Dispatch>
21834 template <
typename Dispatch>
21836 const VULKAN_HPP_NAMESPACE::Event * pEvents,
21837 const VULKAN_HPP_NAMESPACE::DependencyInfo * pDependencyInfos,
21841 d.vkCmdWaitEvents2KHR(
21842 m_commandBuffer, eventCount,
reinterpret_cast<const VkEvent *
>( pEvents ),
reinterpret_cast<const VkDependencyInfo *
>( pDependencyInfos ) );
21845 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21846 template <
typename Dispatch>
21847 VULKAN_HPP_INLINE void CommandBuffer::waitEvents2KHR( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Event>
const & events,
21848 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DependencyInfo>
const & dependencyInfos,
21852 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21853 VULKAN_HPP_ASSERT( d.vkCmdWaitEvents2KHR &&
"Function <vkCmdWaitEvents2KHR> requires <VK_KHR_synchronization2> or <VK_VERSION_1_3>" );
21855 # ifdef VULKAN_HPP_NO_EXCEPTIONS
21858 if ( events.size() != dependencyInfos.size() )
21864 d.vkCmdWaitEvents2KHR( m_commandBuffer,
21866 reinterpret_cast<const VkEvent *
>( events.data() ),
21871 template <
typename Dispatch>
21872 VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier2KHR(
const VULKAN_HPP_NAMESPACE::DependencyInfo * pDependencyInfo,
21876 d.vkCmdPipelineBarrier2KHR( m_commandBuffer,
reinterpret_cast<const VkDependencyInfo *
>( pDependencyInfo ) );
21879 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21880 template <
typename Dispatch>
21881 VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier2KHR(
const VULKAN_HPP_NAMESPACE::DependencyInfo & dependencyInfo,
21885 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21886 VULKAN_HPP_ASSERT( d.vkCmdPipelineBarrier2KHR &&
"Function <vkCmdPipelineBarrier2KHR> requires <VK_KHR_synchronization2> or <VK_VERSION_1_3>" );
21889 d.vkCmdPipelineBarrier2KHR( m_commandBuffer,
reinterpret_cast<const VkDependencyInfo *
>( &dependencyInfo ) );
21893 template <
typename Dispatch>
21895 VULKAN_HPP_NAMESPACE::QueryPool queryPool,
21903 template <
typename Dispatch>
21905 const VULKAN_HPP_NAMESPACE::SubmitInfo2 * pSubmits,
21906 VULKAN_HPP_NAMESPACE::Fence fence,
21910 return static_cast<Result>(
21911 d.vkQueueSubmit2KHR( m_queue, submitCount,
reinterpret_cast<const VkSubmitInfo2 *
>( pSubmits ),
static_cast<VkFence>( fence ) ) );
21914 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21915 template <
typename Dispatch>
21917 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SubmitInfo2>
const & submits, VULKAN_HPP_NAMESPACE::Fence fence, Dispatch
const & d )
const
21920 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21921 VULKAN_HPP_ASSERT( d.vkQueueSubmit2KHR &&
"Function <vkQueueSubmit2KHR> requires <VK_KHR_synchronization2> or <VK_VERSION_1_3>" );
21925 d.vkQueueSubmit2KHR( m_queue, submits.size(),
reinterpret_cast<const VkSubmitInfo2 *
>( submits.data() ),
static_cast<VkFence>( fence ) ) );
21932 template <
typename Dispatch>
21934 VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
21940 d.vkCmdWriteBufferMarker2AMD(
21944 template <
typename Dispatch>
21946 VULKAN_HPP_NAMESPACE::CheckpointData2NV * pCheckpointData,
21950 d.vkGetQueueCheckpointData2NV( m_queue, pCheckpointDataCount,
reinterpret_cast<VkCheckpointData2NV *
>( pCheckpointData ) );
21953 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
21954 template <
typename Checkpo
intData2NVAllocator,
typename Dispatch>
21956 Queue::getCheckpointData2NV( Dispatch
const & d )
const
21959 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21960 VULKAN_HPP_ASSERT( d.vkGetQueueCheckpointData2NV &&
"Function <vkGetQueueCheckpointData2NV> requires <VK_KHR_synchronization2>" );
21963 std::vector<VULKAN_HPP_NAMESPACE::CheckpointData2NV, CheckpointData2NVAllocator> checkpointData;
21964 uint32_t checkpointDataCount;
21965 d.vkGetQueueCheckpointData2NV( m_queue, &checkpointDataCount,
nullptr );
21966 checkpointData.resize( checkpointDataCount );
21967 d.vkGetQueueCheckpointData2NV( m_queue, &checkpointDataCount,
reinterpret_cast<VkCheckpointData2NV *
>( checkpointData.data() ) );
21970 if ( checkpointDataCount < checkpointData.size() )
21972 checkpointData.resize( checkpointDataCount );
21974 return checkpointData;
21977 template <
typename CheckpointData2NVAllocator,
21979 typename std::enable_if<std::is_same<typename CheckpointData2NVAllocator::value_type, VULKAN_HPP_NAMESPACE::CheckpointData2NV>::value,
int>::type>
21981 Queue::getCheckpointData2NV( CheckpointData2NVAllocator & checkpointData2NVAllocator, Dispatch
const & d )
const
21984 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
21985 VULKAN_HPP_ASSERT( d.vkGetQueueCheckpointData2NV &&
"Function <vkGetQueueCheckpointData2NV> requires <VK_KHR_synchronization2>" );
21988 std::vector<VULKAN_HPP_NAMESPACE::CheckpointData2NV, CheckpointData2NVAllocator> checkpointData( checkpointData2NVAllocator );
21989 uint32_t checkpointDataCount;
21990 d.vkGetQueueCheckpointData2NV( m_queue, &checkpointDataCount,
nullptr );
21991 checkpointData.resize( checkpointDataCount );
21992 d.vkGetQueueCheckpointData2NV( m_queue, &checkpointDataCount,
reinterpret_cast<VkCheckpointData2NV *
>( checkpointData.data() ) );
21995 if ( checkpointDataCount < checkpointData.size() )
21997 checkpointData.resize( checkpointDataCount );
21999 return checkpointData;
22005 template <
typename Dispatch>
22006 VULKAN_HPP_INLINE void Device::getDescriptorSetLayoutSizeEXT( VULKAN_HPP_NAMESPACE::DescriptorSetLayout layout,
22014 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22015 template <
typename Dispatch>
22017 Device::getDescriptorSetLayoutSizeEXT( VULKAN_HPP_NAMESPACE::DescriptorSetLayout layout, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22020 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22021 VULKAN_HPP_ASSERT( d.vkGetDescriptorSetLayoutSizeEXT &&
"Function <vkGetDescriptorSetLayoutSizeEXT> requires <VK_EXT_descriptor_buffer>" );
22027 return layoutSizeInBytes;
22031 template <
typename Dispatch>
22032 VULKAN_HPP_INLINE void Device::getDescriptorSetLayoutBindingOffsetEXT( VULKAN_HPP_NAMESPACE::DescriptorSetLayout layout,
22041 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22042 template <
typename Dispatch>
22044 VULKAN_HPP_NAMESPACE::DescriptorSetLayout layout, uint32_t binding, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22047 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22049 "Function <vkGetDescriptorSetLayoutBindingOffsetEXT> requires <VK_EXT_descriptor_buffer>" );
22059 template <
typename Dispatch>
22060 VULKAN_HPP_INLINE void Device::getDescriptorEXT(
const VULKAN_HPP_NAMESPACE::DescriptorGetInfoEXT * pDescriptorInfo,
22062 void * pDescriptor,
22066 d.vkGetDescriptorEXT( m_device,
reinterpret_cast<const VkDescriptorGetInfoEXT *
>( pDescriptorInfo ), dataSize, pDescriptor );
22069 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22070 template <
typename Dispatch>
22071 VULKAN_HPP_INLINE void Device::getDescriptorEXT(
const VULKAN_HPP_NAMESPACE::DescriptorGetInfoEXT & descriptorInfo,
22073 void * pDescriptor,
22077 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22078 VULKAN_HPP_ASSERT( d.vkGetDescriptorEXT &&
"Function <vkGetDescriptorEXT> requires <VK_EXT_descriptor_buffer>" );
22081 d.vkGetDescriptorEXT( m_device,
reinterpret_cast<const VkDescriptorGetInfoEXT *
>( &descriptorInfo ), dataSize, pDescriptor );
22084 template <
typename DescriptorType,
typename Dispatch>
22089 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22090 VULKAN_HPP_ASSERT( d.vkGetDescriptorEXT &&
"Function <vkGetDescriptorEXT> requires <VK_EXT_descriptor_buffer>" );
22094 d.vkGetDescriptorEXT(
22101 template <
typename Dispatch>
22103 const VULKAN_HPP_NAMESPACE::DescriptorBufferBindingInfoEXT * pBindingInfos,
22110 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22111 template <
typename Dispatch>
22113 CommandBuffer::bindDescriptorBuffersEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DescriptorBufferBindingInfoEXT>
const & bindingInfos,
22117 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22118 VULKAN_HPP_ASSERT( d.vkCmdBindDescriptorBuffersEXT &&
"Function <vkCmdBindDescriptorBuffersEXT> requires <VK_EXT_descriptor_buffer>" );
22121 d.vkCmdBindDescriptorBuffersEXT( m_commandBuffer, bindingInfos.size(),
reinterpret_cast<const VkDescriptorBufferBindingInfoEXT *
>( bindingInfos.data() ) );
22125 template <
typename Dispatch>
22127 VULKAN_HPP_NAMESPACE::PipelineLayout layout,
22130 const uint32_t * pBufferIndices,
22135 d.vkCmdSetDescriptorBufferOffsetsEXT( m_commandBuffer,
22144 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22145 template <
typename Dispatch>
22147 VULKAN_HPP_NAMESPACE::PipelineLayout layout,
22149 VULKAN_HPP_NAMESPACE::ArrayProxy<const uint32_t>
const & bufferIndices,
22150 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize>
const & offsets,
22154 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22155 VULKAN_HPP_ASSERT( d.vkCmdSetDescriptorBufferOffsetsEXT &&
"Function <vkCmdSetDescriptorBufferOffsetsEXT> requires <VK_EXT_descriptor_buffer>" );
22157 # ifdef VULKAN_HPP_NO_EXCEPTIONS
22160 if ( bufferIndices.size() != offsets.size() )
22166 d.vkCmdSetDescriptorBufferOffsetsEXT( m_commandBuffer,
22170 bufferIndices.size(),
22171 bufferIndices.data(),
22172 reinterpret_cast<const VkDeviceSize *
>( offsets.data() ) );
22176 template <
typename Dispatch>
22178 VULKAN_HPP_NAMESPACE::PipelineLayout layout,
22183 d.vkCmdBindDescriptorBufferEmbeddedSamplersEXT(
22187 template <
typename Dispatch>
22189 const VULKAN_HPP_NAMESPACE::BufferCaptureDescriptorDataInfoEXT * pInfo,
void * pData, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22192 return static_cast<Result>(
22196 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22197 template <
typename DataType,
typename Dispatch>
22199 Device::getBufferOpaqueCaptureDescriptorDataEXT(
const VULKAN_HPP_NAMESPACE::BufferCaptureDescriptorDataInfoEXT & info, Dispatch
const & d )
const
22202 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22204 "Function <vkGetBufferOpaqueCaptureDescriptorDataEXT> requires <VK_EXT_descriptor_buffer>" );
22216 template <
typename Dispatch>
22218 const VULKAN_HPP_NAMESPACE::ImageCaptureDescriptorDataInfoEXT * pInfo,
void * pData, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22221 return static_cast<Result>(
22225 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22226 template <
typename DataType,
typename Dispatch>
22228 Device::getImageOpaqueCaptureDescriptorDataEXT(
const VULKAN_HPP_NAMESPACE::ImageCaptureDescriptorDataInfoEXT & info, Dispatch
const & d )
const
22231 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22233 "Function <vkGetImageOpaqueCaptureDescriptorDataEXT> requires <VK_EXT_descriptor_buffer>" );
22245 template <
typename Dispatch>
22247 const VULKAN_HPP_NAMESPACE::ImageViewCaptureDescriptorDataInfoEXT * pInfo,
void * pData, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22250 return static_cast<Result>(
22254 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22255 template <
typename DataType,
typename Dispatch>
22257 Device::getImageViewOpaqueCaptureDescriptorDataEXT(
const VULKAN_HPP_NAMESPACE::ImageViewCaptureDescriptorDataInfoEXT & info, Dispatch
const & d )
const
22260 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22262 "Function <vkGetImageViewOpaqueCaptureDescriptorDataEXT> requires <VK_EXT_descriptor_buffer>" );
22274 template <
typename Dispatch>
22276 const VULKAN_HPP_NAMESPACE::SamplerCaptureDescriptorDataInfoEXT * pInfo,
void * pData, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22279 return static_cast<Result>(
22283 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22284 template <
typename DataType,
typename Dispatch>
22286 Device::getSamplerOpaqueCaptureDescriptorDataEXT(
const VULKAN_HPP_NAMESPACE::SamplerCaptureDescriptorDataInfoEXT & info, Dispatch
const & d )
const
22289 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22291 "Function <vkGetSamplerOpaqueCaptureDescriptorDataEXT> requires <VK_EXT_descriptor_buffer>" );
22303 template <
typename Dispatch>
22305 const VULKAN_HPP_NAMESPACE::AccelerationStructureCaptureDescriptorDataInfoEXT * pInfo,
void * pData, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22308 return static_cast<Result>( d.vkGetAccelerationStructureOpaqueCaptureDescriptorDataEXT(
22312 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22313 template <
typename DataType,
typename Dispatch>
22315 Device::getAccelerationStructureOpaqueCaptureDescriptorDataEXT(
const VULKAN_HPP_NAMESPACE::AccelerationStructureCaptureDescriptorDataInfoEXT & info,
22316 Dispatch
const & d )
const
22319 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22320 VULKAN_HPP_ASSERT( d.vkGetAccelerationStructureOpaqueCaptureDescriptorDataEXT &&
22321 "Function <vkGetAccelerationStructureOpaqueCaptureDescriptorDataEXT> requires <VK_EXT_descriptor_buffer>" );
22335 template <
typename Dispatch>
22341 d.vkCmdSetFragmentShadingRateEnumNV(
22347 template <
typename Dispatch>
22349 CommandBuffer::drawMeshTasksEXT( uint32_t groupCountX, uint32_t groupCountY, uint32_t groupCountZ, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22352 d.vkCmdDrawMeshTasksEXT( m_commandBuffer, groupCountX, groupCountY, groupCountZ );
22355 template <
typename Dispatch>
22358 uint32_t drawCount,
22363 d.vkCmdDrawMeshTasksIndirectEXT( m_commandBuffer,
static_cast<VkBuffer>( buffer ),
static_cast<VkDeviceSize>( offset ), drawCount, stride );
22366 template <
typename Dispatch>
22367 VULKAN_HPP_INLINE void CommandBuffer::drawMeshTasksIndirectCountEXT( VULKAN_HPP_NAMESPACE::Buffer buffer,
22369 VULKAN_HPP_NAMESPACE::Buffer countBuffer,
22371 uint32_t maxDrawCount,
22376 d.vkCmdDrawMeshTasksIndirectCountEXT( m_commandBuffer,
22379 static_cast<VkBuffer>( countBuffer ),
22387 template <
typename Dispatch>
22388 VULKAN_HPP_INLINE void CommandBuffer::copyBuffer2KHR(
const VULKAN_HPP_NAMESPACE::CopyBufferInfo2 * pCopyBufferInfo,
22392 d.vkCmdCopyBuffer2KHR( m_commandBuffer,
reinterpret_cast<const VkCopyBufferInfo2 *
>( pCopyBufferInfo ) );
22395 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22396 template <
typename Dispatch>
22397 VULKAN_HPP_INLINE void CommandBuffer::copyBuffer2KHR(
const VULKAN_HPP_NAMESPACE::CopyBufferInfo2 & copyBufferInfo,
22401 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22402 VULKAN_HPP_ASSERT( d.vkCmdCopyBuffer2KHR &&
"Function <vkCmdCopyBuffer2KHR> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
22405 d.vkCmdCopyBuffer2KHR( m_commandBuffer,
reinterpret_cast<const VkCopyBufferInfo2 *
>( ©BufferInfo ) );
22409 template <
typename Dispatch>
22410 VULKAN_HPP_INLINE void CommandBuffer::copyImage2KHR(
const VULKAN_HPP_NAMESPACE::CopyImageInfo2 * pCopyImageInfo,
22414 d.vkCmdCopyImage2KHR( m_commandBuffer,
reinterpret_cast<const VkCopyImageInfo2 *
>( pCopyImageInfo ) );
22417 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22418 template <
typename Dispatch>
22419 VULKAN_HPP_INLINE void CommandBuffer::copyImage2KHR(
const VULKAN_HPP_NAMESPACE::CopyImageInfo2 & copyImageInfo,
22423 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22424 VULKAN_HPP_ASSERT( d.vkCmdCopyImage2KHR &&
"Function <vkCmdCopyImage2KHR> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
22427 d.vkCmdCopyImage2KHR( m_commandBuffer,
reinterpret_cast<const VkCopyImageInfo2 *
>( ©ImageInfo ) );
22431 template <
typename Dispatch>
22432 VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage2KHR(
const VULKAN_HPP_NAMESPACE::CopyBufferToImageInfo2 * pCopyBufferToImageInfo,
22436 d.vkCmdCopyBufferToImage2KHR( m_commandBuffer,
reinterpret_cast<const VkCopyBufferToImageInfo2 *
>( pCopyBufferToImageInfo ) );
22439 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22440 template <
typename Dispatch>
22441 VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage2KHR(
const VULKAN_HPP_NAMESPACE::CopyBufferToImageInfo2 & copyBufferToImageInfo,
22445 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22446 VULKAN_HPP_ASSERT( d.vkCmdCopyBufferToImage2KHR &&
"Function <vkCmdCopyBufferToImage2KHR> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
22449 d.vkCmdCopyBufferToImage2KHR( m_commandBuffer,
reinterpret_cast<const VkCopyBufferToImageInfo2 *
>( ©BufferToImageInfo ) );
22453 template <
typename Dispatch>
22454 VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer2KHR(
const VULKAN_HPP_NAMESPACE::CopyImageToBufferInfo2 * pCopyImageToBufferInfo,
22458 d.vkCmdCopyImageToBuffer2KHR( m_commandBuffer,
reinterpret_cast<const VkCopyImageToBufferInfo2 *
>( pCopyImageToBufferInfo ) );
22461 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22462 template <
typename Dispatch>
22463 VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer2KHR(
const VULKAN_HPP_NAMESPACE::CopyImageToBufferInfo2 & copyImageToBufferInfo,
22467 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22468 VULKAN_HPP_ASSERT( d.vkCmdCopyImageToBuffer2KHR &&
"Function <vkCmdCopyImageToBuffer2KHR> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
22471 d.vkCmdCopyImageToBuffer2KHR( m_commandBuffer,
reinterpret_cast<const VkCopyImageToBufferInfo2 *
>( ©ImageToBufferInfo ) );
22475 template <
typename Dispatch>
22476 VULKAN_HPP_INLINE void CommandBuffer::blitImage2KHR(
const VULKAN_HPP_NAMESPACE::BlitImageInfo2 * pBlitImageInfo,
22480 d.vkCmdBlitImage2KHR( m_commandBuffer,
reinterpret_cast<const VkBlitImageInfo2 *
>( pBlitImageInfo ) );
22483 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22484 template <
typename Dispatch>
22485 VULKAN_HPP_INLINE void CommandBuffer::blitImage2KHR(
const VULKAN_HPP_NAMESPACE::BlitImageInfo2 & blitImageInfo,
22489 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22490 VULKAN_HPP_ASSERT( d.vkCmdBlitImage2KHR &&
"Function <vkCmdBlitImage2KHR> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
22493 d.vkCmdBlitImage2KHR( m_commandBuffer,
reinterpret_cast<const VkBlitImageInfo2 *
>( &blitImageInfo ) );
22497 template <
typename Dispatch>
22498 VULKAN_HPP_INLINE void CommandBuffer::resolveImage2KHR(
const VULKAN_HPP_NAMESPACE::ResolveImageInfo2 * pResolveImageInfo,
22502 d.vkCmdResolveImage2KHR( m_commandBuffer,
reinterpret_cast<const VkResolveImageInfo2 *
>( pResolveImageInfo ) );
22505 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22506 template <
typename Dispatch>
22507 VULKAN_HPP_INLINE void CommandBuffer::resolveImage2KHR(
const VULKAN_HPP_NAMESPACE::ResolveImageInfo2 & resolveImageInfo,
22511 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22512 VULKAN_HPP_ASSERT( d.vkCmdResolveImage2KHR &&
"Function <vkCmdResolveImage2KHR> requires <VK_KHR_copy_commands2> or <VK_VERSION_1_3>" );
22515 d.vkCmdResolveImage2KHR( m_commandBuffer,
reinterpret_cast<const VkResolveImageInfo2 *
>( &resolveImageInfo ) );
22521 template <
typename Dispatch>
22523 VULKAN_HPP_NAMESPACE::DeviceFaultInfoEXT * pFaultInfo,
22527 return static_cast<Result>( d.vkGetDeviceFaultInfoEXT(
22531 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22532 template <
typename Dispatch>
22534 Device::getFaultInfoEXT( Dispatch
const & d )
const
22537 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22538 VULKAN_HPP_ASSERT( d.vkGetDeviceFaultInfoEXT &&
"Function <vkGetDeviceFaultInfoEXT> requires <VK_EXT_device_fault>" );
22541 std::pair<VULKAN_HPP_NAMESPACE::DeviceFaultCountsEXT, VULKAN_HPP_NAMESPACE::DeviceFaultInfoEXT> data_;
22542 VULKAN_HPP_NAMESPACE::DeviceFaultCountsEXT & faultCounts = data_.first;
22543 VULKAN_HPP_NAMESPACE::DeviceFaultInfoEXT & faultInfo = data_.second;
22547 result,
VULKAN_HPP_NAMESPACE_STRING "::Device::getFaultInfoEXT", { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eIncomplete } );
22554 #if defined( VK_USE_PLATFORM_WIN32_KHR )
22557 # ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
22558 template <
typename Dispatch>
22563 return static_cast<Result>( d.vkAcquireWinrtDisplayNV( m_physicalDevice,
static_cast<VkDisplayKHR>( display ) ) );
22566 template <
typename Dispatch>
22568 PhysicalDevice::acquireWinrtDisplayNV( VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch
const & d )
const
22571 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22572 VULKAN_HPP_ASSERT( d.vkAcquireWinrtDisplayNV &&
"Function <vkAcquireWinrtDisplayNV> requires <VK_NV_acquire_winrt_display>" );
22583 template <
typename Dispatch>
22585 VULKAN_HPP_NAMESPACE::DisplayKHR * pDisplay,
22589 return static_cast<Result>( d.vkGetWinrtDisplayNV( m_physicalDevice, deviceRelativeId,
reinterpret_cast<VkDisplayKHR *
>( pDisplay ) ) );
22592 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22593 template <
typename Dispatch>
22595 PhysicalDevice::getWinrtDisplayNV( uint32_t deviceRelativeId, Dispatch
const & d )
const
22598 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22599 VULKAN_HPP_ASSERT( d.vkGetWinrtDisplayNV &&
"Function <vkGetWinrtDisplayNV> requires <VK_NV_acquire_winrt_display>" );
22602 VULKAN_HPP_NAMESPACE::DisplayKHR display;
22610 # ifndef VULKAN_HPP_NO_SMART_HANDLE
22611 template <
typename Dispatch>
22613 PhysicalDevice::getWinrtDisplayNVUnique( uint32_t deviceRelativeId, Dispatch
const & d )
const
22616 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22617 VULKAN_HPP_ASSERT( d.vkGetWinrtDisplayNV &&
"Function <vkGetWinrtDisplayNV> requires <VK_NV_acquire_winrt_display>" );
22620 VULKAN_HPP_NAMESPACE::DisplayKHR display;
22626 UniqueHandle<VULKAN_HPP_NAMESPACE::DisplayKHR, Dispatch>( display, ObjectRelease<PhysicalDevice, Dispatch>( *
this, d ) ) );
22632 #if defined( VK_USE_PLATFORM_DIRECTFB_EXT )
22635 template <
typename Dispatch>
22637 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
22638 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
22642 return static_cast<Result>( d.vkCreateDirectFBSurfaceEXT( m_instance,
22643 reinterpret_cast<const VkDirectFBSurfaceCreateInfoEXT *
>( pCreateInfo ),
22648 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22649 template <
typename Dispatch>
22651 Instance::createDirectFBSurfaceEXT(
const VULKAN_HPP_NAMESPACE::DirectFBSurfaceCreateInfoEXT & createInfo,
22652 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
22653 Dispatch
const & d )
const
22656 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22657 VULKAN_HPP_ASSERT( d.vkCreateDirectFBSurfaceEXT &&
"Function <vkCreateDirectFBSurfaceEXT> requires <VK_EXT_directfb_surface>" );
22660 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
22663 reinterpret_cast<const VkDirectFBSurfaceCreateInfoEXT *
>( &createInfo ),
22664 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
22671 # ifndef VULKAN_HPP_NO_SMART_HANDLE
22672 template <
typename Dispatch>
22674 Instance::createDirectFBSurfaceEXTUnique(
const VULKAN_HPP_NAMESPACE::DirectFBSurfaceCreateInfoEXT & createInfo,
22675 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
22676 Dispatch
const & d )
const
22679 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22680 VULKAN_HPP_ASSERT( d.vkCreateDirectFBSurfaceEXT &&
"Function <vkCreateDirectFBSurfaceEXT> requires <VK_EXT_directfb_surface>" );
22683 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
22686 reinterpret_cast<const VkDirectFBSurfaceCreateInfoEXT *
>( &createInfo ),
22687 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
22692 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
22697 template <
typename Dispatch>
22703 return static_cast<Bool32>( d.vkGetPhysicalDeviceDirectFBPresentationSupportEXT( m_physicalDevice, queueFamilyIndex, dfb ) );
22706 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22707 template <
typename Dispatch>
22709 PhysicalDevice::getDirectFBPresentationSupportEXT( uint32_t queueFamilyIndex, IDirectFB & dfb, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22712 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22714 "Function <vkGetPhysicalDeviceDirectFBPresentationSupportEXT> requires <VK_EXT_directfb_surface>" );
22717 VkBool32 result = d.vkGetPhysicalDeviceDirectFBPresentationSupportEXT( m_physicalDevice, queueFamilyIndex, &dfb );
22726 template <
typename Dispatch>
22728 const VULKAN_HPP_NAMESPACE::VertexInputBindingDescription2EXT * pVertexBindingDescriptions,
22729 uint32_t vertexAttributeDescriptionCount,
22730 const VULKAN_HPP_NAMESPACE::VertexInputAttributeDescription2EXT * pVertexAttributeDescriptions,
22734 d.vkCmdSetVertexInputEXT( m_commandBuffer,
22735 vertexBindingDescriptionCount,
22737 vertexAttributeDescriptionCount,
22741 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22742 template <
typename Dispatch>
22744 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::VertexInputBindingDescription2EXT>
const & vertexBindingDescriptions,
22745 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::VertexInputAttributeDescription2EXT>
const & vertexAttributeDescriptions,
22749 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22750 VULKAN_HPP_ASSERT( d.vkCmdSetVertexInputEXT &&
"Function <vkCmdSetVertexInputEXT> requires <VK_EXT_shader_object> or <VK_EXT_vertex_input_dynamic_state>" );
22753 d.vkCmdSetVertexInputEXT( m_commandBuffer,
22754 vertexBindingDescriptions.size(),
22756 vertexAttributeDescriptions.size(),
22761 #if defined( VK_USE_PLATFORM_FUCHSIA )
22764 template <
typename Dispatch>
22766 Device::getMemoryZirconHandleFUCHSIA(
const VULKAN_HPP_NAMESPACE::MemoryGetZirconHandleInfoFUCHSIA * pGetZirconHandleInfo,
22767 zx_handle_t * pZirconHandle,
22771 return static_cast<Result>(
22772 d.vkGetMemoryZirconHandleFUCHSIA( m_device,
reinterpret_cast<const VkMemoryGetZirconHandleInfoFUCHSIA *
>( pGetZirconHandleInfo ), pZirconHandle ) );
22775 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22776 template <
typename Dispatch>
22778 Device::getMemoryZirconHandleFUCHSIA(
const VULKAN_HPP_NAMESPACE::MemoryGetZirconHandleInfoFUCHSIA & getZirconHandleInfo, Dispatch
const & d )
const
22781 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22782 VULKAN_HPP_ASSERT( d.vkGetMemoryZirconHandleFUCHSIA &&
"Function <vkGetMemoryZirconHandleFUCHSIA> requires <VK_FUCHSIA_external_memory>" );
22785 zx_handle_t zirconHandle;
22787 d.vkGetMemoryZirconHandleFUCHSIA( m_device,
reinterpret_cast<const VkMemoryGetZirconHandleInfoFUCHSIA *
>( &getZirconHandleInfo ), &zirconHandle ) );
22794 template <
typename Dispatch>
22797 zx_handle_t zirconHandle,
22798 VULKAN_HPP_NAMESPACE::MemoryZirconHandlePropertiesFUCHSIA * pMemoryZirconHandleProperties,
22802 return static_cast<Result>(
22803 d.vkGetMemoryZirconHandlePropertiesFUCHSIA( m_device,
22806 reinterpret_cast<VkMemoryZirconHandlePropertiesFUCHSIA *
>( pMemoryZirconHandleProperties ) ) );
22809 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22810 template <
typename Dispatch>
22813 zx_handle_t zirconHandle,
22814 Dispatch
const & d )
const
22817 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22819 "Function <vkGetMemoryZirconHandlePropertiesFUCHSIA> requires <VK_FUCHSIA_external_memory>" );
22822 VULKAN_HPP_NAMESPACE::MemoryZirconHandlePropertiesFUCHSIA memoryZirconHandleProperties;
22824 d.vkGetMemoryZirconHandlePropertiesFUCHSIA( m_device,
22827 reinterpret_cast<VkMemoryZirconHandlePropertiesFUCHSIA *
>( &memoryZirconHandleProperties ) ) );
22835 #if defined( VK_USE_PLATFORM_FUCHSIA )
22838 template <
typename Dispatch>
22840 const VULKAN_HPP_NAMESPACE::ImportSemaphoreZirconHandleInfoFUCHSIA * pImportSemaphoreZirconHandleInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
22843 return static_cast<Result>( d.vkImportSemaphoreZirconHandleFUCHSIA(
22844 m_device,
reinterpret_cast<const VkImportSemaphoreZirconHandleInfoFUCHSIA *
>( pImportSemaphoreZirconHandleInfo ) ) );
22847 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22848 template <
typename Dispatch>
22850 Device::importSemaphoreZirconHandleFUCHSIA(
const VULKAN_HPP_NAMESPACE::ImportSemaphoreZirconHandleInfoFUCHSIA & importSemaphoreZirconHandleInfo,
22851 Dispatch
const & d )
const
22854 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22855 VULKAN_HPP_ASSERT( d.vkImportSemaphoreZirconHandleFUCHSIA &&
"Function <vkImportSemaphoreZirconHandleFUCHSIA> requires <VK_FUCHSIA_external_semaphore>" );
22859 m_device,
reinterpret_cast<const VkImportSemaphoreZirconHandleInfoFUCHSIA *
>( &importSemaphoreZirconHandleInfo ) ) );
22866 template <
typename Dispatch>
22868 Device::getSemaphoreZirconHandleFUCHSIA(
const VULKAN_HPP_NAMESPACE::SemaphoreGetZirconHandleInfoFUCHSIA * pGetZirconHandleInfo,
22869 zx_handle_t * pZirconHandle,
22873 return static_cast<Result>(
22874 d.vkGetSemaphoreZirconHandleFUCHSIA( m_device,
reinterpret_cast<const VkSemaphoreGetZirconHandleInfoFUCHSIA *
>( pGetZirconHandleInfo ), pZirconHandle ) );
22877 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22878 template <
typename Dispatch>
22880 Device::getSemaphoreZirconHandleFUCHSIA(
const VULKAN_HPP_NAMESPACE::SemaphoreGetZirconHandleInfoFUCHSIA & getZirconHandleInfo, Dispatch
const & d )
const
22883 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22884 VULKAN_HPP_ASSERT( d.vkGetSemaphoreZirconHandleFUCHSIA &&
"Function <vkGetSemaphoreZirconHandleFUCHSIA> requires <VK_FUCHSIA_external_semaphore>" );
22887 zx_handle_t zirconHandle;
22889 d.vkGetSemaphoreZirconHandleFUCHSIA( m_device,
reinterpret_cast<const VkSemaphoreGetZirconHandleInfoFUCHSIA *
>( &getZirconHandleInfo ), &zirconHandle ) );
22897 #if defined( VK_USE_PLATFORM_FUCHSIA )
22900 template <
typename Dispatch>
22902 Device::createBufferCollectionFUCHSIA(
const VULKAN_HPP_NAMESPACE::BufferCollectionCreateInfoFUCHSIA * pCreateInfo,
22903 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
22904 VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA * pCollection,
22908 return static_cast<Result>( d.vkCreateBufferCollectionFUCHSIA( m_device,
22909 reinterpret_cast<const VkBufferCollectionCreateInfoFUCHSIA *
>( pCreateInfo ),
22911 reinterpret_cast<VkBufferCollectionFUCHSIA *
>( pCollection ) ) );
22914 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22915 template <
typename Dispatch>
22917 Device::createBufferCollectionFUCHSIA(
const VULKAN_HPP_NAMESPACE::BufferCollectionCreateInfoFUCHSIA & createInfo,
22918 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
22919 Dispatch
const & d )
const
22922 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22923 VULKAN_HPP_ASSERT( d.vkCreateBufferCollectionFUCHSIA &&
"Function <vkCreateBufferCollectionFUCHSIA> requires <VK_FUCHSIA_buffer_collection>" );
22926 VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection;
22929 reinterpret_cast<const VkBufferCollectionCreateInfoFUCHSIA *
>( &createInfo ),
22930 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
22931 reinterpret_cast<VkBufferCollectionFUCHSIA *
>( &collection ) ) );
22937 # ifndef VULKAN_HPP_NO_SMART_HANDLE
22938 template <
typename Dispatch>
22940 Device::createBufferCollectionFUCHSIAUnique(
const VULKAN_HPP_NAMESPACE::BufferCollectionCreateInfoFUCHSIA & createInfo,
22941 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
22942 Dispatch
const & d )
const
22945 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22946 VULKAN_HPP_ASSERT( d.vkCreateBufferCollectionFUCHSIA &&
"Function <vkCreateBufferCollectionFUCHSIA> requires <VK_FUCHSIA_buffer_collection>" );
22949 VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection;
22952 reinterpret_cast<const VkBufferCollectionCreateInfoFUCHSIA *
>( &createInfo ),
22953 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
22954 reinterpret_cast<VkBufferCollectionFUCHSIA *
>( &collection ) ) );
22958 result, UniqueHandle<VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA, Dispatch>( collection, ObjectDestroy<Device, Dispatch>( *
this, allocator, d ) ) );
22963 template <
typename Dispatch>
22965 Device::setBufferCollectionImageConstraintsFUCHSIA( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection,
22966 const VULKAN_HPP_NAMESPACE::ImageConstraintsInfoFUCHSIA * pImageConstraintsInfo,
22970 return static_cast<Result>( d.vkSetBufferCollectionImageConstraintsFUCHSIA(
22971 m_device,
static_cast<VkBufferCollectionFUCHSIA
>( collection ),
reinterpret_cast<const VkImageConstraintsInfoFUCHSIA *
>( pImageConstraintsInfo ) ) );
22974 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
22975 template <
typename Dispatch>
22977 Device::setBufferCollectionImageConstraintsFUCHSIA( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection,
22978 const VULKAN_HPP_NAMESPACE::ImageConstraintsInfoFUCHSIA & imageConstraintsInfo,
22979 Dispatch
const & d )
const
22982 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
22984 "Function <vkSetBufferCollectionImageConstraintsFUCHSIA> requires <VK_FUCHSIA_buffer_collection>" );
22988 m_device,
static_cast<VkBufferCollectionFUCHSIA
>( collection ),
reinterpret_cast<const VkImageConstraintsInfoFUCHSIA *
>( &imageConstraintsInfo ) ) );
22995 template <
typename Dispatch>
22997 Device::setBufferCollectionBufferConstraintsFUCHSIA( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection,
22998 const VULKAN_HPP_NAMESPACE::BufferConstraintsInfoFUCHSIA * pBufferConstraintsInfo,
23002 return static_cast<Result>( d.vkSetBufferCollectionBufferConstraintsFUCHSIA(
23003 m_device,
static_cast<VkBufferCollectionFUCHSIA
>( collection ),
reinterpret_cast<const VkBufferConstraintsInfoFUCHSIA *
>( pBufferConstraintsInfo ) ) );
23006 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23007 template <
typename Dispatch>
23009 Device::setBufferCollectionBufferConstraintsFUCHSIA( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection,
23010 const VULKAN_HPP_NAMESPACE::BufferConstraintsInfoFUCHSIA & bufferConstraintsInfo,
23011 Dispatch
const & d )
const
23014 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23016 "Function <vkSetBufferCollectionBufferConstraintsFUCHSIA> requires <VK_FUCHSIA_buffer_collection>" );
23020 m_device,
static_cast<VkBufferCollectionFUCHSIA
>( collection ),
reinterpret_cast<const VkBufferConstraintsInfoFUCHSIA *
>( &bufferConstraintsInfo ) ) );
23027 template <
typename Dispatch>
23028 VULKAN_HPP_INLINE void Device::destroyBufferCollectionFUCHSIA( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection,
23029 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
23033 d.vkDestroyBufferCollectionFUCHSIA(
23034 m_device,
static_cast<VkBufferCollectionFUCHSIA
>( collection ),
reinterpret_cast<const VkAllocationCallbacks *
>( pAllocator ) );
23037 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23038 template <
typename Dispatch>
23039 VULKAN_HPP_INLINE void Device::destroyBufferCollectionFUCHSIA( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection,
23040 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
23044 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23045 VULKAN_HPP_ASSERT( d.vkDestroyBufferCollectionFUCHSIA &&
"Function <vkDestroyBufferCollectionFUCHSIA> requires <VK_FUCHSIA_buffer_collection>" );
23048 d.vkDestroyBufferCollectionFUCHSIA(
23050 static_cast<VkBufferCollectionFUCHSIA
>( collection ),
23051 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
23055 template <
typename Dispatch>
23056 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection,
23057 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
23061 d.vkDestroyBufferCollectionFUCHSIA(
23062 m_device,
static_cast<VkBufferCollectionFUCHSIA
>( collection ),
reinterpret_cast<const VkAllocationCallbacks *
>( pAllocator ) );
23065 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23066 template <
typename Dispatch>
23067 VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection,
23068 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
23072 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23073 VULKAN_HPP_ASSERT( d.vkDestroyBufferCollectionFUCHSIA &&
"Function <vkDestroyBufferCollectionFUCHSIA> requires <VK_FUCHSIA_buffer_collection>" );
23076 d.vkDestroyBufferCollectionFUCHSIA(
23078 static_cast<VkBufferCollectionFUCHSIA
>( collection ),
23079 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
23083 template <
typename Dispatch>
23085 Device::getBufferCollectionPropertiesFUCHSIA( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection,
23086 VULKAN_HPP_NAMESPACE::BufferCollectionPropertiesFUCHSIA * pProperties,
23090 return static_cast<Result>( d.vkGetBufferCollectionPropertiesFUCHSIA(
23091 m_device,
static_cast<VkBufferCollectionFUCHSIA
>( collection ),
reinterpret_cast<VkBufferCollectionPropertiesFUCHSIA *
>( pProperties ) ) );
23094 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23095 template <
typename Dispatch>
23097 Device::getBufferCollectionPropertiesFUCHSIA( VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, Dispatch
const & d )
const
23100 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23102 "Function <vkGetBufferCollectionPropertiesFUCHSIA> requires <VK_FUCHSIA_buffer_collection>" );
23105 VULKAN_HPP_NAMESPACE::BufferCollectionPropertiesFUCHSIA properties;
23107 m_device,
static_cast<VkBufferCollectionFUCHSIA
>( collection ),
reinterpret_cast<VkBufferCollectionPropertiesFUCHSIA *
>( &properties ) ) );
23117 template <
typename Dispatch>
23119 VULKAN_HPP_NAMESPACE::Extent2D * pMaxWorkgroupSize,
23123 return static_cast<Result>( d.vkGetDeviceSubpassShadingMaxWorkgroupSizeHUAWEI(
23124 m_device,
static_cast<VkRenderPass>( renderpass ),
reinterpret_cast<VkExtent2D *
>( pMaxWorkgroupSize ) ) );
23127 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23128 template <
typename Dispatch>
23130 Device::getSubpassShadingMaxWorkgroupSizeHUAWEI( VULKAN_HPP_NAMESPACE::RenderPass renderpass, Dispatch
const & d )
const
23133 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23135 "Function <vkGetDeviceSubpassShadingMaxWorkgroupSizeHUAWEI> requires <VK_HUAWEI_subpass_shading>" );
23138 VULKAN_HPP_NAMESPACE::Extent2D maxWorkgroupSize;
23140 m_device,
static_cast<VkRenderPass>( renderpass ),
reinterpret_cast<VkExtent2D *
>( &maxWorkgroupSize ) ) );
23143 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eIncomplete } );
23149 template <
typename Dispatch>
23153 d.vkCmdSubpassShadingHUAWEI( m_commandBuffer );
23158 template <
typename Dispatch>
23159 VULKAN_HPP_INLINE void CommandBuffer::bindInvocationMaskHUAWEI( VULKAN_HPP_NAMESPACE::ImageView imageView,
23164 d.vkCmdBindInvocationMaskHUAWEI( m_commandBuffer,
static_cast<VkImageView>( imageView ),
static_cast<VkImageLayout>( imageLayout ) );
23169 template <
typename Dispatch>
23171 Device::getMemoryRemoteAddressNV(
const VULKAN_HPP_NAMESPACE::MemoryGetRemoteAddressInfoNV * pMemoryGetRemoteAddressInfo,
23176 return static_cast<Result>( d.vkGetMemoryRemoteAddressNV(
23180 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23181 template <
typename Dispatch>
23183 Device::getMemoryRemoteAddressNV(
const VULKAN_HPP_NAMESPACE::MemoryGetRemoteAddressInfoNV & memoryGetRemoteAddressInfo, Dispatch
const & d )
const
23186 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23187 VULKAN_HPP_ASSERT( d.vkGetMemoryRemoteAddressNV &&
"Function <vkGetMemoryRemoteAddressNV> requires <VK_NV_external_memory_rdma>" );
23201 template <
typename Dispatch>
23203 VULKAN_HPP_NAMESPACE::BaseOutStructure * pPipelineProperties,
23207 return static_cast<Result>( d.vkGetPipelinePropertiesEXT(
23211 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23212 template <
typename Dispatch>
23217 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23218 VULKAN_HPP_ASSERT( d.vkGetPipelinePropertiesEXT &&
"Function <vkGetPipelinePropertiesEXT> requires <VK_EXT_pipeline_properties>" );
23221 VULKAN_HPP_NAMESPACE::BaseOutStructure pipelineProperties;
23232 template <
typename Dispatch>
23236 d.vkCmdSetPatchControlPointsEXT( m_commandBuffer, patchControlPoints );
23239 template <
typename Dispatch>
23244 d.vkCmdSetRasterizerDiscardEnableEXT( m_commandBuffer,
static_cast<VkBool32>( rasterizerDiscardEnable ) );
23247 template <
typename Dispatch>
23251 d.vkCmdSetDepthBiasEnableEXT( m_commandBuffer,
static_cast<VkBool32>( depthBiasEnable ) );
23254 template <
typename Dispatch>
23258 d.vkCmdSetLogicOpEXT( m_commandBuffer,
static_cast<VkLogicOp>( logicOp ) );
23261 template <
typename Dispatch>
23266 d.vkCmdSetPrimitiveRestartEnableEXT( m_commandBuffer,
static_cast<VkBool32>( primitiveRestartEnable ) );
23269 #if defined( VK_USE_PLATFORM_SCREEN_QNX )
23272 template <
typename Dispatch>
23274 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
23275 VULKAN_HPP_NAMESPACE::SurfaceKHR * pSurface,
23279 return static_cast<Result>( d.vkCreateScreenSurfaceQNX( m_instance,
23280 reinterpret_cast<const VkScreenSurfaceCreateInfoQNX *
>( pCreateInfo ),
23285 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23286 template <
typename Dispatch>
23288 Instance::createScreenSurfaceQNX(
const VULKAN_HPP_NAMESPACE::ScreenSurfaceCreateInfoQNX & createInfo,
23289 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
23290 Dispatch
const & d )
const
23293 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23294 VULKAN_HPP_ASSERT( d.vkCreateScreenSurfaceQNX &&
"Function <vkCreateScreenSurfaceQNX> requires <VK_QNX_screen_surface>" );
23297 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
23300 reinterpret_cast<const VkScreenSurfaceCreateInfoQNX *
>( &createInfo ),
23301 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
23308 # ifndef VULKAN_HPP_NO_SMART_HANDLE
23309 template <
typename Dispatch>
23311 Instance::createScreenSurfaceQNXUnique(
const VULKAN_HPP_NAMESPACE::ScreenSurfaceCreateInfoQNX & createInfo,
23312 Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
23313 Dispatch
const & d )
const
23316 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23317 VULKAN_HPP_ASSERT( d.vkCreateScreenSurfaceQNX &&
"Function <vkCreateScreenSurfaceQNX> requires <VK_QNX_screen_surface>" );
23320 VULKAN_HPP_NAMESPACE::SurfaceKHR surface;
23323 reinterpret_cast<const VkScreenSurfaceCreateInfoQNX *
>( &createInfo ),
23324 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
23329 result, UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( surface, ObjectDestroy<Instance, Dispatch>( *
this, allocator, d ) ) );
23334 template <
typename Dispatch>
23336 struct _screen_window * window,
23340 return static_cast<Bool32>( d.vkGetPhysicalDeviceScreenPresentationSupportQNX( m_physicalDevice, queueFamilyIndex, window ) );
23343 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23344 template <
typename Dispatch>
23346 PhysicalDevice::getScreenPresentationSupportQNX( uint32_t queueFamilyIndex,
struct _screen_window & window, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
23349 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23351 "Function <vkGetPhysicalDeviceScreenPresentationSupportQNX> requires <VK_QNX_screen_surface>" );
23354 VkBool32 result = d.vkGetPhysicalDeviceScreenPresentationSupportQNX( m_physicalDevice, queueFamilyIndex, &window );
23363 template <
typename Dispatch>
23369 d.vkCmdSetColorWriteEnableEXT( m_commandBuffer, attachmentCount,
reinterpret_cast<const VkBool32 *
>( pColorWriteEnables ) );
23372 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23373 template <
typename Dispatch>
23374 VULKAN_HPP_INLINE void CommandBuffer::setColorWriteEnableEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Bool32>
const & colorWriteEnables,
23378 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23379 VULKAN_HPP_ASSERT( d.vkCmdSetColorWriteEnableEXT &&
"Function <vkCmdSetColorWriteEnableEXT> requires <VK_EXT_color_write_enable>" );
23382 d.vkCmdSetColorWriteEnableEXT( m_commandBuffer, colorWriteEnables.size(),
reinterpret_cast<const VkBool32 *
>( colorWriteEnables.data() ) );
23388 template <
typename Dispatch>
23393 d.vkCmdTraceRaysIndirect2KHR( m_commandBuffer,
static_cast<VkDeviceAddress>( indirectDeviceAddress ) );
23398 template <
typename Dispatch>
23400 const VULKAN_HPP_NAMESPACE::MultiDrawInfoEXT * pVertexInfo,
23401 uint32_t instanceCount,
23402 uint32_t firstInstance,
23407 d.vkCmdDrawMultiEXT( m_commandBuffer, drawCount,
reinterpret_cast<const VkMultiDrawInfoEXT *
>( pVertexInfo ), instanceCount, firstInstance, stride );
23410 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23411 template <
typename Dispatch>
23412 VULKAN_HPP_INLINE void CommandBuffer::drawMultiEXT( VULKAN_HPP_NAMESPACE::StridedArrayProxy<const VULKAN_HPP_NAMESPACE::MultiDrawInfoEXT>
const & vertexInfo,
23413 uint32_t instanceCount,
23414 uint32_t firstInstance,
23418 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23419 VULKAN_HPP_ASSERT( d.vkCmdDrawMultiEXT &&
"Function <vkCmdDrawMultiEXT> requires <VK_EXT_multi_draw>" );
23422 d.vkCmdDrawMultiEXT( m_commandBuffer,
23427 vertexInfo.stride() );
23431 template <
typename Dispatch>
23433 const VULKAN_HPP_NAMESPACE::MultiDrawIndexedInfoEXT * pIndexInfo,
23434 uint32_t instanceCount,
23435 uint32_t firstInstance,
23437 const int32_t * pVertexOffset,
23441 d.vkCmdDrawMultiIndexedEXT(
23442 m_commandBuffer, drawCount,
reinterpret_cast<const VkMultiDrawIndexedInfoEXT *
>( pIndexInfo ), instanceCount, firstInstance, stride, pVertexOffset );
23445 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23446 template <
typename Dispatch>
23448 CommandBuffer::drawMultiIndexedEXT( VULKAN_HPP_NAMESPACE::StridedArrayProxy<const VULKAN_HPP_NAMESPACE::MultiDrawIndexedInfoEXT>
const & indexInfo,
23449 uint32_t instanceCount,
23450 uint32_t firstInstance,
23455 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23456 VULKAN_HPP_ASSERT( d.vkCmdDrawMultiIndexedEXT &&
"Function <vkCmdDrawMultiIndexedEXT> requires <VK_EXT_multi_draw>" );
23459 d.vkCmdDrawMultiIndexedEXT( m_commandBuffer,
23464 indexInfo.stride(),
23465 static_cast<const int32_t *
>( vertexOffset ) );
23471 template <
typename Dispatch>
23473 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
23474 VULKAN_HPP_NAMESPACE::MicromapEXT * pMicromap,
23478 return static_cast<Result>( d.vkCreateMicromapEXT( m_device,
23484 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23485 template <
typename Dispatch>
23487 Device::createMicromapEXT(
const VULKAN_HPP_NAMESPACE::MicromapCreateInfoEXT & createInfo,
23489 Dispatch
const & d )
const
23492 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23493 VULKAN_HPP_ASSERT( d.vkCreateMicromapEXT &&
"Function <vkCreateMicromapEXT> requires <VK_EXT_opacity_micromap>" );
23496 VULKAN_HPP_NAMESPACE::MicromapEXT micromap;
23498 d.vkCreateMicromapEXT( m_device,
23500 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
23507 # ifndef VULKAN_HPP_NO_SMART_HANDLE
23508 template <
typename Dispatch>
23510 Device::createMicromapEXTUnique(
const VULKAN_HPP_NAMESPACE::MicromapCreateInfoEXT & createInfo,
23512 Dispatch
const & d )
const
23515 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23516 VULKAN_HPP_ASSERT( d.vkCreateMicromapEXT &&
"Function <vkCreateMicromapEXT> requires <VK_EXT_opacity_micromap>" );
23519 VULKAN_HPP_NAMESPACE::MicromapEXT micromap;
23521 d.vkCreateMicromapEXT( m_device,
23523 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
23533 template <
typename Dispatch>
23535 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
23542 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23543 template <
typename Dispatch>
23549 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23550 VULKAN_HPP_ASSERT( d.vkDestroyMicromapEXT &&
"Function <vkDestroyMicromapEXT> requires <VK_EXT_opacity_micromap>" );
23553 d.vkDestroyMicromapEXT( m_device,
23555 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
23559 template <
typename Dispatch>
23561 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
23568 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23569 template <
typename Dispatch>
23575 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23576 VULKAN_HPP_ASSERT( d.vkDestroyMicromapEXT &&
"Function <vkDestroyMicromapEXT> requires <VK_EXT_opacity_micromap>" );
23579 d.vkDestroyMicromapEXT( m_device,
23581 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
23585 template <
typename Dispatch>
23587 const VULKAN_HPP_NAMESPACE::MicromapBuildInfoEXT * pInfos,
23591 d.vkCmdBuildMicromapsEXT( m_commandBuffer, infoCount,
reinterpret_cast<const VkMicromapBuildInfoEXT *
>( pInfos ) );
23594 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23595 template <
typename Dispatch>
23596 VULKAN_HPP_INLINE void CommandBuffer::buildMicromapsEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::MicromapBuildInfoEXT>
const & infos,
23600 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23601 VULKAN_HPP_ASSERT( d.vkCmdBuildMicromapsEXT &&
"Function <vkCmdBuildMicromapsEXT> requires <VK_EXT_opacity_micromap>" );
23604 d.vkCmdBuildMicromapsEXT( m_commandBuffer, infos.size(),
reinterpret_cast<const VkMicromapBuildInfoEXT *
>( infos.data() ) );
23608 template <
typename Dispatch>
23610 uint32_t infoCount,
23611 const VULKAN_HPP_NAMESPACE::MicromapBuildInfoEXT * pInfos,
23615 return static_cast<Result>( d.vkBuildMicromapsEXT(
23619 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23620 template <
typename Dispatch>
23622 Device::buildMicromapsEXT( VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
23623 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::MicromapBuildInfoEXT>
const & infos,
23624 Dispatch
const & d )
const
23627 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23628 VULKAN_HPP_ASSERT( d.vkBuildMicromapsEXT &&
"Function <vkBuildMicromapsEXT> requires <VK_EXT_opacity_micromap>" );
23636 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR } );
23642 template <
typename Dispatch>
23644 const VULKAN_HPP_NAMESPACE::CopyMicromapInfoEXT * pInfo,
23648 return static_cast<Result>(
23652 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23653 template <
typename Dispatch>
23655 const VULKAN_HPP_NAMESPACE::CopyMicromapInfoEXT & info,
23656 Dispatch
const & d )
const
23659 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23660 VULKAN_HPP_ASSERT( d.vkCopyMicromapEXT &&
"Function <vkCopyMicromapEXT> requires <VK_EXT_opacity_micromap>" );
23668 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR } );
23674 template <
typename Dispatch>
23676 const VULKAN_HPP_NAMESPACE::CopyMicromapToMemoryInfoEXT * pInfo,
23680 return static_cast<Result>( d.vkCopyMicromapToMemoryEXT(
23684 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23685 template <
typename Dispatch>
23687 VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
const VULKAN_HPP_NAMESPACE::CopyMicromapToMemoryInfoEXT & info, Dispatch
const & d )
const
23690 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23691 VULKAN_HPP_ASSERT( d.vkCopyMicromapToMemoryEXT &&
"Function <vkCopyMicromapToMemoryEXT> requires <VK_EXT_opacity_micromap>" );
23699 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR } );
23705 template <
typename Dispatch>
23707 const VULKAN_HPP_NAMESPACE::CopyMemoryToMicromapInfoEXT * pInfo,
23711 return static_cast<Result>( d.vkCopyMemoryToMicromapEXT(
23715 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23716 template <
typename Dispatch>
23718 VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation,
const VULKAN_HPP_NAMESPACE::CopyMemoryToMicromapInfoEXT & info, Dispatch
const & d )
const
23721 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23722 VULKAN_HPP_ASSERT( d.vkCopyMemoryToMicromapEXT &&
"Function <vkCopyMemoryToMicromapEXT> requires <VK_EXT_opacity_micromap>" );
23730 { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR } );
23736 template <
typename Dispatch>
23738 const VULKAN_HPP_NAMESPACE::MicromapEXT * pMicromaps,
23746 return static_cast<Result>( d.vkWriteMicromapsPropertiesEXT(
23747 m_device, micromapCount,
reinterpret_cast<const VkMicromapEXT *
>( pMicromaps ),
static_cast<VkQueryType>( queryType ), dataSize, pData, stride ) );
23750 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23751 template <
typename DataType,
typename DataTypeAllocator,
typename Dispatch>
23753 Device::writeMicromapsPropertiesEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::MicromapEXT>
const & micromaps,
23757 Dispatch
const & d )
const
23760 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23761 VULKAN_HPP_ASSERT( d.vkWriteMicromapsPropertiesEXT &&
"Function <vkWriteMicromapsPropertiesEXT> requires <VK_EXT_opacity_micromap>" );
23765 std::vector<DataType, DataTypeAllocator> data( dataSize /
sizeof( DataType ) );
23769 reinterpret_cast<const VkMicromapEXT *
>( micromaps.data() ),
23771 data.size() *
sizeof( DataType ),
23772 reinterpret_cast<void *
>( data.data() ),
23779 template <
typename DataType,
typename Dispatch>
23781 Device::writeMicromapsPropertyEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::MicromapEXT>
const & micromaps,
23784 Dispatch
const & d )
const
23787 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23788 VULKAN_HPP_ASSERT( d.vkWriteMicromapsPropertiesEXT &&
"Function <vkWriteMicromapsPropertiesEXT> requires <VK_EXT_opacity_micromap>" );
23795 reinterpret_cast<const VkMicromapEXT *
>( micromaps.data() ),
23797 sizeof( DataType ),
23798 reinterpret_cast<void *
>( &data ),
23806 template <
typename Dispatch>
23810 d.vkCmdCopyMicromapEXT( m_commandBuffer,
reinterpret_cast<const VkCopyMicromapInfoEXT *
>( pInfo ) );
23813 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23814 template <
typename Dispatch>
23818 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23819 VULKAN_HPP_ASSERT( d.vkCmdCopyMicromapEXT &&
"Function <vkCmdCopyMicromapEXT> requires <VK_EXT_opacity_micromap>" );
23822 d.vkCmdCopyMicromapEXT( m_commandBuffer,
reinterpret_cast<const VkCopyMicromapInfoEXT *
>( &info ) );
23826 template <
typename Dispatch>
23827 VULKAN_HPP_INLINE void CommandBuffer::copyMicromapToMemoryEXT(
const VULKAN_HPP_NAMESPACE::CopyMicromapToMemoryInfoEXT * pInfo,
23834 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23835 template <
typename Dispatch>
23836 VULKAN_HPP_INLINE void CommandBuffer::copyMicromapToMemoryEXT(
const VULKAN_HPP_NAMESPACE::CopyMicromapToMemoryInfoEXT & info,
23840 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23841 VULKAN_HPP_ASSERT( d.vkCmdCopyMicromapToMemoryEXT &&
"Function <vkCmdCopyMicromapToMemoryEXT> requires <VK_EXT_opacity_micromap>" );
23848 template <
typename Dispatch>
23849 VULKAN_HPP_INLINE void CommandBuffer::copyMemoryToMicromapEXT(
const VULKAN_HPP_NAMESPACE::CopyMemoryToMicromapInfoEXT * pInfo,
23856 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23857 template <
typename Dispatch>
23858 VULKAN_HPP_INLINE void CommandBuffer::copyMemoryToMicromapEXT(
const VULKAN_HPP_NAMESPACE::CopyMemoryToMicromapInfoEXT & info,
23862 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23863 VULKAN_HPP_ASSERT( d.vkCmdCopyMemoryToMicromapEXT &&
"Function <vkCmdCopyMemoryToMicromapEXT> requires <VK_EXT_opacity_micromap>" );
23870 template <
typename Dispatch>
23872 const VULKAN_HPP_NAMESPACE::MicromapEXT * pMicromaps,
23874 VULKAN_HPP_NAMESPACE::QueryPool queryPool,
23875 uint32_t firstQuery,
23879 d.vkCmdWriteMicromapsPropertiesEXT( m_commandBuffer,
23887 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23888 template <
typename Dispatch>
23890 CommandBuffer::writeMicromapsPropertiesEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::MicromapEXT>
const & micromaps,
23892 VULKAN_HPP_NAMESPACE::QueryPool queryPool,
23893 uint32_t firstQuery,
23897 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23898 VULKAN_HPP_ASSERT( d.vkCmdWriteMicromapsPropertiesEXT &&
"Function <vkCmdWriteMicromapsPropertiesEXT> requires <VK_EXT_opacity_micromap>" );
23901 d.vkCmdWriteMicromapsPropertiesEXT( m_commandBuffer,
23903 reinterpret_cast<const VkMicromapEXT *
>( micromaps.data() ),
23910 template <
typename Dispatch>
23911 VULKAN_HPP_INLINE void Device::getMicromapCompatibilityEXT(
const VULKAN_HPP_NAMESPACE::MicromapVersionInfoEXT * pVersionInfo,
23916 d.vkGetDeviceMicromapCompatibilityEXT( m_device,
23921 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23922 template <
typename Dispatch>
23924 Device::getMicromapCompatibilityEXT(
const VULKAN_HPP_NAMESPACE::MicromapVersionInfoEXT & versionInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
23927 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23928 VULKAN_HPP_ASSERT( d.vkGetDeviceMicromapCompatibilityEXT &&
"Function <vkGetDeviceMicromapCompatibilityEXT> requires <VK_EXT_opacity_micromap>" );
23932 d.vkGetDeviceMicromapCompatibilityEXT( m_device,
23936 return compatibility;
23940 template <
typename Dispatch>
23942 const VULKAN_HPP_NAMESPACE::MicromapBuildInfoEXT * pBuildInfo,
23943 VULKAN_HPP_NAMESPACE::MicromapBuildSizesInfoEXT * pSizeInfo,
23947 d.vkGetMicromapBuildSizesEXT( m_device,
23953 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
23954 template <
typename Dispatch>
23957 const VULKAN_HPP_NAMESPACE::MicromapBuildInfoEXT & buildInfo,
23961 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
23962 VULKAN_HPP_ASSERT( d.vkGetMicromapBuildSizesEXT &&
"Function <vkGetMicromapBuildSizesEXT> requires <VK_EXT_opacity_micromap>" );
23965 VULKAN_HPP_NAMESPACE::MicromapBuildSizesInfoEXT sizeInfo;
23966 d.vkGetMicromapBuildSizesEXT( m_device,
23977 template <
typename Dispatch>
23979 CommandBuffer::drawClusterHUAWEI( uint32_t groupCountX, uint32_t groupCountY, uint32_t groupCountZ, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
23982 d.vkCmdDrawClusterHUAWEI( m_commandBuffer, groupCountX, groupCountY, groupCountZ );
23985 template <
typename Dispatch>
23991 d.vkCmdDrawClusterIndirectHUAWEI( m_commandBuffer,
static_cast<VkBuffer>( buffer ),
static_cast<VkDeviceSize>( offset ) );
23996 template <
typename Dispatch>
24000 d.vkSetDeviceMemoryPriorityEXT( m_device,
static_cast<VkDeviceMemory>( memory ), priority );
24005 template <
typename Dispatch>
24006 VULKAN_HPP_INLINE void Device::getBufferMemoryRequirementsKHR(
const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements * pInfo,
24007 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
24011 d.vkGetDeviceBufferMemoryRequirementsKHR(
24015 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24016 template <
typename Dispatch>
24018 Device::getBufferMemoryRequirementsKHR(
const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
24021 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24023 "Function <vkGetDeviceBufferMemoryRequirementsKHR> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
24026 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
24027 d.vkGetDeviceBufferMemoryRequirementsKHR(
24030 return memoryRequirements;
24033 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
24035 Device::getBufferMemoryRequirementsKHR(
const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
24038 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24040 "Function <vkGetDeviceBufferMemoryRequirementsKHR> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
24043 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
24044 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
24045 d.vkGetDeviceBufferMemoryRequirementsKHR(
24048 return structureChain;
24052 template <
typename Dispatch>
24053 VULKAN_HPP_INLINE void Device::getImageMemoryRequirementsKHR(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements * pInfo,
24054 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
24058 d.vkGetDeviceImageMemoryRequirementsKHR(
24062 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24063 template <
typename Dispatch>
24065 Device::getImageMemoryRequirementsKHR(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
24068 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24070 "Function <vkGetDeviceImageMemoryRequirementsKHR> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
24073 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
24074 d.vkGetDeviceImageMemoryRequirementsKHR(
24077 return memoryRequirements;
24080 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
24082 Device::getImageMemoryRequirementsKHR(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
24085 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24087 "Function <vkGetDeviceImageMemoryRequirementsKHR> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
24090 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
24091 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
24092 d.vkGetDeviceImageMemoryRequirementsKHR(
24095 return structureChain;
24099 template <
typename Dispatch>
24100 VULKAN_HPP_INLINE void Device::getImageSparseMemoryRequirementsKHR(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements * pInfo,
24101 uint32_t * pSparseMemoryRequirementCount,
24102 VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2 * pSparseMemoryRequirements,
24106 d.vkGetDeviceImageSparseMemoryRequirementsKHR( m_device,
24108 pSparseMemoryRequirementCount,
24112 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24113 template <
typename SparseImageMemoryRequirements2Allocator,
typename Dispatch>
24115 Device::getImageSparseMemoryRequirementsKHR(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements & info, Dispatch
const & d )
const
24118 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24120 "Function <vkGetDeviceImageSparseMemoryRequirementsKHR> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
24123 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements;
24124 uint32_t sparseMemoryRequirementCount;
24125 d.vkGetDeviceImageSparseMemoryRequirementsKHR(
24127 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
24128 d.vkGetDeviceImageSparseMemoryRequirementsKHR( m_device,
24130 &sparseMemoryRequirementCount,
24133 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
24134 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
24136 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
24138 return sparseMemoryRequirements;
24141 template <
typename SparseImageMemoryRequirements2Allocator,
24143 typename std::enable_if<
24144 std::is_same<typename SparseImageMemoryRequirements2Allocator::value_type, VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2>::value,
24147 Device::getImageSparseMemoryRequirementsKHR(
const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements & info,
24148 SparseImageMemoryRequirements2Allocator & sparseImageMemoryRequirements2Allocator,
24149 Dispatch
const & d )
const
24152 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24154 "Function <vkGetDeviceImageSparseMemoryRequirementsKHR> requires <VK_KHR_maintenance4> or <VK_VERSION_1_3>" );
24157 std::vector<VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements(
24158 sparseImageMemoryRequirements2Allocator );
24159 uint32_t sparseMemoryRequirementCount;
24160 d.vkGetDeviceImageSparseMemoryRequirementsKHR(
24162 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
24163 d.vkGetDeviceImageSparseMemoryRequirementsKHR( m_device,
24165 &sparseMemoryRequirementCount,
24168 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
24169 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
24171 sparseMemoryRequirements.resize( sparseMemoryRequirementCount );
24173 return sparseMemoryRequirements;
24179 template <
typename Dispatch>
24180 VULKAN_HPP_INLINE void Device::getDescriptorSetLayoutHostMappingInfoVALVE(
const VULKAN_HPP_NAMESPACE::DescriptorSetBindingReferenceVALVE * pBindingReference,
24181 VULKAN_HPP_NAMESPACE::DescriptorSetLayoutHostMappingInfoVALVE * pHostMapping,
24185 d.vkGetDescriptorSetLayoutHostMappingInfoVALVE( m_device,
24190 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24191 template <
typename Dispatch>
24193 Device::getDescriptorSetLayoutHostMappingInfoVALVE(
const VULKAN_HPP_NAMESPACE::DescriptorSetBindingReferenceVALVE & bindingReference,
24197 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24199 "Function <vkGetDescriptorSetLayoutHostMappingInfoVALVE> requires <VK_VALVE_descriptor_set_host_mapping>" );
24202 VULKAN_HPP_NAMESPACE::DescriptorSetLayoutHostMappingInfoVALVE hostMapping;
24203 d.vkGetDescriptorSetLayoutHostMappingInfoVALVE( m_device,
24207 return hostMapping;
24211 template <
typename Dispatch>
24213 Device::getDescriptorSetHostMappingVALVE( VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet,
void ** ppData, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
24216 d.vkGetDescriptorSetHostMappingVALVE( m_device,
static_cast<VkDescriptorSet>( descriptorSet ), ppData );
24219 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24220 template <
typename Dispatch>
24225 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24227 "Function <vkGetDescriptorSetHostMappingVALVE> requires <VK_VALVE_descriptor_set_host_mapping>" );
24231 d.vkGetDescriptorSetHostMappingVALVE( m_device,
static_cast<VkDescriptorSet>( descriptorSet ), &pData );
24239 template <
typename Dispatch>
24241 uint32_t copyCount,
24246 d.vkCmdCopyMemoryIndirectNV( m_commandBuffer,
static_cast<VkDeviceAddress>( copyBufferAddress ), copyCount, stride );
24249 template <
typename Dispatch>
24251 uint32_t copyCount,
24253 VULKAN_HPP_NAMESPACE::Image dstImage,
24255 const VULKAN_HPP_NAMESPACE::ImageSubresourceLayers * pImageSubresources,
24259 d.vkCmdCopyMemoryToImageIndirectNV( m_commandBuffer,
24263 static_cast<VkImage>( dstImage ),
24268 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24269 template <
typename Dispatch>
24273 VULKAN_HPP_NAMESPACE::Image dstImage,
24275 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageSubresourceLayers>
const & imageSubresources,
24279 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24280 VULKAN_HPP_ASSERT( d.vkCmdCopyMemoryToImageIndirectNV &&
"Function <vkCmdCopyMemoryToImageIndirectNV> requires <VK_NV_copy_memory_indirect>" );
24283 d.vkCmdCopyMemoryToImageIndirectNV( m_commandBuffer,
24285 imageSubresources.size(),
24287 static_cast<VkImage>( dstImage ),
24295 template <
typename Dispatch>
24297 const VULKAN_HPP_NAMESPACE::DecompressMemoryRegionNV * pDecompressMemoryRegions,
24301 d.vkCmdDecompressMemoryNV( m_commandBuffer, decompressRegionCount,
reinterpret_cast<const VkDecompressMemoryRegionNV *
>( pDecompressMemoryRegions ) );
24304 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24305 template <
typename Dispatch>
24307 CommandBuffer::decompressMemoryNV( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DecompressMemoryRegionNV>
const & decompressMemoryRegions,
24311 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24312 VULKAN_HPP_ASSERT( d.vkCmdDecompressMemoryNV &&
"Function <vkCmdDecompressMemoryNV> requires <VK_NV_memory_decompression>" );
24315 d.vkCmdDecompressMemoryNV(
24316 m_commandBuffer, decompressMemoryRegions.size(),
reinterpret_cast<const VkDecompressMemoryRegionNV *
>( decompressMemoryRegions.data() ) );
24320 template <
typename Dispatch>
24327 d.vkCmdDecompressMemoryIndirectCountNV(
24328 m_commandBuffer,
static_cast<VkDeviceAddress>( indirectCommandsAddress ),
static_cast<VkDeviceAddress>( indirectCommandsCountAddress ), stride );
24333 template <
typename Dispatch>
24334 VULKAN_HPP_INLINE void Device::getPipelineIndirectMemoryRequirementsNV(
const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo * pCreateInfo,
24335 VULKAN_HPP_NAMESPACE::MemoryRequirements2 * pMemoryRequirements,
24339 d.vkGetPipelineIndirectMemoryRequirementsNV(
24343 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24344 template <
typename Dispatch>
24346 Device::getPipelineIndirectMemoryRequirementsNV(
const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo & createInfo,
24350 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24352 "Function <vkGetPipelineIndirectMemoryRequirementsNV> requires <VK_NV_device_generated_commands_compute>" );
24355 VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
24356 d.vkGetPipelineIndirectMemoryRequirementsNV(
24359 return memoryRequirements;
24362 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
24364 Device::getPipelineIndirectMemoryRequirementsNV(
const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo & createInfo,
24368 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24370 "Function <vkGetPipelineIndirectMemoryRequirementsNV> requires <VK_NV_device_generated_commands_compute>" );
24373 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
24374 VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
24375 d.vkGetPipelineIndirectMemoryRequirementsNV(
24378 return structureChain;
24382 template <
typename Dispatch>
24384 VULKAN_HPP_NAMESPACE::Pipeline pipeline,
24388 d.vkCmdUpdatePipelineIndirectBufferNV( m_commandBuffer,
static_cast<VkPipelineBindPoint>( pipelineBindPoint ),
static_cast<VkPipeline>( pipeline ) );
24391 template <
typename Dispatch>
24400 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24401 template <
typename Dispatch>
24403 Device::getPipelineIndirectAddressNV(
const VULKAN_HPP_NAMESPACE::PipelineIndirectDeviceAddressInfoNV & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
24406 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24408 "Function <vkGetPipelineIndirectDeviceAddressNV> requires <VK_NV_device_generated_commands_compute>" );
24419 template <
typename Dispatch>
24427 template <
typename Dispatch>
24431 d.vkCmdSetDepthClampEnableEXT( m_commandBuffer,
static_cast<VkBool32>( depthClampEnable ) );
24434 template <
typename Dispatch>
24438 d.vkCmdSetPolygonModeEXT( m_commandBuffer,
static_cast<VkPolygonMode>( polygonMode ) );
24441 template <
typename Dispatch>
24446 d.vkCmdSetRasterizationSamplesEXT( m_commandBuffer,
static_cast<VkSampleCountFlagBits>( rasterizationSamples ) );
24449 template <
typename Dispatch>
24458 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24459 template <
typename Dispatch>
24461 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SampleMask>
const & sampleMask,
24465 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24466 VULKAN_HPP_ASSERT( d.vkCmdSetSampleMaskEXT &&
"Function <vkCmdSetSampleMaskEXT> requires <VK_EXT_extended_dynamic_state3> or <VK_EXT_shader_object>" );
24468 # ifdef VULKAN_HPP_NO_EXCEPTIONS
24469 VULKAN_HPP_ASSERT( sampleMask.size() == (
static_cast<uint32_t
>( samples ) + 31 ) / 32 );
24471 if ( sampleMask.size() != (
static_cast<uint32_t
>( samples ) + 31 ) / 32 )
24481 template <
typename Dispatch>
24486 d.vkCmdSetAlphaToCoverageEnableEXT( m_commandBuffer,
static_cast<VkBool32>( alphaToCoverageEnable ) );
24489 template <
typename Dispatch>
24493 d.vkCmdSetAlphaToOneEnableEXT( m_commandBuffer,
static_cast<VkBool32>( alphaToOneEnable ) );
24496 template <
typename Dispatch>
24500 d.vkCmdSetLogicOpEnableEXT( m_commandBuffer,
static_cast<VkBool32>( logicOpEnable ) );
24503 template <
typename Dispatch>
24505 uint32_t attachmentCount,
24510 d.vkCmdSetColorBlendEnableEXT( m_commandBuffer, firstAttachment, attachmentCount,
reinterpret_cast<const VkBool32 *
>( pColorBlendEnables ) );
24513 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24514 template <
typename Dispatch>
24516 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::Bool32>
const & colorBlendEnables,
24520 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24522 "Function <vkCmdSetColorBlendEnableEXT> requires <VK_EXT_extended_dynamic_state3> or <VK_EXT_shader_object>" );
24525 d.vkCmdSetColorBlendEnableEXT( m_commandBuffer, firstAttachment, colorBlendEnables.size(),
reinterpret_cast<const VkBool32 *
>( colorBlendEnables.data() ) );
24529 template <
typename Dispatch>
24531 uint32_t attachmentCount,
24532 const VULKAN_HPP_NAMESPACE::ColorBlendEquationEXT * pColorBlendEquations,
24536 d.vkCmdSetColorBlendEquationEXT(
24537 m_commandBuffer, firstAttachment, attachmentCount,
reinterpret_cast<const VkColorBlendEquationEXT *
>( pColorBlendEquations ) );
24540 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24541 template <
typename Dispatch>
24543 CommandBuffer::setColorBlendEquationEXT( uint32_t firstAttachment,
24544 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ColorBlendEquationEXT>
const & colorBlendEquations,
24548 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24550 "Function <vkCmdSetColorBlendEquationEXT> requires <VK_EXT_extended_dynamic_state3> or <VK_EXT_shader_object>" );
24553 d.vkCmdSetColorBlendEquationEXT(
24554 m_commandBuffer, firstAttachment, colorBlendEquations.size(),
reinterpret_cast<const VkColorBlendEquationEXT *
>( colorBlendEquations.data() ) );
24558 template <
typename Dispatch>
24560 uint32_t attachmentCount,
24565 d.vkCmdSetColorWriteMaskEXT( m_commandBuffer, firstAttachment, attachmentCount,
reinterpret_cast<const VkColorComponentFlags *
>( pColorWriteMasks ) );
24568 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24569 template <
typename Dispatch>
24571 CommandBuffer::setColorWriteMaskEXT( uint32_t firstAttachment,
24572 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ColorComponentFlags>
const & colorWriteMasks,
24576 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24578 "Function <vkCmdSetColorWriteMaskEXT> requires <VK_EXT_extended_dynamic_state3> or <VK_EXT_shader_object>" );
24581 d.vkCmdSetColorWriteMaskEXT(
24582 m_commandBuffer, firstAttachment, colorWriteMasks.size(),
reinterpret_cast<const VkColorComponentFlags *
>( colorWriteMasks.data() ) );
24586 template <
typename Dispatch>
24590 d.vkCmdSetRasterizationStreamEXT( m_commandBuffer, rasterizationStream );
24593 template <
typename Dispatch>
24602 template <
typename Dispatch>
24603 VULKAN_HPP_INLINE void CommandBuffer::setExtraPrimitiveOverestimationSizeEXT(
float extraPrimitiveOverestimationSize,
24607 d.vkCmdSetExtraPrimitiveOverestimationSizeEXT( m_commandBuffer, extraPrimitiveOverestimationSize );
24610 template <
typename Dispatch>
24614 d.vkCmdSetDepthClipEnableEXT( m_commandBuffer,
static_cast<VkBool32>( depthClipEnable ) );
24617 template <
typename Dispatch>
24622 d.vkCmdSetSampleLocationsEnableEXT( m_commandBuffer,
static_cast<VkBool32>( sampleLocationsEnable ) );
24625 template <
typename Dispatch>
24627 uint32_t attachmentCount,
24628 const VULKAN_HPP_NAMESPACE::ColorBlendAdvancedEXT * pColorBlendAdvanced,
24632 d.vkCmdSetColorBlendAdvancedEXT(
24633 m_commandBuffer, firstAttachment, attachmentCount,
reinterpret_cast<const VkColorBlendAdvancedEXT *
>( pColorBlendAdvanced ) );
24636 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24637 template <
typename Dispatch>
24639 CommandBuffer::setColorBlendAdvancedEXT( uint32_t firstAttachment,
24640 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ColorBlendAdvancedEXT>
const & colorBlendAdvanced,
24644 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24646 "Function <vkCmdSetColorBlendAdvancedEXT> requires <VK_EXT_extended_dynamic_state3> or <VK_EXT_shader_object>" );
24649 d.vkCmdSetColorBlendAdvancedEXT(
24650 m_commandBuffer, firstAttachment, colorBlendAdvanced.size(),
reinterpret_cast<const VkColorBlendAdvancedEXT *
>( colorBlendAdvanced.data() ) );
24654 template <
typename Dispatch>
24659 d.vkCmdSetProvokingVertexModeEXT( m_commandBuffer,
static_cast<VkProvokingVertexModeEXT>( provokingVertexMode ) );
24662 template <
typename Dispatch>
24670 template <
typename Dispatch>
24674 d.vkCmdSetLineStippleEnableEXT( m_commandBuffer,
static_cast<VkBool32>( stippledLineEnable ) );
24677 template <
typename Dispatch>
24682 d.vkCmdSetDepthClipNegativeOneToOneEXT( m_commandBuffer,
static_cast<VkBool32>( negativeOneToOne ) );
24685 template <
typename Dispatch>
24690 d.vkCmdSetViewportWScalingEnableNV( m_commandBuffer,
static_cast<VkBool32>( viewportWScalingEnable ) );
24693 template <
typename Dispatch>
24695 uint32_t viewportCount,
24696 const VULKAN_HPP_NAMESPACE::ViewportSwizzleNV * pViewportSwizzles,
24700 d.vkCmdSetViewportSwizzleNV( m_commandBuffer, firstViewport, viewportCount,
reinterpret_cast<const VkViewportSwizzleNV *
>( pViewportSwizzles ) );
24703 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24704 template <
typename Dispatch>
24706 CommandBuffer::setViewportSwizzleNV( uint32_t firstViewport,
24707 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ViewportSwizzleNV>
const & viewportSwizzles,
24711 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24713 "Function <vkCmdSetViewportSwizzleNV> requires <VK_EXT_extended_dynamic_state3> or <VK_EXT_shader_object>" );
24716 d.vkCmdSetViewportSwizzleNV(
24717 m_commandBuffer, firstViewport, viewportSwizzles.size(),
reinterpret_cast<const VkViewportSwizzleNV *
>( viewportSwizzles.data() ) );
24721 template <
typename Dispatch>
24726 d.vkCmdSetCoverageToColorEnableNV( m_commandBuffer,
static_cast<VkBool32>( coverageToColorEnable ) );
24729 template <
typename Dispatch>
24733 d.vkCmdSetCoverageToColorLocationNV( m_commandBuffer, coverageToColorLocation );
24736 template <
typename Dispatch>
24744 template <
typename Dispatch>
24749 d.vkCmdSetCoverageModulationTableEnableNV( m_commandBuffer,
static_cast<VkBool32>( coverageModulationTableEnable ) );
24752 template <
typename Dispatch>
24753 VULKAN_HPP_INLINE void CommandBuffer::setCoverageModulationTableNV( uint32_t coverageModulationTableCount,
24754 const float * pCoverageModulationTable,
24758 d.vkCmdSetCoverageModulationTableNV( m_commandBuffer, coverageModulationTableCount, pCoverageModulationTable );
24761 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24762 template <
typename Dispatch>
24763 VULKAN_HPP_INLINE void CommandBuffer::setCoverageModulationTableNV( VULKAN_HPP_NAMESPACE::ArrayProxy<const float>
const & coverageModulationTable,
24767 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24769 "Function <vkCmdSetCoverageModulationTableNV> requires <VK_EXT_extended_dynamic_state3> or <VK_EXT_shader_object>" );
24772 d.vkCmdSetCoverageModulationTableNV( m_commandBuffer, coverageModulationTable.size(), coverageModulationTable.data() );
24776 template <
typename Dispatch>
24781 d.vkCmdSetShadingRateImageEnableNV( m_commandBuffer,
static_cast<VkBool32>( shadingRateImageEnable ) );
24784 template <
typename Dispatch>
24789 d.vkCmdSetRepresentativeFragmentTestEnableNV( m_commandBuffer,
static_cast<VkBool32>( representativeFragmentTestEnable ) );
24792 template <
typename Dispatch>
24802 template <
typename Dispatch>
24803 VULKAN_HPP_INLINE void Device::getShaderModuleIdentifierEXT( VULKAN_HPP_NAMESPACE::ShaderModule shaderModule,
24804 VULKAN_HPP_NAMESPACE::ShaderModuleIdentifierEXT * pIdentifier,
24811 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24812 template <
typename Dispatch>
24814 Device::getShaderModuleIdentifierEXT( VULKAN_HPP_NAMESPACE::ShaderModule shaderModule, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
24817 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24818 VULKAN_HPP_ASSERT( d.vkGetShaderModuleIdentifierEXT &&
"Function <vkGetShaderModuleIdentifierEXT> requires <VK_EXT_shader_module_identifier>" );
24821 VULKAN_HPP_NAMESPACE::ShaderModuleIdentifierEXT identifier;
24828 template <
typename Dispatch>
24829 VULKAN_HPP_INLINE void Device::getShaderModuleCreateInfoIdentifierEXT(
const VULKAN_HPP_NAMESPACE::ShaderModuleCreateInfo * pCreateInfo,
24830 VULKAN_HPP_NAMESPACE::ShaderModuleIdentifierEXT * pIdentifier,
24834 d.vkGetShaderModuleCreateInfoIdentifierEXT(
24838 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24839 template <
typename Dispatch>
24841 Device::getShaderModuleCreateInfoIdentifierEXT(
const VULKAN_HPP_NAMESPACE::ShaderModuleCreateInfo & createInfo,
24845 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24847 "Function <vkGetShaderModuleCreateInfoIdentifierEXT> requires <VK_EXT_shader_module_identifier>" );
24850 VULKAN_HPP_NAMESPACE::ShaderModuleIdentifierEXT identifier;
24851 d.vkGetShaderModuleCreateInfoIdentifierEXT(
24860 template <
typename Dispatch>
24862 PhysicalDevice::getOpticalFlowImageFormatsNV(
const VULKAN_HPP_NAMESPACE::OpticalFlowImageFormatInfoNV * pOpticalFlowImageFormatInfo,
24863 uint32_t * pFormatCount,
24864 VULKAN_HPP_NAMESPACE::OpticalFlowImageFormatPropertiesNV * pImageFormatProperties,
24868 return static_cast<Result>(
24869 d.vkGetPhysicalDeviceOpticalFlowImageFormatsNV( m_physicalDevice,
24875 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24876 template <
typename OpticalFlowImageFormatPropertiesNVAllocator,
typename Dispatch>
24879 PhysicalDevice::getOpticalFlowImageFormatsNV(
const VULKAN_HPP_NAMESPACE::OpticalFlowImageFormatInfoNV & opticalFlowImageFormatInfo,
24880 Dispatch
const & d )
const
24883 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24885 "Function <vkGetPhysicalDeviceOpticalFlowImageFormatsNV> requires <VK_NV_optical_flow>" );
24888 std::vector<VULKAN_HPP_NAMESPACE::OpticalFlowImageFormatPropertiesNV, OpticalFlowImageFormatPropertiesNVAllocator> imageFormatProperties;
24889 uint32_t formatCount;
24895 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && formatCount )
24897 imageFormatProperties.resize( formatCount );
24899 d.vkGetPhysicalDeviceOpticalFlowImageFormatsNV( m_physicalDevice,
24904 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
24907 if ( formatCount < imageFormatProperties.size() )
24909 imageFormatProperties.resize( formatCount );
24914 template <
typename OpticalFlowImageFormatPropertiesNVAllocator,
24916 typename std::enable_if<
24917 std::is_same<typename OpticalFlowImageFormatPropertiesNVAllocator::value_type, VULKAN_HPP_NAMESPACE::OpticalFlowImageFormatPropertiesNV>::value,
24921 PhysicalDevice::getOpticalFlowImageFormatsNV(
const VULKAN_HPP_NAMESPACE::OpticalFlowImageFormatInfoNV & opticalFlowImageFormatInfo,
24922 OpticalFlowImageFormatPropertiesNVAllocator & opticalFlowImageFormatPropertiesNVAllocator,
24923 Dispatch
const & d )
const
24926 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24928 "Function <vkGetPhysicalDeviceOpticalFlowImageFormatsNV> requires <VK_NV_optical_flow>" );
24931 std::vector<VULKAN_HPP_NAMESPACE::OpticalFlowImageFormatPropertiesNV, OpticalFlowImageFormatPropertiesNVAllocator> imageFormatProperties(
24932 opticalFlowImageFormatPropertiesNVAllocator );
24933 uint32_t formatCount;
24939 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && formatCount )
24941 imageFormatProperties.resize( formatCount );
24943 d.vkGetPhysicalDeviceOpticalFlowImageFormatsNV( m_physicalDevice,
24948 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
24951 if ( formatCount < imageFormatProperties.size() )
24953 imageFormatProperties.resize( formatCount );
24959 template <
typename Dispatch>
24961 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
24962 VULKAN_HPP_NAMESPACE::OpticalFlowSessionNV * pSession,
24966 return static_cast<Result>( d.vkCreateOpticalFlowSessionNV( m_device,
24972 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
24973 template <
typename Dispatch>
24975 Device::createOpticalFlowSessionNV(
const VULKAN_HPP_NAMESPACE::OpticalFlowSessionCreateInfoNV & createInfo,
24977 Dispatch
const & d )
const
24980 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
24981 VULKAN_HPP_ASSERT( d.vkCreateOpticalFlowSessionNV &&
"Function <vkCreateOpticalFlowSessionNV> requires <VK_NV_optical_flow>" );
24984 VULKAN_HPP_NAMESPACE::OpticalFlowSessionNV session;
24988 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
24995 # ifndef VULKAN_HPP_NO_SMART_HANDLE
24996 template <
typename Dispatch>
24998 Device::createOpticalFlowSessionNVUnique(
const VULKAN_HPP_NAMESPACE::OpticalFlowSessionCreateInfoNV & createInfo,
25000 Dispatch
const & d )
const
25003 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25004 VULKAN_HPP_ASSERT( d.vkCreateOpticalFlowSessionNV &&
"Function <vkCreateOpticalFlowSessionNV> requires <VK_NV_optical_flow>" );
25007 VULKAN_HPP_NAMESPACE::OpticalFlowSessionNV session;
25011 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
25021 template <
typename Dispatch>
25022 VULKAN_HPP_INLINE void Device::destroyOpticalFlowSessionNV( VULKAN_HPP_NAMESPACE::OpticalFlowSessionNV session,
25023 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
25030 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25031 template <
typename Dispatch>
25032 VULKAN_HPP_INLINE void Device::destroyOpticalFlowSessionNV( VULKAN_HPP_NAMESPACE::OpticalFlowSessionNV session,
25037 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25038 VULKAN_HPP_ASSERT( d.vkDestroyOpticalFlowSessionNV &&
"Function <vkDestroyOpticalFlowSessionNV> requires <VK_NV_optical_flow>" );
25041 d.vkDestroyOpticalFlowSessionNV(
25044 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
25048 template <
typename Dispatch>
25050 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
25057 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25058 template <
typename Dispatch>
25064 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25065 VULKAN_HPP_ASSERT( d.vkDestroyOpticalFlowSessionNV &&
"Function <vkDestroyOpticalFlowSessionNV> requires <VK_NV_optical_flow>" );
25068 d.vkDestroyOpticalFlowSessionNV(
25071 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
25075 #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE
25076 template <
typename Dispatch>
25079 VULKAN_HPP_NAMESPACE::ImageView view,
25084 return static_cast<Result>( d.vkBindOpticalFlowSessionImageNV( m_device,
25091 template <
typename Dispatch>
25093 Device::bindOpticalFlowSessionImageNV( VULKAN_HPP_NAMESPACE::OpticalFlowSessionNV session,
25095 VULKAN_HPP_NAMESPACE::ImageView view,
25097 Dispatch
const & d )
const
25100 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25101 VULKAN_HPP_ASSERT( d.vkBindOpticalFlowSessionImageNV &&
"Function <vkBindOpticalFlowSessionImageNV> requires <VK_NV_optical_flow>" );
25116 template <
typename Dispatch>
25117 VULKAN_HPP_INLINE void CommandBuffer::opticalFlowExecuteNV( VULKAN_HPP_NAMESPACE::OpticalFlowSessionNV session,
25118 const VULKAN_HPP_NAMESPACE::OpticalFlowExecuteInfoNV * pExecuteInfo,
25122 d.vkCmdOpticalFlowExecuteNV(
25126 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25127 template <
typename Dispatch>
25128 VULKAN_HPP_INLINE void CommandBuffer::opticalFlowExecuteNV( VULKAN_HPP_NAMESPACE::OpticalFlowSessionNV session,
25129 const VULKAN_HPP_NAMESPACE::OpticalFlowExecuteInfoNV & executeInfo,
25133 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25134 VULKAN_HPP_ASSERT( d.vkCmdOpticalFlowExecuteNV &&
"Function <vkCmdOpticalFlowExecuteNV> requires <VK_NV_optical_flow>" );
25137 d.vkCmdOpticalFlowExecuteNV(
25144 template <
typename Dispatch>
25152 d.vkCmdBindIndexBuffer2KHR( m_commandBuffer,
25159 template <
typename Dispatch>
25160 VULKAN_HPP_INLINE void Device::getRenderingAreaGranularityKHR(
const VULKAN_HPP_NAMESPACE::RenderingAreaInfoKHR * pRenderingAreaInfo,
25161 VULKAN_HPP_NAMESPACE::Extent2D * pGranularity,
25165 d.vkGetRenderingAreaGranularityKHR(
25169 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25170 template <
typename Dispatch>
25172 Device::getRenderingAreaGranularityKHR(
const VULKAN_HPP_NAMESPACE::RenderingAreaInfoKHR & renderingAreaInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
25175 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25176 VULKAN_HPP_ASSERT( d.vkGetRenderingAreaGranularityKHR &&
"Function <vkGetRenderingAreaGranularityKHR> requires <VK_KHR_maintenance5>" );
25179 VULKAN_HPP_NAMESPACE::Extent2D granularity;
25180 d.vkGetRenderingAreaGranularityKHR(
25183 return granularity;
25187 template <
typename Dispatch>
25188 VULKAN_HPP_INLINE void Device::getImageSubresourceLayoutKHR(
const VULKAN_HPP_NAMESPACE::DeviceImageSubresourceInfoKHR * pInfo,
25189 VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR * pLayout,
25193 d.vkGetDeviceImageSubresourceLayoutKHR(
25197 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25198 template <
typename Dispatch>
25200 Device::getImageSubresourceLayoutKHR(
const VULKAN_HPP_NAMESPACE::DeviceImageSubresourceInfoKHR & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
25203 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25204 VULKAN_HPP_ASSERT( d.vkGetDeviceImageSubresourceLayoutKHR &&
"Function <vkGetDeviceImageSubresourceLayoutKHR> requires <VK_KHR_maintenance5>" );
25207 VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR layout;
25208 d.vkGetDeviceImageSubresourceLayoutKHR(
25214 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
25216 Device::getImageSubresourceLayoutKHR(
const VULKAN_HPP_NAMESPACE::DeviceImageSubresourceInfoKHR & info, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
25219 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25220 VULKAN_HPP_ASSERT( d.vkGetDeviceImageSubresourceLayoutKHR &&
"Function <vkGetDeviceImageSubresourceLayoutKHR> requires <VK_KHR_maintenance5>" );
25223 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
25224 VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR & layout = structureChain.template get<VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR>();
25225 d.vkGetDeviceImageSubresourceLayoutKHR(
25228 return structureChain;
25232 template <
typename Dispatch>
25234 const VULKAN_HPP_NAMESPACE::ImageSubresource2KHR * pSubresource,
25235 VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR * pLayout,
25239 d.vkGetImageSubresourceLayout2KHR( m_device,
25240 static_cast<VkImage>( image ),
25245 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25246 template <
typename Dispatch>
25248 VULKAN_HPP_NAMESPACE::Image image,
const VULKAN_HPP_NAMESPACE::ImageSubresource2KHR & subresource, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
25251 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25253 d.vkGetImageSubresourceLayout2KHR &&
25254 "Function <vkGetImageSubresourceLayout2KHR> requires <VK_EXT_host_image_copy> or <VK_EXT_image_compression_control> or <VK_KHR_maintenance5>" );
25257 VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR layout;
25258 d.vkGetImageSubresourceLayout2KHR( m_device,
25259 static_cast<VkImage>( image ),
25266 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
25268 VULKAN_HPP_NAMESPACE::Image image,
const VULKAN_HPP_NAMESPACE::ImageSubresource2KHR & subresource, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
25271 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25273 d.vkGetImageSubresourceLayout2KHR &&
25274 "Function <vkGetImageSubresourceLayout2KHR> requires <VK_EXT_host_image_copy> or <VK_EXT_image_compression_control> or <VK_KHR_maintenance5>" );
25277 VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
25278 VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR & layout = structureChain.template get<VULKAN_HPP_NAMESPACE::SubresourceLayout2KHR>();
25279 d.vkGetImageSubresourceLayout2KHR( m_device,
25280 static_cast<VkImage>( image ),
25284 return structureChain;
25290 template <
typename Dispatch>
25292 const VULKAN_HPP_NAMESPACE::ShaderCreateInfoEXT * pCreateInfos,
25293 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
25294 VULKAN_HPP_NAMESPACE::ShaderEXT * pShaders,
25298 return static_cast<Result>( d.vkCreateShadersEXT( m_device,
25305 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25306 template <
typename ShaderEXTAllocator,
typename Dispatch>
25308 Device::createShadersEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ShaderCreateInfoEXT>
const & createInfos,
25310 Dispatch
const & d )
const
25313 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25314 VULKAN_HPP_ASSERT( d.vkCreateShadersEXT &&
"Function <vkCreateShadersEXT> requires <VK_EXT_shader_object>" );
25317 std::vector<VULKAN_HPP_NAMESPACE::ShaderEXT, ShaderEXTAllocator> shaders( createInfos.size() );
25319 d.vkCreateShadersEXT( m_device,
25320 createInfos.size(),
25322 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
25323 reinterpret_cast<VkShaderEXT *
>( shaders.data() ) ) );
25329 template <
typename ShaderEXTAllocator,
25331 typename std::enable_if<std::is_same<typename ShaderEXTAllocator::value_type, VULKAN_HPP_NAMESPACE::ShaderEXT>::value,
int>::type>
25333 Device::createShadersEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ShaderCreateInfoEXT>
const & createInfos,
25335 ShaderEXTAllocator & shaderEXTAllocator,
25336 Dispatch
const & d )
const
25339 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25340 VULKAN_HPP_ASSERT( d.vkCreateShadersEXT &&
"Function <vkCreateShadersEXT> requires <VK_EXT_shader_object>" );
25343 std::vector<VULKAN_HPP_NAMESPACE::ShaderEXT, ShaderEXTAllocator> shaders( createInfos.size(), shaderEXTAllocator );
25345 d.vkCreateShadersEXT( m_device,
25346 createInfos.size(),
25348 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
25349 reinterpret_cast<VkShaderEXT *
>( shaders.data() ) ) );
25355 template <
typename Dispatch>
25357 Device::createShaderEXT(
const VULKAN_HPP_NAMESPACE::ShaderCreateInfoEXT & createInfo,
25359 Dispatch
const & d )
const
25362 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25363 VULKAN_HPP_ASSERT( d.vkCreateShadersEXT &&
"Function <vkCreateShadersEXT> requires <VK_EXT_shader_object>" );
25366 VULKAN_HPP_NAMESPACE::ShaderEXT shader;
25368 d.vkCreateShadersEXT( m_device,
25371 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
25378 # ifndef VULKAN_HPP_NO_SMART_HANDLE
25379 template <
typename Dispatch,
typename ShaderEXTAllocator>
25382 Device::createShadersEXTUnique( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ShaderCreateInfoEXT>
const & createInfos,
25384 Dispatch
const & d )
const
25387 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25388 VULKAN_HPP_ASSERT( d.vkCreateShadersEXT &&
"Function <vkCreateShadersEXT> requires <VK_EXT_shader_object>" );
25391 std::vector<VULKAN_HPP_NAMESPACE::ShaderEXT> shaders( createInfos.size() );
25393 d.vkCreateShadersEXT( m_device,
25394 createInfos.size(),
25396 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
25397 reinterpret_cast<VkShaderEXT *
>( shaders.data() ) ) );
25399 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::ShaderEXT, Dispatch>, ShaderEXTAllocator> uniqueShaders;
25400 uniqueShaders.reserve( createInfos.size() );
25402 for (
auto const & shader : shaders )
25411 typename ShaderEXTAllocator,
25412 typename std::enable_if<std::is_same<typename ShaderEXTAllocator::value_type, UniqueHandle<VULKAN_HPP_NAMESPACE::ShaderEXT, Dispatch>>::value,
int>::type>
25415 Device::createShadersEXTUnique( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ShaderCreateInfoEXT>
const & createInfos,
25417 ShaderEXTAllocator & shaderEXTAllocator,
25418 Dispatch
const & d )
const
25421 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25422 VULKAN_HPP_ASSERT( d.vkCreateShadersEXT &&
"Function <vkCreateShadersEXT> requires <VK_EXT_shader_object>" );
25425 std::vector<VULKAN_HPP_NAMESPACE::ShaderEXT> shaders( createInfos.size() );
25427 d.vkCreateShadersEXT( m_device,
25428 createInfos.size(),
25430 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
25431 reinterpret_cast<VkShaderEXT *
>( shaders.data() ) ) );
25433 std::vector<UniqueHandle<VULKAN_HPP_NAMESPACE::ShaderEXT, Dispatch>, ShaderEXTAllocator> uniqueShaders( shaderEXTAllocator );
25434 uniqueShaders.reserve( createInfos.size() );
25436 for (
auto const & shader : shaders )
25443 template <
typename Dispatch>
25445 Device::createShaderEXTUnique(
const VULKAN_HPP_NAMESPACE::ShaderCreateInfoEXT & createInfo,
25447 Dispatch
const & d )
const
25450 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25451 VULKAN_HPP_ASSERT( d.vkCreateShadersEXT &&
"Function <vkCreateShadersEXT> requires <VK_EXT_shader_object>" );
25454 VULKAN_HPP_NAMESPACE::ShaderEXT shader;
25456 d.vkCreateShadersEXT( m_device,
25459 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ),
25469 template <
typename Dispatch>
25471 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
25478 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25479 template <
typename Dispatch>
25485 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25486 VULKAN_HPP_ASSERT( d.vkDestroyShaderEXT &&
"Function <vkDestroyShaderEXT> requires <VK_EXT_shader_object>" );
25489 d.vkDestroyShaderEXT( m_device,
25491 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
25495 template <
typename Dispatch>
25497 const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
25504 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25505 template <
typename Dispatch>
25511 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25512 VULKAN_HPP_ASSERT( d.vkDestroyShaderEXT &&
"Function <vkDestroyShaderEXT> requires <VK_EXT_shader_object>" );
25515 d.vkDestroyShaderEXT( m_device,
25517 reinterpret_cast<const VkAllocationCallbacks *
>(
static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *
>( allocator ) ) );
25521 template <
typename Dispatch>
25523 Device::getShaderBinaryDataEXT( VULKAN_HPP_NAMESPACE::ShaderEXT shader,
size_t * pDataSize,
void * pData, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
25526 return static_cast<Result>( d.vkGetShaderBinaryDataEXT( m_device,
static_cast<VkShaderEXT>( shader ), pDataSize, pData ) );
25529 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25530 template <
typename U
int8_tAllocator,
typename Dispatch>
25532 Device::getShaderBinaryDataEXT( VULKAN_HPP_NAMESPACE::ShaderEXT shader, Dispatch
const & d )
const
25535 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25536 VULKAN_HPP_ASSERT( d.vkGetShaderBinaryDataEXT &&
"Function <vkGetShaderBinaryDataEXT> requires <VK_EXT_shader_object>" );
25539 std::vector<uint8_t, Uint8_tAllocator> data;
25545 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
25547 data.resize( dataSize );
25549 d.vkGetShaderBinaryDataEXT( m_device,
static_cast<VkShaderEXT>( shader ), &dataSize,
reinterpret_cast<void *
>( data.data() ) ) );
25551 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
25554 if ( dataSize < data.size() )
25556 data.resize( dataSize );
25561 template <
typename Uint8_tAllocator,
25563 typename std::enable_if<std::is_same<typename Uint8_tAllocator::value_type, uint8_t>::value,
int>::type>
25565 Device::getShaderBinaryDataEXT( VULKAN_HPP_NAMESPACE::ShaderEXT shader, Uint8_tAllocator & uint8_tAllocator, Dispatch
const & d )
const
25568 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25569 VULKAN_HPP_ASSERT( d.vkGetShaderBinaryDataEXT &&
"Function <vkGetShaderBinaryDataEXT> requires <VK_EXT_shader_object>" );
25572 std::vector<uint8_t, Uint8_tAllocator> data( uint8_tAllocator );
25578 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && dataSize )
25580 data.resize( dataSize );
25582 d.vkGetShaderBinaryDataEXT( m_device,
static_cast<VkShaderEXT>( shader ), &dataSize,
reinterpret_cast<void *
>( data.data() ) ) );
25584 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
25587 if ( dataSize < data.size() )
25589 data.resize( dataSize );
25595 template <
typename Dispatch>
25598 const VULKAN_HPP_NAMESPACE::ShaderEXT * pShaders,
25602 d.vkCmdBindShadersEXT(
25606 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25607 template <
typename Dispatch>
25608 VULKAN_HPP_INLINE void CommandBuffer::bindShadersEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ShaderStageFlagBits>
const & stages,
25609 VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ShaderEXT>
const & shaders,
25613 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25614 VULKAN_HPP_ASSERT( d.vkCmdBindShadersEXT &&
"Function <vkCmdBindShadersEXT> requires <VK_EXT_shader_object>" );
25616 # ifdef VULKAN_HPP_NO_EXCEPTIONS
25619 if ( stages.size() != shaders.size() )
25625 d.vkCmdBindShadersEXT( m_commandBuffer,
25628 reinterpret_cast<const VkShaderEXT *
>( shaders.data() ) );
25634 template <
typename Dispatch>
25636 uint32_t * pPropertiesCount,
25637 VULKAN_HPP_NAMESPACE::TilePropertiesQCOM * pProperties,
25641 return static_cast<Result>( d.vkGetFramebufferTilePropertiesQCOM(
25645 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25646 template <
typename TilePropertiesQCOMAllocator,
typename Dispatch>
25648 Device::getFramebufferTilePropertiesQCOM( VULKAN_HPP_NAMESPACE::Framebuffer framebuffer, Dispatch
const & d )
const
25651 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25652 VULKAN_HPP_ASSERT( d.vkGetFramebufferTilePropertiesQCOM &&
"Function <vkGetFramebufferTilePropertiesQCOM> requires <VK_QCOM_tile_properties>" );
25655 std::vector<VULKAN_HPP_NAMESPACE::TilePropertiesQCOM, TilePropertiesQCOMAllocator> properties;
25656 uint32_t propertiesCount;
25661 d.vkGetFramebufferTilePropertiesQCOM( m_device,
static_cast<VkFramebuffer>( framebuffer ), &propertiesCount,
nullptr ) );
25662 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertiesCount )
25664 properties.resize( propertiesCount );
25668 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
25671 if ( propertiesCount < properties.size() )
25673 properties.resize( propertiesCount );
25678 template <
typename TilePropertiesQCOMAllocator,
25680 typename std::enable_if<std::is_same<typename TilePropertiesQCOMAllocator::value_type, VULKAN_HPP_NAMESPACE::TilePropertiesQCOM>::value,
int>::type>
25682 Device::getFramebufferTilePropertiesQCOM( VULKAN_HPP_NAMESPACE::Framebuffer framebuffer,
25683 TilePropertiesQCOMAllocator & tilePropertiesQCOMAllocator,
25684 Dispatch
const & d )
const
25687 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25688 VULKAN_HPP_ASSERT( d.vkGetFramebufferTilePropertiesQCOM &&
"Function <vkGetFramebufferTilePropertiesQCOM> requires <VK_QCOM_tile_properties>" );
25691 std::vector<VULKAN_HPP_NAMESPACE::TilePropertiesQCOM, TilePropertiesQCOMAllocator> properties( tilePropertiesQCOMAllocator );
25692 uint32_t propertiesCount;
25697 d.vkGetFramebufferTilePropertiesQCOM( m_device,
static_cast<VkFramebuffer>( framebuffer ), &propertiesCount,
nullptr ) );
25698 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertiesCount )
25700 properties.resize( propertiesCount );
25704 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
25707 if ( propertiesCount < properties.size() )
25709 properties.resize( propertiesCount );
25715 template <
typename Dispatch>
25717 VULKAN_HPP_NAMESPACE::TilePropertiesQCOM * pProperties,
25721 return static_cast<Result>( d.vkGetDynamicRenderingTilePropertiesQCOM(
25725 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25726 template <
typename Dispatch>
25728 Device::getDynamicRenderingTilePropertiesQCOM(
const VULKAN_HPP_NAMESPACE::RenderingInfo & renderingInfo, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
25731 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25732 VULKAN_HPP_ASSERT( d.vkGetDynamicRenderingTilePropertiesQCOM &&
"Function <vkGetDynamicRenderingTilePropertiesQCOM> requires <VK_QCOM_tile_properties>" );
25735 VULKAN_HPP_NAMESPACE::TilePropertiesQCOM properties;
25736 d.vkGetDynamicRenderingTilePropertiesQCOM(
25745 template <
typename Dispatch>
25747 const VULKAN_HPP_NAMESPACE::LatencySleepModeInfoNV * pSleepModeInfo,
25751 return static_cast<Result>(
25755 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25756 template <
typename Dispatch>
25758 const VULKAN_HPP_NAMESPACE::LatencySleepModeInfoNV & sleepModeInfo,
25759 Dispatch
const & d )
const
25762 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25763 VULKAN_HPP_ASSERT( d.vkSetLatencySleepModeNV &&
"Function <vkSetLatencySleepModeNV> requires <VK_NV_low_latency2>" );
25774 template <
typename Dispatch>
25776 const VULKAN_HPP_NAMESPACE::LatencySleepInfoNV * pSleepInfo,
25780 return static_cast<Result>(
25784 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25785 template <
typename Dispatch>
25787 Device::latencySleepNV( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain,
const VULKAN_HPP_NAMESPACE::LatencySleepInfoNV & sleepInfo, Dispatch
const & d )
const
25790 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25791 VULKAN_HPP_ASSERT( d.vkLatencySleepNV &&
"Function <vkLatencySleepNV> requires <VK_NV_low_latency2>" );
25802 template <
typename Dispatch>
25804 const VULKAN_HPP_NAMESPACE::SetLatencyMarkerInfoNV * pLatencyMarkerInfo,
25811 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25812 template <
typename Dispatch>
25814 const VULKAN_HPP_NAMESPACE::SetLatencyMarkerInfoNV & latencyMarkerInfo,
25818 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25819 VULKAN_HPP_ASSERT( d.vkSetLatencyMarkerNV &&
"Function <vkSetLatencyMarkerNV> requires <VK_NV_low_latency2>" );
25826 template <
typename Dispatch>
25828 VULKAN_HPP_NAMESPACE::GetLatencyMarkerInfoNV * pLatencyMarkerInfo,
25835 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25836 template <
typename Dispatch>
25838 Device::getLatencyTimingsNV( VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
25841 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25842 VULKAN_HPP_ASSERT( d.vkGetLatencyTimingsNV &&
"Function <vkGetLatencyTimingsNV> requires <VK_NV_low_latency2>" );
25845 VULKAN_HPP_NAMESPACE::GetLatencyMarkerInfoNV latencyMarkerInfo;
25848 return latencyMarkerInfo;
25852 template <
typename Dispatch>
25853 VULKAN_HPP_INLINE void Queue::notifyOutOfBandNV(
const VULKAN_HPP_NAMESPACE::OutOfBandQueueTypeInfoNV * pQueueTypeInfo,
25860 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25861 template <
typename Dispatch>
25862 VULKAN_HPP_INLINE void Queue::notifyOutOfBandNV(
const VULKAN_HPP_NAMESPACE::OutOfBandQueueTypeInfoNV & queueTypeInfo,
25866 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25867 VULKAN_HPP_ASSERT( d.vkQueueNotifyOutOfBandNV &&
"Function <vkQueueNotifyOutOfBandNV> requires <VK_NV_low_latency2>" );
25876 template <
typename Dispatch>
25878 uint32_t * pPropertyCount, VULKAN_HPP_NAMESPACE::CooperativeMatrixPropertiesKHR * pProperties, Dispatch
const & d )
const VULKAN_HPP_NOEXCEPT
25881 return static_cast<Result>( d.vkGetPhysicalDeviceCooperativeMatrixPropertiesKHR(
25885 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25886 template <
typename CooperativeMatrixPropertiesKHRAllocator,
typename Dispatch>
25889 PhysicalDevice::getCooperativeMatrixPropertiesKHR( Dispatch
const & d )
const
25892 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25894 "Function <vkGetPhysicalDeviceCooperativeMatrixPropertiesKHR> requires <VK_KHR_cooperative_matrix>" );
25897 std::vector<VULKAN_HPP_NAMESPACE::CooperativeMatrixPropertiesKHR, CooperativeMatrixPropertiesKHRAllocator> properties;
25898 uint32_t propertyCount;
25902 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceCooperativeMatrixPropertiesKHR( m_physicalDevice, &propertyCount,
nullptr ) );
25903 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
25905 properties.resize( propertyCount );
25909 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
25912 if ( propertyCount < properties.size() )
25914 properties.resize( propertyCount );
25919 template <
typename CooperativeMatrixPropertiesKHRAllocator,
25921 typename std::enable_if<
25922 std::is_same<typename CooperativeMatrixPropertiesKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::CooperativeMatrixPropertiesKHR>::value,
25926 PhysicalDevice::getCooperativeMatrixPropertiesKHR( CooperativeMatrixPropertiesKHRAllocator & cooperativeMatrixPropertiesKHRAllocator,
25927 Dispatch
const & d )
const
25930 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25932 "Function <vkGetPhysicalDeviceCooperativeMatrixPropertiesKHR> requires <VK_KHR_cooperative_matrix>" );
25935 std::vector<VULKAN_HPP_NAMESPACE::CooperativeMatrixPropertiesKHR, CooperativeMatrixPropertiesKHRAllocator> properties(
25936 cooperativeMatrixPropertiesKHRAllocator );
25937 uint32_t propertyCount;
25941 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceCooperativeMatrixPropertiesKHR( m_physicalDevice, &propertyCount,
nullptr ) );
25942 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && propertyCount )
25944 properties.resize( propertyCount );
25948 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
25951 if ( propertyCount < properties.size() )
25953 properties.resize( propertyCount );
25961 template <
typename Dispatch>
25966 d.vkCmdSetAttachmentFeedbackLoopEnableEXT( m_commandBuffer,
static_cast<VkImageAspectFlags>( aspectMask ) );
25969 #if defined( VK_USE_PLATFORM_SCREEN_QNX )
25972 template <
typename Dispatch>
25974 VULKAN_HPP_NAMESPACE::ScreenBufferPropertiesQNX * pProperties,
25978 return static_cast<Result>( d.vkGetScreenBufferPropertiesQNX( m_device, buffer,
reinterpret_cast<VkScreenBufferPropertiesQNX *
>( pProperties ) ) );
25981 # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
25982 template <
typename Dispatch>
25984 Device::getScreenBufferPropertiesQNX(
const struct _screen_buffer & buffer, Dispatch
const & d )
const
25987 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
25988 VULKAN_HPP_ASSERT( d.vkGetScreenBufferPropertiesQNX &&
"Function <vkGetScreenBufferPropertiesQNX> requires <VK_QNX_external_memory_screen_buffer>" );
25991 VULKAN_HPP_NAMESPACE::ScreenBufferPropertiesQNX properties;
25993 d.vkGetScreenBufferPropertiesQNX( m_device, &buffer,
reinterpret_cast<VkScreenBufferPropertiesQNX *
>( &properties ) ) );
25999 template <
typename X,
typename Y,
typename... Z,
typename Dispatch>
26001 Device::getScreenBufferPropertiesQNX(
const struct _screen_buffer & buffer, Dispatch
const & d )
const
26004 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26005 VULKAN_HPP_ASSERT( d.vkGetScreenBufferPropertiesQNX &&
"Function <vkGetScreenBufferPropertiesQNX> requires <VK_QNX_external_memory_screen_buffer>" );
26008 StructureChain<X, Y, Z...> structureChain;
26009 VULKAN_HPP_NAMESPACE::ScreenBufferPropertiesQNX & properties = structureChain.template get<VULKAN_HPP_NAMESPACE::ScreenBufferPropertiesQNX>();
26011 d.vkGetScreenBufferPropertiesQNX( m_device, &buffer,
reinterpret_cast<VkScreenBufferPropertiesQNX *
>( &properties ) ) );
26021 template <
typename Dispatch>
26027 return static_cast<Result>(
26028 d.vkGetPhysicalDeviceCalibrateableTimeDomainsKHR( m_physicalDevice, pTimeDomainCount,
reinterpret_cast<VkTimeDomainKHR *
>( pTimeDomains ) ) );
26031 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
26032 template <
typename TimeDomainKHRAllocator,
typename Dispatch>
26034 PhysicalDevice::getCalibrateableTimeDomainsKHR( Dispatch
const & d )
const
26037 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26039 "Function <vkGetPhysicalDeviceCalibrateableTimeDomainsKHR> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
26042 std::vector<VULKAN_HPP_NAMESPACE::TimeDomainKHR, TimeDomainKHRAllocator> timeDomains;
26043 uint32_t timeDomainCount;
26047 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceCalibrateableTimeDomainsKHR( m_physicalDevice, &timeDomainCount,
nullptr ) );
26048 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && timeDomainCount )
26050 timeDomains.resize( timeDomainCount );
26052 d.vkGetPhysicalDeviceCalibrateableTimeDomainsKHR( m_physicalDevice, &timeDomainCount,
reinterpret_cast<VkTimeDomainKHR *
>( timeDomains.data() ) ) );
26054 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
26057 if ( timeDomainCount < timeDomains.size() )
26059 timeDomains.resize( timeDomainCount );
26064 template <
typename TimeDomainKHRAllocator,
26066 typename std::enable_if<std::is_same<typename TimeDomainKHRAllocator::value_type, VULKAN_HPP_NAMESPACE::TimeDomainKHR>::value,
int>::type>
26068 PhysicalDevice::getCalibrateableTimeDomainsKHR( TimeDomainKHRAllocator & timeDomainKHRAllocator, Dispatch
const & d )
const
26071 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26073 "Function <vkGetPhysicalDeviceCalibrateableTimeDomainsKHR> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
26076 std::vector<VULKAN_HPP_NAMESPACE::TimeDomainKHR, TimeDomainKHRAllocator> timeDomains( timeDomainKHRAllocator );
26077 uint32_t timeDomainCount;
26081 result =
static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkGetPhysicalDeviceCalibrateableTimeDomainsKHR( m_physicalDevice, &timeDomainCount,
nullptr ) );
26082 if ( ( result == VULKAN_HPP_NAMESPACE::Result::eSuccess ) && timeDomainCount )
26084 timeDomains.resize( timeDomainCount );
26086 d.vkGetPhysicalDeviceCalibrateableTimeDomainsKHR( m_physicalDevice, &timeDomainCount,
reinterpret_cast<VkTimeDomainKHR *
>( timeDomains.data() ) ) );
26088 }
while ( result == VULKAN_HPP_NAMESPACE::Result::eIncomplete );
26091 if ( timeDomainCount < timeDomains.size() )
26093 timeDomains.resize( timeDomainCount );
26099 template <
typename Dispatch>
26101 const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoKHR * pTimestampInfos,
26102 uint64_t * pTimestamps,
26103 uint64_t * pMaxDeviation,
26107 return static_cast<Result>( d.vkGetCalibratedTimestampsKHR(
26111 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
26112 template <
typename U
int64_tAllocator,
typename Dispatch>
26114 Device::getCalibratedTimestampsKHR( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoKHR>
const & timestampInfos,
26115 Dispatch
const & d )
const
26118 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26120 "Function <vkGetCalibratedTimestampsKHR> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
26123 std::pair<std::vector<uint64_t, Uint64_tAllocator>, uint64_t> data_(
26124 std::piecewise_construct, std::forward_as_tuple( timestampInfos.size() ), std::forward_as_tuple( 0 ) );
26125 std::vector<uint64_t, Uint64_tAllocator> & timestamps = data_.first;
26126 uint64_t & maxDeviation = data_.second;
26128 m_device, timestampInfos.size(),
reinterpret_cast<const VkCalibratedTimestampInfoKHR *
>( timestampInfos.data() ), timestamps.data(), &maxDeviation ) );
26134 template <
typename Uint64_tAllocator,
26136 typename std::enable_if<std::is_same<typename Uint64_tAllocator::value_type, uint64_t>::value,
int>::type>
26138 Device::getCalibratedTimestampsKHR( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoKHR>
const & timestampInfos,
26139 Uint64_tAllocator & uint64_tAllocator,
26140 Dispatch
const & d )
const
26143 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26145 "Function <vkGetCalibratedTimestampsKHR> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
26148 std::pair<std::vector<uint64_t, Uint64_tAllocator>, uint64_t> data_(
26149 std::piecewise_construct, std::forward_as_tuple( timestampInfos.size(), uint64_tAllocator ), std::forward_as_tuple( 0 ) );
26150 std::vector<uint64_t, Uint64_tAllocator> & timestamps = data_.first;
26151 uint64_t & maxDeviation = data_.second;
26153 m_device, timestampInfos.size(),
reinterpret_cast<const VkCalibratedTimestampInfoKHR *
>( timestampInfos.data() ), timestamps.data(), &maxDeviation ) );
26159 template <
typename Dispatch>
26161 Device::getCalibratedTimestampKHR(
const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoKHR & timestampInfo, Dispatch
const & d )
const
26164 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26166 "Function <vkGetCalibratedTimestampsKHR> requires <VK_EXT_calibrated_timestamps> or <VK_KHR_calibrated_timestamps>" );
26169 std::pair<uint64_t, uint64_t> data_;
26170 uint64_t & timestamp = data_.first;
26171 uint64_t & maxDeviation = data_.second;
26173 d.vkGetCalibratedTimestampsKHR( m_device, 1,
reinterpret_cast<const VkCalibratedTimestampInfoKHR *
>( ×tampInfo ), ×tamp, &maxDeviation ) );
26182 template <
typename Dispatch>
26183 VULKAN_HPP_INLINE void CommandBuffer::bindDescriptorSets2KHR(
const VULKAN_HPP_NAMESPACE::BindDescriptorSetsInfoKHR * pBindDescriptorSetsInfo,
26187 d.vkCmdBindDescriptorSets2KHR( m_commandBuffer,
reinterpret_cast<const VkBindDescriptorSetsInfoKHR *
>( pBindDescriptorSetsInfo ) );
26190 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
26191 template <
typename Dispatch>
26192 VULKAN_HPP_INLINE void CommandBuffer::bindDescriptorSets2KHR(
const VULKAN_HPP_NAMESPACE::BindDescriptorSetsInfoKHR & bindDescriptorSetsInfo,
26196 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26197 VULKAN_HPP_ASSERT( d.vkCmdBindDescriptorSets2KHR &&
"Function <vkCmdBindDescriptorSets2KHR> requires <VK_KHR_maintenance6>" );
26200 d.vkCmdBindDescriptorSets2KHR( m_commandBuffer,
reinterpret_cast<const VkBindDescriptorSetsInfoKHR *
>( &bindDescriptorSetsInfo ) );
26204 template <
typename Dispatch>
26205 VULKAN_HPP_INLINE void CommandBuffer::pushConstants2KHR(
const VULKAN_HPP_NAMESPACE::PushConstantsInfoKHR * pPushConstantsInfo,
26209 d.vkCmdPushConstants2KHR( m_commandBuffer,
reinterpret_cast<const VkPushConstantsInfoKHR *
>( pPushConstantsInfo ) );
26212 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
26213 template <
typename Dispatch>
26214 VULKAN_HPP_INLINE void CommandBuffer::pushConstants2KHR(
const VULKAN_HPP_NAMESPACE::PushConstantsInfoKHR & pushConstantsInfo,
26218 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26219 VULKAN_HPP_ASSERT( d.vkCmdPushConstants2KHR &&
"Function <vkCmdPushConstants2KHR> requires <VK_KHR_maintenance6>" );
26222 d.vkCmdPushConstants2KHR( m_commandBuffer,
reinterpret_cast<const VkPushConstantsInfoKHR *
>( &pushConstantsInfo ) );
26226 template <
typename Dispatch>
26227 VULKAN_HPP_INLINE void CommandBuffer::pushDescriptorSet2KHR(
const VULKAN_HPP_NAMESPACE::PushDescriptorSetInfoKHR * pPushDescriptorSetInfo,
26231 d.vkCmdPushDescriptorSet2KHR( m_commandBuffer,
reinterpret_cast<const VkPushDescriptorSetInfoKHR *
>( pPushDescriptorSetInfo ) );
26234 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
26235 template <
typename Dispatch>
26236 VULKAN_HPP_INLINE void CommandBuffer::pushDescriptorSet2KHR(
const VULKAN_HPP_NAMESPACE::PushDescriptorSetInfoKHR & pushDescriptorSetInfo,
26240 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26241 VULKAN_HPP_ASSERT( d.vkCmdPushDescriptorSet2KHR &&
"Function <vkCmdPushDescriptorSet2KHR> requires <VK_KHR_maintenance6>" );
26244 d.vkCmdPushDescriptorSet2KHR( m_commandBuffer,
reinterpret_cast<const VkPushDescriptorSetInfoKHR *
>( &pushDescriptorSetInfo ) );
26248 template <
typename Dispatch>
26250 CommandBuffer::pushDescriptorSetWithTemplate2KHR(
const VULKAN_HPP_NAMESPACE::PushDescriptorSetWithTemplateInfoKHR * pPushDescriptorSetWithTemplateInfo,
26254 d.vkCmdPushDescriptorSetWithTemplate2KHR( m_commandBuffer,
26258 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
26259 template <
typename Dispatch>
26261 CommandBuffer::pushDescriptorSetWithTemplate2KHR(
const VULKAN_HPP_NAMESPACE::PushDescriptorSetWithTemplateInfoKHR & pushDescriptorSetWithTemplateInfo,
26265 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26266 VULKAN_HPP_ASSERT( d.vkCmdPushDescriptorSetWithTemplate2KHR &&
"Function <vkCmdPushDescriptorSetWithTemplate2KHR> requires <VK_KHR_maintenance6>" );
26269 d.vkCmdPushDescriptorSetWithTemplate2KHR( m_commandBuffer,
26274 template <
typename Dispatch>
26276 CommandBuffer::setDescriptorBufferOffsets2EXT(
const VULKAN_HPP_NAMESPACE::SetDescriptorBufferOffsetsInfoEXT * pSetDescriptorBufferOffsetsInfo,
26283 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
26284 template <
typename Dispatch>
26286 CommandBuffer::setDescriptorBufferOffsets2EXT(
const VULKAN_HPP_NAMESPACE::SetDescriptorBufferOffsetsInfoEXT & setDescriptorBufferOffsetsInfo,
26290 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26291 VULKAN_HPP_ASSERT( d.vkCmdSetDescriptorBufferOffsets2EXT &&
"Function <vkCmdSetDescriptorBufferOffsets2EXT> requires <VK_KHR_maintenance6>" );
26298 template <
typename Dispatch>
26300 const VULKAN_HPP_NAMESPACE::BindDescriptorBufferEmbeddedSamplersInfoEXT * pBindDescriptorBufferEmbeddedSamplersInfo,
26304 d.vkCmdBindDescriptorBufferEmbeddedSamplers2EXT(
26308 #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
26309 template <
typename Dispatch>
26311 const VULKAN_HPP_NAMESPACE::BindDescriptorBufferEmbeddedSamplersInfoEXT & bindDescriptorBufferEmbeddedSamplersInfo,
26315 # if ( VULKAN_HPP_DISPATCH_LOADER_DYNAMIC == 1 )
26317 "Function <vkCmdBindDescriptorBufferEmbeddedSamplers2EXT> requires <VK_KHR_maintenance6>" );
26320 d.vkCmdBindDescriptorBufferEmbeddedSamplers2EXT(
vk::ShaderModule createShaderModule(vk::Device const &device, vk::ShaderStageFlagBits shaderStage, std::string const &shaderText)
vk::RenderPass createRenderPass(vk::Device const &device, vk::Format colorFormat, vk::Format depthFormat, vk::AttachmentLoadOp loadOp, vk::ImageLayout colorFinalLayout)
vk::Pipeline createGraphicsPipeline(vk::Device const &device, vk::PipelineCache const &pipelineCache, std::pair< vk::ShaderModule, vk::SpecializationInfo const * > const &vertexShaderData, std::pair< vk::ShaderModule, vk::SpecializationInfo const * > const &fragmentShaderData, uint32_t vertexStride, std::vector< std::pair< vk::Format, uint32_t >> const &vertexInputAttributeFormatOffset, vk::FrontFace frontFace, bool depthBuffered, vk::PipelineLayout const &pipelineLayout, vk::RenderPass const &renderPass)
vk::Device createDevice(vk::PhysicalDevice const &physicalDevice, uint32_t queueFamilyIndex, std::vector< std::string > const &extensions, vk::PhysicalDeviceFeatures const *physicalDeviceFeatures, void const *pNext)
vk::DescriptorSetLayout createDescriptorSetLayout(vk::Device const &device, std::vector< std::tuple< vk::DescriptorType, uint32_t, vk::ShaderStageFlags >> const &bindingData, vk::DescriptorSetLayoutCreateFlags flags)
void updateDescriptorSets(vk::Device const &device, vk::DescriptorSet const &descriptorSet, std::vector< std::tuple< vk::DescriptorType, vk::Buffer const &, vk::DeviceSize, vk::BufferView const & >> const &bufferData, vk::su::TextureData const &textureData, uint32_t bindingOffset)
vk::DescriptorPool createDescriptorPool(vk::Device const &device, std::vector< vk::DescriptorPoolSize > const &poolSizes)
Flags< CommandBufferResetFlagBits > CommandBufferResetFlags
SurfaceCounterFlagBitsEXT
Flags< CullModeFlagBits > CullModeFlags
Flags< DebugUtilsMessageTypeFlagBitsEXT > DebugUtilsMessageTypeFlagsEXT
AccelerationStructureBuildTypeKHR
Flags< ImageCreateFlagBits > ImageCreateFlags
OpticalFlowSessionBindingPointNV
Flags< ImageUsageFlagBits > ImageUsageFlags
Flags< ColorComponentFlagBits > ColorComponentFlags
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValueType< uint32_t >::type enumerateInstanceVersion(Dispatch const &d)
Flags< ImageAspectFlagBits > ImageAspectFlags
Flags< CommandPoolResetFlagBits > CommandPoolResetFlags
Flags< MemoryMapFlagBits > MemoryMapFlags
Flags< QueryControlFlagBits > QueryControlFlags
Flags< ExternalMemoryHandleTypeFlagBitsNV > ExternalMemoryHandleTypeFlagsNV
PerformanceParameterTypeINTEL
Flags< PipelineStageFlagBits2 > PipelineStageFlags2
VULKAN_HPP_INLINE void resultCheck(Result result, char const *message)
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValueType< vk::Instance >::type createInstance(const vk::InstanceCreateInfo &createInfo, Optional< const vk::AllocationCallbacks > allocator, Dispatch const &d)
AccelerationStructureCompatibilityKHR
Flags< DependencyFlagBits > DependencyFlags
Flags< ShaderStageFlagBits > ShaderStageFlags
FragmentShadingRateCombinerOpKHR
Flags< DebugReportFlagBitsEXT > DebugReportFlagsEXT
DebugUtilsMessageSeverityFlagBitsEXT
Flags< PeerMemoryFeatureFlagBits > PeerMemoryFeatureFlags
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValueType< std::vector< vk::ExtensionProperties, ExtensionPropertiesAllocator > >::type enumerateInstanceExtensionProperties(Optional< const std::string > layerName, ExtensionPropertiesAllocator &extensionPropertiesAllocator, Dispatch const &d)
VULKAN_HPP_INLINE ResultValueType< void >::type createResultValueType(Result result)
CopyAccelerationStructureModeKHR
ExternalMemoryHandleTypeFlagBits
Flags< QueryResultFlagBits > QueryResultFlags
ConservativeRasterizationModeEXT
Flags< PipelineStageFlagBits > PipelineStageFlags
MemoryRequirements2 MemoryRequirements2KHR
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValueType< std::vector< vk::LayerProperties, LayerPropertiesAllocator > >::type enumerateInstanceLayerProperties(LayerPropertiesAllocator &layerPropertiesAllocator, Dispatch const &d)
PipelineInfoKHR PipelineInfoEXT
Flags< DeviceGroupPresentModeFlagBitsKHR > DeviceGroupPresentModeFlagsKHR
Flags< CommandPoolTrimFlagBits > CommandPoolTrimFlags
Flags< StencilFaceFlagBits > StencilFaceFlags
Flags< DescriptorPoolResetFlagBits > DescriptorPoolResetFlags
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValueType< UniqueHandle< vk::Instance, Dispatch > >::type createInstanceUnique(const vk::InstanceCreateInfo &createInfo, Optional< const vk::AllocationCallbacks > allocator, Dispatch const &d)
uint64_t VkVideoSessionParametersKHR
VkFlags VkPipelineStageFlags
uint64_t VkIndirectCommandsLayoutNV
VkFlags VkImageAspectFlags
#define VK_HEADER_VERSION
VkFlags VkImageUsageFlags
VkTessellationDomainOrigin
struct VkPhysicalDevice_T * VkPhysicalDevice
VkFlags VkDependencyFlags
VkCoarseSampleOrderTypeNV
struct VkQueue_T * VkQueue
VkFlags VkStencilFaceFlags
VkFlags VkExternalMemoryHandleTypeFlagsNV
VkCoverageReductionModeNV
uint64_t VkDisplayModeKHR
VkFlags VkCommandPoolTrimFlags
VkAccelerationStructureCompatibilityKHR
VkOpticalFlowSessionBindingPointNV
struct VkInstance_T * VkInstance
struct VkDevice_T * VkDevice
uint64_t VkDeferredOperationKHR
VkSurfaceCounterFlagBitsEXT
VkDebugReportObjectTypeEXT
uint64_t VkOpticalFlowSessionNV
VkFlags VkPeerMemoryFeatureFlags
VkExternalMemoryHandleTypeFlagBits
VkFlags VkDebugUtilsMessageTypeFlagsEXT
VkFragmentShadingRateCombinerOpKHR
void(* PFN_vkVoidFunction)(void)
uint64_t VkDescriptorPool
VkConservativeRasterizationModeEXT
uint64_t VkAccelerationStructureNV
VkFlags VkQueryControlFlags
uint64_t VkPipelineLayout
VkFlags VkCommandPoolResetFlags
VkPerformanceParameterTypeINTEL
VkCoverageModulationModeNV
VkFlags VkShaderStageFlags
VkCopyAccelerationStructureModeKHR
uint64_t VkDescriptorUpdateTemplate
VkAccelerationStructureBuildTypeKHR
uint64_t VkDescriptorSetLayout
uint64_t VkAccelerationStructureKHR
uint64_t VkCudaFunctionNV
uint64_t VkSamplerYcbcrConversion
VkFlags VkDeviceGroupPresentModeFlagsKHR
VkFlags VkQueryResultFlags
uint64_t VkDebugUtilsMessengerEXT
VkFlags VkImageCreateFlags
struct VkCommandBuffer_T * VkCommandBuffer
VkFlags VkDebugReportFlagsEXT
uint64_t VkDebugReportCallbackEXT
VkLineRasterizationModeEXT
uint64_t VkPrivateDataSlot
VkFlags VkDescriptorPoolResetFlags
uint64_t VkPerformanceConfigurationINTEL
VkFlags64 VkPipelineStageFlags2
VkDiscardRectangleModeEXT
uint64_t VkValidationCacheEXT
VkFlags VkCommandBufferResetFlags
uint64_t VkVideoSessionKHR
VkFlags VkColorComponentFlags
VkDebugUtilsMessageSeverityFlagBitsEXT
#define VULKAN_HPP_ASSERT
#define VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS
#define VULKAN_HPP_NODISCARD
#define VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS
#define VULKAN_HPP_NAMESPACE_STRING
#define VULKAN_HPP_NOEXCEPT
#define VULKAN_HPP_INLINE