mirror of
https://gitlab.freedesktop.org/mesa/mesa.git
synced 2026-01-10 16:50:13 +01:00
radv: move buffer related code to radv_buffer.c
radv_device.c is getting too big and this follows the Vulkan common runtime infrastructure. Signed-off-by: Samuel Pitoiset <samuel.pitoiset@gmail.com> Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/21767>
This commit is contained in:
parent
17c5a91028
commit
4316a64e27
3 changed files with 269 additions and 238 deletions
|
|
@ -60,6 +60,7 @@ libradv_files = files(
|
|||
'winsys/null/radv_null_winsys_public.h',
|
||||
'radv_acceleration_structure.c',
|
||||
'radv_android.c',
|
||||
'radv_buffer.c',
|
||||
'radv_cmd_buffer.c',
|
||||
'radv_cp_reg_shadowing.c',
|
||||
'radv_cs.h',
|
||||
|
|
|
|||
268
src/amd/vulkan/radv_buffer.c
Normal file
268
src/amd/vulkan/radv_buffer.c
Normal file
|
|
@ -0,0 +1,268 @@
|
|||
/*
|
||||
* Copyright © 2016 Red Hat.
|
||||
* Copyright © 2016 Bas Nieuwenhuizen
|
||||
*
|
||||
* based in part on anv driver which is:
|
||||
* Copyright © 2015 Intel Corporation
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice (including the next
|
||||
* paragraph) shall be included in all copies or substantial portions of the
|
||||
* Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
||||
* IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
#include "radv_private.h"
|
||||
|
||||
#include "vk_buffer.h"
|
||||
|
||||
void
|
||||
radv_buffer_init(struct radv_buffer *buffer, struct radv_device *device,
|
||||
struct radeon_winsys_bo *bo, uint64_t size,
|
||||
uint64_t offset)
|
||||
{
|
||||
VkBufferCreateInfo createInfo = {
|
||||
.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO,
|
||||
.size = size,
|
||||
};
|
||||
|
||||
vk_buffer_init(&device->vk, &buffer->vk, &createInfo);
|
||||
|
||||
buffer->bo = bo;
|
||||
buffer->offset = offset;
|
||||
}
|
||||
|
||||
void
|
||||
radv_buffer_finish(struct radv_buffer *buffer)
|
||||
{
|
||||
vk_buffer_finish(&buffer->vk);
|
||||
}
|
||||
|
||||
static void
|
||||
radv_destroy_buffer(struct radv_device *device, const VkAllocationCallbacks *pAllocator,
|
||||
struct radv_buffer *buffer)
|
||||
{
|
||||
if ((buffer->vk.create_flags & VK_BUFFER_CREATE_SPARSE_BINDING_BIT) && buffer->bo)
|
||||
device->ws->buffer_destroy(device->ws, buffer->bo);
|
||||
|
||||
radv_rmv_log_resource_destroy(device, (uint64_t)radv_buffer_to_handle(buffer));
|
||||
radv_buffer_finish(buffer);
|
||||
vk_free2(&device->vk.alloc, pAllocator, buffer);
|
||||
}
|
||||
|
||||
VkResult
|
||||
radv_create_buffer(struct radv_device *device, const VkBufferCreateInfo *pCreateInfo,
|
||||
const VkAllocationCallbacks *pAllocator, VkBuffer *pBuffer, bool is_internal)
|
||||
{
|
||||
struct radv_buffer *buffer;
|
||||
|
||||
assert(pCreateInfo->sType == VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO);
|
||||
|
||||
#ifdef ANDROID
|
||||
/* reject buffers that are larger than maxBufferSize on Android, which
|
||||
* might not have VK_KHR_maintenance4
|
||||
*/
|
||||
if (pCreateInfo->size > RADV_MAX_MEMORY_ALLOCATION_SIZE)
|
||||
return vk_error(device, VK_ERROR_OUT_OF_DEVICE_MEMORY);
|
||||
#endif
|
||||
|
||||
buffer = vk_alloc2(&device->vk.alloc, pAllocator, sizeof(*buffer), 8,
|
||||
VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
|
||||
if (buffer == NULL)
|
||||
return vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
|
||||
|
||||
vk_buffer_init(&device->vk, &buffer->vk, pCreateInfo);
|
||||
buffer->bo = NULL;
|
||||
buffer->offset = 0;
|
||||
|
||||
if (pCreateInfo->flags & VK_BUFFER_CREATE_SPARSE_BINDING_BIT) {
|
||||
enum radeon_bo_flag flags = RADEON_FLAG_VIRTUAL;
|
||||
if (pCreateInfo->flags & VK_BUFFER_CREATE_DEVICE_ADDRESS_CAPTURE_REPLAY_BIT)
|
||||
flags |= RADEON_FLAG_REPLAYABLE;
|
||||
|
||||
uint64_t replay_address = 0;
|
||||
const VkBufferOpaqueCaptureAddressCreateInfo *replay_info =
|
||||
vk_find_struct_const(pCreateInfo->pNext, BUFFER_OPAQUE_CAPTURE_ADDRESS_CREATE_INFO);
|
||||
if (replay_info && replay_info->opaqueCaptureAddress)
|
||||
replay_address = replay_info->opaqueCaptureAddress;
|
||||
|
||||
VkResult result =
|
||||
device->ws->buffer_create(device->ws, align64(buffer->vk.size, 4096), 4096, 0, flags,
|
||||
RADV_BO_PRIORITY_VIRTUAL, replay_address, &buffer->bo);
|
||||
if (result != VK_SUCCESS) {
|
||||
radv_destroy_buffer(device, pAllocator, buffer);
|
||||
return vk_error(device, result);
|
||||
}
|
||||
radv_rmv_log_bo_allocate(device, buffer->bo, buffer->vk.size, true);
|
||||
}
|
||||
|
||||
*pBuffer = radv_buffer_to_handle(buffer);
|
||||
vk_rmv_log_buffer_create(&device->vk, false, *pBuffer);
|
||||
if (buffer->bo)
|
||||
radv_rmv_log_buffer_bind(device, *pBuffer);
|
||||
return VK_SUCCESS;
|
||||
}
|
||||
|
||||
VKAPI_ATTR VkResult VKAPI_CALL
|
||||
radv_CreateBuffer(VkDevice _device, const VkBufferCreateInfo *pCreateInfo,
|
||||
const VkAllocationCallbacks *pAllocator, VkBuffer *pBuffer)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
return radv_create_buffer(device, pCreateInfo, pAllocator, pBuffer, false);
|
||||
}
|
||||
|
||||
VKAPI_ATTR void VKAPI_CALL
|
||||
radv_DestroyBuffer(VkDevice _device, VkBuffer _buffer, const VkAllocationCallbacks *pAllocator)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, _buffer);
|
||||
|
||||
if (!buffer)
|
||||
return;
|
||||
|
||||
radv_destroy_buffer(device, pAllocator, buffer);
|
||||
}
|
||||
|
||||
VKAPI_ATTR VkResult VKAPI_CALL
|
||||
radv_BindBufferMemory2(VkDevice _device, uint32_t bindInfoCount,
|
||||
const VkBindBufferMemoryInfo *pBindInfos)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
|
||||
for (uint32_t i = 0; i < bindInfoCount; ++i) {
|
||||
RADV_FROM_HANDLE(radv_device_memory, mem, pBindInfos[i].memory);
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, pBindInfos[i].buffer);
|
||||
|
||||
if (mem->alloc_size) {
|
||||
VkBufferMemoryRequirementsInfo2 info = {
|
||||
.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_REQUIREMENTS_INFO_2,
|
||||
.buffer = pBindInfos[i].buffer,
|
||||
};
|
||||
VkMemoryRequirements2 reqs = {
|
||||
.sType = VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2,
|
||||
};
|
||||
|
||||
radv_GetBufferMemoryRequirements2(_device, &info, &reqs);
|
||||
|
||||
if (pBindInfos[i].memoryOffset + reqs.memoryRequirements.size > mem->alloc_size) {
|
||||
return vk_errorf(device, VK_ERROR_UNKNOWN,
|
||||
"Device memory object too small for the buffer.\n");
|
||||
}
|
||||
}
|
||||
|
||||
buffer->bo = mem->bo;
|
||||
buffer->offset = pBindInfos[i].memoryOffset;
|
||||
radv_rmv_log_buffer_bind(device, pBindInfos[i].buffer);
|
||||
}
|
||||
return VK_SUCCESS;
|
||||
}
|
||||
|
||||
static void
|
||||
radv_get_buffer_memory_requirements(struct radv_device *device, VkDeviceSize size,
|
||||
VkBufferCreateFlags flags, VkBufferCreateFlags usage,
|
||||
VkMemoryRequirements2 *pMemoryRequirements)
|
||||
{
|
||||
pMemoryRequirements->memoryRequirements.memoryTypeBits =
|
||||
((1u << device->physical_device->memory_properties.memoryTypeCount) - 1u) &
|
||||
~device->physical_device->memory_types_32bit;
|
||||
|
||||
/* Allow 32-bit address-space for DGC usage, as this buffer will contain
|
||||
* cmd buffer upload buffers, and those get passed to shaders through 32-bit
|
||||
* pointers.
|
||||
*
|
||||
* We only allow it with this usage set, to "protect" the 32-bit address space
|
||||
* from being overused. The actual requirement is done as part of
|
||||
* vkGetGeneratedCommandsMemoryRequirementsNV. (we have to make sure their
|
||||
* intersection is non-zero at least)
|
||||
*/
|
||||
if ((usage & VK_BUFFER_USAGE_INDIRECT_BUFFER_BIT) && device->uses_device_generated_commands)
|
||||
pMemoryRequirements->memoryRequirements.memoryTypeBits |=
|
||||
device->physical_device->memory_types_32bit;
|
||||
|
||||
/* Force 32-bit address-space for descriptor buffers usage because they are passed to shaders
|
||||
* through 32-bit pointers.
|
||||
*/
|
||||
if (usage & (VK_BUFFER_USAGE_RESOURCE_DESCRIPTOR_BUFFER_BIT_EXT |
|
||||
VK_BUFFER_USAGE_SAMPLER_DESCRIPTOR_BUFFER_BIT_EXT))
|
||||
pMemoryRequirements->memoryRequirements.memoryTypeBits =
|
||||
device->physical_device->memory_types_32bit;
|
||||
|
||||
if (flags & VK_BUFFER_CREATE_SPARSE_BINDING_BIT)
|
||||
pMemoryRequirements->memoryRequirements.alignment = 4096;
|
||||
else
|
||||
pMemoryRequirements->memoryRequirements.alignment = 16;
|
||||
|
||||
/* Top level acceleration structures need the bottom 6 bits to store
|
||||
* the root ids of instances. The hardware also needs bvh nodes to
|
||||
* be 64 byte aligned.
|
||||
*/
|
||||
if (usage & VK_BUFFER_USAGE_ACCELERATION_STRUCTURE_STORAGE_BIT_KHR)
|
||||
pMemoryRequirements->memoryRequirements.alignment =
|
||||
MAX2(pMemoryRequirements->memoryRequirements.alignment, 64);
|
||||
|
||||
pMemoryRequirements->memoryRequirements.size =
|
||||
align64(size, pMemoryRequirements->memoryRequirements.alignment);
|
||||
|
||||
vk_foreach_struct(ext, pMemoryRequirements->pNext)
|
||||
{
|
||||
switch (ext->sType) {
|
||||
case VK_STRUCTURE_TYPE_MEMORY_DEDICATED_REQUIREMENTS: {
|
||||
VkMemoryDedicatedRequirements *req = (VkMemoryDedicatedRequirements *)ext;
|
||||
req->requiresDedicatedAllocation = false;
|
||||
req->prefersDedicatedAllocation = req->requiresDedicatedAllocation;
|
||||
break;
|
||||
}
|
||||
default:
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
VKAPI_ATTR void VKAPI_CALL
|
||||
radv_GetBufferMemoryRequirements2(VkDevice _device, const VkBufferMemoryRequirementsInfo2 *pInfo,
|
||||
VkMemoryRequirements2 *pMemoryRequirements)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, pInfo->buffer);
|
||||
|
||||
radv_get_buffer_memory_requirements(device, buffer->vk.size, buffer->vk.create_flags,
|
||||
buffer->vk.usage, pMemoryRequirements);
|
||||
}
|
||||
|
||||
VKAPI_ATTR void VKAPI_CALL
|
||||
radv_GetDeviceBufferMemoryRequirements(VkDevice _device,
|
||||
const VkDeviceBufferMemoryRequirements *pInfo,
|
||||
VkMemoryRequirements2 *pMemoryRequirements)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
|
||||
radv_get_buffer_memory_requirements(device, pInfo->pCreateInfo->size, pInfo->pCreateInfo->flags,
|
||||
pInfo->pCreateInfo->usage, pMemoryRequirements);
|
||||
}
|
||||
|
||||
VKAPI_ATTR VkDeviceAddress VKAPI_CALL
|
||||
radv_GetBufferDeviceAddress(VkDevice device, const VkBufferDeviceAddressInfo *pInfo)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, pInfo->buffer);
|
||||
return radv_buffer_get_va(buffer->bo) + buffer->offset;
|
||||
}
|
||||
|
||||
VKAPI_ATTR uint64_t VKAPI_CALL
|
||||
radv_GetBufferOpaqueCaptureAddress(VkDevice device, const VkBufferDeviceAddressInfo *pInfo)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, pInfo->buffer);
|
||||
return buffer->bo ? radv_buffer_get_va(buffer->bo) + buffer->offset : 0;
|
||||
}
|
||||
|
|
@ -1506,89 +1506,6 @@ radv_InvalidateMappedMemoryRanges(VkDevice _device, uint32_t memoryRangeCount,
|
|||
return VK_SUCCESS;
|
||||
}
|
||||
|
||||
static void
|
||||
radv_get_buffer_memory_requirements(struct radv_device *device, VkDeviceSize size,
|
||||
VkBufferCreateFlags flags, VkBufferCreateFlags usage,
|
||||
VkMemoryRequirements2 *pMemoryRequirements)
|
||||
{
|
||||
pMemoryRequirements->memoryRequirements.memoryTypeBits =
|
||||
((1u << device->physical_device->memory_properties.memoryTypeCount) - 1u) &
|
||||
~device->physical_device->memory_types_32bit;
|
||||
|
||||
/* Allow 32-bit address-space for DGC usage, as this buffer will contain
|
||||
* cmd buffer upload buffers, and those get passed to shaders through 32-bit
|
||||
* pointers.
|
||||
*
|
||||
* We only allow it with this usage set, to "protect" the 32-bit address space
|
||||
* from being overused. The actual requirement is done as part of
|
||||
* vkGetGeneratedCommandsMemoryRequirementsNV. (we have to make sure their
|
||||
* intersection is non-zero at least)
|
||||
*/
|
||||
if ((usage & VK_BUFFER_USAGE_INDIRECT_BUFFER_BIT) && device->uses_device_generated_commands)
|
||||
pMemoryRequirements->memoryRequirements.memoryTypeBits |=
|
||||
device->physical_device->memory_types_32bit;
|
||||
|
||||
/* Force 32-bit address-space for descriptor buffers usage because they are passed to shaders
|
||||
* through 32-bit pointers.
|
||||
*/
|
||||
if (usage & (VK_BUFFER_USAGE_RESOURCE_DESCRIPTOR_BUFFER_BIT_EXT |
|
||||
VK_BUFFER_USAGE_SAMPLER_DESCRIPTOR_BUFFER_BIT_EXT))
|
||||
pMemoryRequirements->memoryRequirements.memoryTypeBits =
|
||||
device->physical_device->memory_types_32bit;
|
||||
|
||||
if (flags & VK_BUFFER_CREATE_SPARSE_BINDING_BIT)
|
||||
pMemoryRequirements->memoryRequirements.alignment = 4096;
|
||||
else
|
||||
pMemoryRequirements->memoryRequirements.alignment = 16;
|
||||
|
||||
/* Top level acceleration structures need the bottom 6 bits to store
|
||||
* the root ids of instances. The hardware also needs bvh nodes to
|
||||
* be 64 byte aligned.
|
||||
*/
|
||||
if (usage & VK_BUFFER_USAGE_ACCELERATION_STRUCTURE_STORAGE_BIT_KHR)
|
||||
pMemoryRequirements->memoryRequirements.alignment =
|
||||
MAX2(pMemoryRequirements->memoryRequirements.alignment, 64);
|
||||
|
||||
pMemoryRequirements->memoryRequirements.size =
|
||||
align64(size, pMemoryRequirements->memoryRequirements.alignment);
|
||||
|
||||
vk_foreach_struct(ext, pMemoryRequirements->pNext)
|
||||
{
|
||||
switch (ext->sType) {
|
||||
case VK_STRUCTURE_TYPE_MEMORY_DEDICATED_REQUIREMENTS: {
|
||||
VkMemoryDedicatedRequirements *req = (VkMemoryDedicatedRequirements *)ext;
|
||||
req->requiresDedicatedAllocation = false;
|
||||
req->prefersDedicatedAllocation = req->requiresDedicatedAllocation;
|
||||
break;
|
||||
}
|
||||
default:
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
VKAPI_ATTR void VKAPI_CALL
|
||||
radv_GetBufferMemoryRequirements2(VkDevice _device, const VkBufferMemoryRequirementsInfo2 *pInfo,
|
||||
VkMemoryRequirements2 *pMemoryRequirements)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, pInfo->buffer);
|
||||
|
||||
radv_get_buffer_memory_requirements(device, buffer->vk.size, buffer->vk.create_flags,
|
||||
buffer->vk.usage, pMemoryRequirements);
|
||||
}
|
||||
|
||||
VKAPI_ATTR void VKAPI_CALL
|
||||
radv_GetDeviceBufferMemoryRequirements(VkDevice _device,
|
||||
const VkDeviceBufferMemoryRequirements *pInfo,
|
||||
VkMemoryRequirements2 *pMemoryRequirements)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
|
||||
radv_get_buffer_memory_requirements(device, pInfo->pCreateInfo->size, pInfo->pCreateInfo->flags,
|
||||
pInfo->pCreateInfo->usage, pMemoryRequirements);
|
||||
}
|
||||
|
||||
VKAPI_ATTR void VKAPI_CALL
|
||||
radv_GetImageMemoryRequirements2(VkDevice _device, const VkImageMemoryRequirementsInfo2 *pInfo,
|
||||
VkMemoryRequirements2 *pMemoryRequirements)
|
||||
|
|
@ -1652,40 +1569,6 @@ radv_GetDeviceMemoryCommitment(VkDevice device, VkDeviceMemory memory,
|
|||
*pCommittedMemoryInBytes = 0;
|
||||
}
|
||||
|
||||
VKAPI_ATTR VkResult VKAPI_CALL
|
||||
radv_BindBufferMemory2(VkDevice _device, uint32_t bindInfoCount,
|
||||
const VkBindBufferMemoryInfo *pBindInfos)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
|
||||
for (uint32_t i = 0; i < bindInfoCount; ++i) {
|
||||
RADV_FROM_HANDLE(radv_device_memory, mem, pBindInfos[i].memory);
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, pBindInfos[i].buffer);
|
||||
|
||||
if (mem->alloc_size) {
|
||||
VkBufferMemoryRequirementsInfo2 info = {
|
||||
.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_REQUIREMENTS_INFO_2,
|
||||
.buffer = pBindInfos[i].buffer,
|
||||
};
|
||||
VkMemoryRequirements2 reqs = {
|
||||
.sType = VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2,
|
||||
};
|
||||
|
||||
radv_GetBufferMemoryRequirements2(_device, &info, &reqs);
|
||||
|
||||
if (pBindInfos[i].memoryOffset + reqs.memoryRequirements.size > mem->alloc_size) {
|
||||
return vk_errorf(device, VK_ERROR_UNKNOWN,
|
||||
"Device memory object too small for the buffer.\n");
|
||||
}
|
||||
}
|
||||
|
||||
buffer->bo = mem->bo;
|
||||
buffer->offset = pBindInfos[i].memoryOffset;
|
||||
radv_rmv_log_buffer_bind(device, pBindInfos[i].buffer);
|
||||
}
|
||||
return VK_SUCCESS;
|
||||
}
|
||||
|
||||
VKAPI_ATTR VkResult VKAPI_CALL
|
||||
radv_BindImageMemory2(VkDevice _device, uint32_t bindInfoCount,
|
||||
const VkBindImageMemoryInfo *pBindInfos)
|
||||
|
|
@ -1758,127 +1641,6 @@ radv_BindImageMemory2(VkDevice _device, uint32_t bindInfoCount,
|
|||
return VK_SUCCESS;
|
||||
}
|
||||
|
||||
void
|
||||
radv_buffer_init(struct radv_buffer *buffer, struct radv_device *device,
|
||||
struct radeon_winsys_bo *bo, uint64_t size,
|
||||
uint64_t offset)
|
||||
{
|
||||
VkBufferCreateInfo createInfo = {
|
||||
.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO,
|
||||
.size = size,
|
||||
};
|
||||
|
||||
vk_buffer_init(&device->vk, &buffer->vk, &createInfo);
|
||||
|
||||
buffer->bo = bo;
|
||||
buffer->offset = offset;
|
||||
}
|
||||
|
||||
void
|
||||
radv_buffer_finish(struct radv_buffer *buffer)
|
||||
{
|
||||
vk_buffer_finish(&buffer->vk);
|
||||
}
|
||||
|
||||
static void
|
||||
radv_destroy_buffer(struct radv_device *device, const VkAllocationCallbacks *pAllocator,
|
||||
struct radv_buffer *buffer)
|
||||
{
|
||||
if ((buffer->vk.create_flags & VK_BUFFER_CREATE_SPARSE_BINDING_BIT) && buffer->bo)
|
||||
device->ws->buffer_destroy(device->ws, buffer->bo);
|
||||
|
||||
radv_rmv_log_resource_destroy(device, (uint64_t)radv_buffer_to_handle(buffer));
|
||||
radv_buffer_finish(buffer);
|
||||
vk_free2(&device->vk.alloc, pAllocator, buffer);
|
||||
}
|
||||
|
||||
VkResult
|
||||
radv_create_buffer(struct radv_device *device, const VkBufferCreateInfo *pCreateInfo,
|
||||
const VkAllocationCallbacks *pAllocator, VkBuffer *pBuffer, bool is_internal)
|
||||
{
|
||||
struct radv_buffer *buffer;
|
||||
|
||||
assert(pCreateInfo->sType == VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO);
|
||||
|
||||
#ifdef ANDROID
|
||||
/* reject buffers that are larger than maxBufferSize on Android, which
|
||||
* might not have VK_KHR_maintenance4
|
||||
*/
|
||||
if (pCreateInfo->size > RADV_MAX_MEMORY_ALLOCATION_SIZE)
|
||||
return vk_error(device, VK_ERROR_OUT_OF_DEVICE_MEMORY);
|
||||
#endif
|
||||
|
||||
buffer = vk_alloc2(&device->vk.alloc, pAllocator, sizeof(*buffer), 8,
|
||||
VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
|
||||
if (buffer == NULL)
|
||||
return vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
|
||||
|
||||
vk_buffer_init(&device->vk, &buffer->vk, pCreateInfo);
|
||||
buffer->bo = NULL;
|
||||
buffer->offset = 0;
|
||||
|
||||
if (pCreateInfo->flags & VK_BUFFER_CREATE_SPARSE_BINDING_BIT) {
|
||||
enum radeon_bo_flag flags = RADEON_FLAG_VIRTUAL;
|
||||
if (pCreateInfo->flags & VK_BUFFER_CREATE_DEVICE_ADDRESS_CAPTURE_REPLAY_BIT)
|
||||
flags |= RADEON_FLAG_REPLAYABLE;
|
||||
|
||||
uint64_t replay_address = 0;
|
||||
const VkBufferOpaqueCaptureAddressCreateInfo *replay_info =
|
||||
vk_find_struct_const(pCreateInfo->pNext, BUFFER_OPAQUE_CAPTURE_ADDRESS_CREATE_INFO);
|
||||
if (replay_info && replay_info->opaqueCaptureAddress)
|
||||
replay_address = replay_info->opaqueCaptureAddress;
|
||||
|
||||
VkResult result =
|
||||
device->ws->buffer_create(device->ws, align64(buffer->vk.size, 4096), 4096, 0, flags,
|
||||
RADV_BO_PRIORITY_VIRTUAL, replay_address, &buffer->bo);
|
||||
if (result != VK_SUCCESS) {
|
||||
radv_destroy_buffer(device, pAllocator, buffer);
|
||||
return vk_error(device, result);
|
||||
}
|
||||
radv_rmv_log_bo_allocate(device, buffer->bo, buffer->vk.size, true);
|
||||
}
|
||||
|
||||
*pBuffer = radv_buffer_to_handle(buffer);
|
||||
vk_rmv_log_buffer_create(&device->vk, false, *pBuffer);
|
||||
if (buffer->bo)
|
||||
radv_rmv_log_buffer_bind(device, *pBuffer);
|
||||
return VK_SUCCESS;
|
||||
}
|
||||
|
||||
VKAPI_ATTR VkResult VKAPI_CALL
|
||||
radv_CreateBuffer(VkDevice _device, const VkBufferCreateInfo *pCreateInfo,
|
||||
const VkAllocationCallbacks *pAllocator, VkBuffer *pBuffer)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
return radv_create_buffer(device, pCreateInfo, pAllocator, pBuffer, false);
|
||||
}
|
||||
|
||||
VKAPI_ATTR void VKAPI_CALL
|
||||
radv_DestroyBuffer(VkDevice _device, VkBuffer _buffer, const VkAllocationCallbacks *pAllocator)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_device, device, _device);
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, _buffer);
|
||||
|
||||
if (!buffer)
|
||||
return;
|
||||
|
||||
radv_destroy_buffer(device, pAllocator, buffer);
|
||||
}
|
||||
|
||||
VKAPI_ATTR VkDeviceAddress VKAPI_CALL
|
||||
radv_GetBufferDeviceAddress(VkDevice device, const VkBufferDeviceAddressInfo *pInfo)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, pInfo->buffer);
|
||||
return radv_buffer_get_va(buffer->bo) + buffer->offset;
|
||||
}
|
||||
|
||||
VKAPI_ATTR uint64_t VKAPI_CALL
|
||||
radv_GetBufferOpaqueCaptureAddress(VkDevice device, const VkBufferDeviceAddressInfo *pInfo)
|
||||
{
|
||||
RADV_FROM_HANDLE(radv_buffer, buffer, pInfo->buffer);
|
||||
return buffer->bo ? radv_buffer_get_va(buffer->bo) + buffer->offset : 0;
|
||||
}
|
||||
|
||||
VKAPI_ATTR uint64_t VKAPI_CALL
|
||||
radv_GetDeviceMemoryOpaqueCaptureAddress(VkDevice device,
|
||||
const VkDeviceMemoryOpaqueCaptureAddressInfo *pInfo)
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue