xref: /aosp_15_r20/external/mesa3d/src/amd/vulkan/radv_buffer.c (revision 6104692788411f58d303aa86923a9ff6ecaded22)
1 /*
2  * Copyright © 2016 Red Hat.
3  * Copyright © 2016 Bas Nieuwenhuizen
4  *
5  * based in part on anv driver which is:
6  * Copyright © 2015 Intel Corporation
7  *
8  * SPDX-License-Identifier: MIT
9  */
10 
11 #include "radv_buffer.h"
12 #include "radv_device.h"
13 #include "radv_device_memory.h"
14 #include "radv_entrypoints.h"
15 #include "radv_instance.h"
16 #include "radv_physical_device.h"
17 #include "radv_rmv.h"
18 
19 #include "vk_common_entrypoints.h"
20 #include "vk_debug_utils.h"
21 #include "vk_log.h"
22 
23 void
radv_buffer_init(struct radv_buffer * buffer,struct radv_device * device,struct radeon_winsys_bo * bo,uint64_t size,uint64_t offset)24 radv_buffer_init(struct radv_buffer *buffer, struct radv_device *device, struct radeon_winsys_bo *bo, uint64_t size,
25                  uint64_t offset)
26 {
27    VkBufferCreateInfo createInfo = {
28       .sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO,
29       .size = size,
30    };
31 
32    vk_buffer_init(&device->vk, &buffer->vk, &createInfo);
33 
34    buffer->bo = bo;
35    buffer->offset = offset;
36 }
37 
38 void
radv_buffer_finish(struct radv_buffer * buffer)39 radv_buffer_finish(struct radv_buffer *buffer)
40 {
41    vk_buffer_finish(&buffer->vk);
42 }
43 
44 static void
radv_destroy_buffer(struct radv_device * device,const VkAllocationCallbacks * pAllocator,struct radv_buffer * buffer)45 radv_destroy_buffer(struct radv_device *device, const VkAllocationCallbacks *pAllocator, struct radv_buffer *buffer)
46 {
47    struct radv_physical_device *pdev = radv_device_physical(device);
48    struct radv_instance *instance = radv_physical_device_instance(pdev);
49 
50    if ((buffer->vk.create_flags & VK_BUFFER_CREATE_SPARSE_BINDING_BIT) && buffer->bo)
51       radv_bo_destroy(device, &buffer->vk.base, buffer->bo);
52 
53    if (buffer->bo_va)
54       vk_address_binding_report(&instance->vk, &buffer->vk.base, buffer->bo_va + buffer->offset, buffer->range,
55                                 VK_DEVICE_ADDRESS_BINDING_TYPE_UNBIND_EXT);
56 
57    radv_rmv_log_resource_destroy(device, (uint64_t)radv_buffer_to_handle(buffer));
58    radv_buffer_finish(buffer);
59    vk_free2(&device->vk.alloc, pAllocator, buffer);
60 }
61 
62 VkResult
radv_create_buffer(struct radv_device * device,const VkBufferCreateInfo * pCreateInfo,const VkAllocationCallbacks * pAllocator,VkBuffer * pBuffer,bool is_internal)63 radv_create_buffer(struct radv_device *device, const VkBufferCreateInfo *pCreateInfo,
64                    const VkAllocationCallbacks *pAllocator, VkBuffer *pBuffer, bool is_internal)
65 {
66    struct radv_buffer *buffer;
67 
68    assert(pCreateInfo->sType == VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO);
69 
70 #if DETECT_OS_ANDROID
71    /* reject buffers that are larger than maxBufferSize on Android, which
72     * might not have VK_KHR_maintenance4
73     */
74    if (pCreateInfo->size > RADV_MAX_MEMORY_ALLOCATION_SIZE)
75       return vk_error(device, VK_ERROR_OUT_OF_DEVICE_MEMORY);
76 #endif
77 
78    buffer = vk_alloc2(&device->vk.alloc, pAllocator, sizeof(*buffer), 8, VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
79    if (buffer == NULL)
80       return vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
81 
82    vk_buffer_init(&device->vk, &buffer->vk, pCreateInfo);
83    buffer->bo = NULL;
84    buffer->offset = 0;
85    buffer->bo_va = 0;
86    buffer->range = 0;
87 
88    if (pCreateInfo->flags & VK_BUFFER_CREATE_SPARSE_BINDING_BIT) {
89       enum radeon_bo_flag flags = RADEON_FLAG_VIRTUAL;
90       if (pCreateInfo->flags & VK_BUFFER_CREATE_DEVICE_ADDRESS_CAPTURE_REPLAY_BIT)
91          flags |= RADEON_FLAG_REPLAYABLE;
92       if (buffer->vk.usage &
93           (VK_BUFFER_USAGE_2_RESOURCE_DESCRIPTOR_BUFFER_BIT_EXT | VK_BUFFER_USAGE_2_SAMPLER_DESCRIPTOR_BUFFER_BIT_EXT))
94          flags |= RADEON_FLAG_32BIT;
95 
96       uint64_t replay_address = 0;
97       const VkBufferOpaqueCaptureAddressCreateInfo *replay_info =
98          vk_find_struct_const(pCreateInfo->pNext, BUFFER_OPAQUE_CAPTURE_ADDRESS_CREATE_INFO);
99       if (replay_info && replay_info->opaqueCaptureAddress)
100          replay_address = replay_info->opaqueCaptureAddress;
101 
102       VkResult result = radv_bo_create(device, &buffer->vk.base, align64(buffer->vk.size, 4096), 4096, 0, flags,
103                                        RADV_BO_PRIORITY_VIRTUAL, replay_address, is_internal, &buffer->bo);
104       if (result != VK_SUCCESS) {
105          radv_destroy_buffer(device, pAllocator, buffer);
106          return vk_error(device, result);
107       }
108    }
109 
110    *pBuffer = radv_buffer_to_handle(buffer);
111    vk_rmv_log_buffer_create(&device->vk, false, *pBuffer);
112    if (buffer->bo)
113       radv_rmv_log_buffer_bind(device, *pBuffer);
114    return VK_SUCCESS;
115 }
116 
117 VKAPI_ATTR VkResult VKAPI_CALL
radv_CreateBuffer(VkDevice _device,const VkBufferCreateInfo * pCreateInfo,const VkAllocationCallbacks * pAllocator,VkBuffer * pBuffer)118 radv_CreateBuffer(VkDevice _device, const VkBufferCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator,
119                   VkBuffer *pBuffer)
120 {
121    VK_FROM_HANDLE(radv_device, device, _device);
122    return radv_create_buffer(device, pCreateInfo, pAllocator, pBuffer, false);
123 }
124 
125 VKAPI_ATTR void VKAPI_CALL
radv_DestroyBuffer(VkDevice _device,VkBuffer _buffer,const VkAllocationCallbacks * pAllocator)126 radv_DestroyBuffer(VkDevice _device, VkBuffer _buffer, const VkAllocationCallbacks *pAllocator)
127 {
128    VK_FROM_HANDLE(radv_device, device, _device);
129    VK_FROM_HANDLE(radv_buffer, buffer, _buffer);
130 
131    if (!buffer)
132       return;
133 
134    radv_destroy_buffer(device, pAllocator, buffer);
135 }
136 
137 VKAPI_ATTR VkResult VKAPI_CALL
radv_BindBufferMemory2(VkDevice _device,uint32_t bindInfoCount,const VkBindBufferMemoryInfo * pBindInfos)138 radv_BindBufferMemory2(VkDevice _device, uint32_t bindInfoCount, const VkBindBufferMemoryInfo *pBindInfos)
139 {
140    VK_FROM_HANDLE(radv_device, device, _device);
141    struct radv_physical_device *pdev = radv_device_physical(device);
142    struct radv_instance *instance = radv_physical_device_instance(pdev);
143 
144    for (uint32_t i = 0; i < bindInfoCount; ++i) {
145       VK_FROM_HANDLE(radv_device_memory, mem, pBindInfos[i].memory);
146       VK_FROM_HANDLE(radv_buffer, buffer, pBindInfos[i].buffer);
147       VkBindMemoryStatusKHR *status = (void *)vk_find_struct_const(&pBindInfos[i], BIND_MEMORY_STATUS_KHR);
148 
149       if (status)
150          *status->pResult = VK_SUCCESS;
151 
152       VkBufferMemoryRequirementsInfo2 info = {
153          .sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_REQUIREMENTS_INFO_2,
154          .buffer = pBindInfos[i].buffer,
155       };
156       VkMemoryRequirements2 reqs = {
157          .sType = VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2,
158       };
159 
160       vk_common_GetBufferMemoryRequirements2(_device, &info, &reqs);
161 
162       if (mem->alloc_size) {
163          if (pBindInfos[i].memoryOffset + reqs.memoryRequirements.size > mem->alloc_size) {
164             if (status)
165                *status->pResult = VK_ERROR_UNKNOWN;
166             return vk_errorf(device, VK_ERROR_UNKNOWN, "Device memory object too small for the buffer.\n");
167          }
168       }
169 
170       buffer->bo = mem->bo;
171       buffer->offset = pBindInfos[i].memoryOffset;
172       buffer->bo_va = radv_buffer_get_va(mem->bo);
173       buffer->range = reqs.memoryRequirements.size;
174 
175       radv_rmv_log_buffer_bind(device, pBindInfos[i].buffer);
176 
177       vk_address_binding_report(&instance->vk, &buffer->vk.base, radv_buffer_get_va(buffer->bo) + buffer->offset,
178                                 buffer->range, VK_DEVICE_ADDRESS_BINDING_TYPE_BIND_EXT);
179    }
180    return VK_SUCCESS;
181 }
182 
183 static void
radv_get_buffer_memory_requirements(struct radv_device * device,VkDeviceSize size,VkBufferCreateFlags flags,VkBufferUsageFlags2KHR usage,VkMemoryRequirements2 * pMemoryRequirements)184 radv_get_buffer_memory_requirements(struct radv_device *device, VkDeviceSize size, VkBufferCreateFlags flags,
185                                     VkBufferUsageFlags2KHR usage, VkMemoryRequirements2 *pMemoryRequirements)
186 {
187    const struct radv_physical_device *pdev = radv_device_physical(device);
188 
189    pMemoryRequirements->memoryRequirements.memoryTypeBits =
190       ((1u << pdev->memory_properties.memoryTypeCount) - 1u) & ~pdev->memory_types_32bit;
191 
192    /* Allow 32-bit address-space for DGC usage, as this buffer will contain
193     * cmd buffer upload buffers, and those get passed to shaders through 32-bit
194     * pointers.
195     *
196     * We only allow it with this usage set, to "protect" the 32-bit address space
197     * from being overused. The actual requirement is done as part of
198     * vkGetGeneratedCommandsMemoryRequirementsNV. (we have to make sure their
199     * intersection is non-zero at least)
200     */
201    if ((usage & VK_BUFFER_USAGE_2_INDIRECT_BUFFER_BIT_KHR) && radv_uses_device_generated_commands(device))
202       pMemoryRequirements->memoryRequirements.memoryTypeBits |= pdev->memory_types_32bit;
203 
204    /* Force 32-bit address-space for descriptor buffers usage because they are passed to shaders
205     * through 32-bit pointers.
206     */
207    if (usage &
208        (VK_BUFFER_USAGE_2_RESOURCE_DESCRIPTOR_BUFFER_BIT_EXT | VK_BUFFER_USAGE_2_SAMPLER_DESCRIPTOR_BUFFER_BIT_EXT))
209       pMemoryRequirements->memoryRequirements.memoryTypeBits = pdev->memory_types_32bit;
210 
211    if (flags & VK_BUFFER_CREATE_SPARSE_BINDING_BIT)
212       pMemoryRequirements->memoryRequirements.alignment = 4096;
213    else
214       pMemoryRequirements->memoryRequirements.alignment = 16;
215 
216    /* Top level acceleration structures need the bottom 6 bits to store
217     * the root ids of instances. The hardware also needs bvh nodes to
218     * be 64 byte aligned.
219     */
220    if (usage & VK_BUFFER_USAGE_2_ACCELERATION_STRUCTURE_STORAGE_BIT_KHR)
221       pMemoryRequirements->memoryRequirements.alignment = MAX2(pMemoryRequirements->memoryRequirements.alignment, 64);
222 
223    pMemoryRequirements->memoryRequirements.size = align64(size, pMemoryRequirements->memoryRequirements.alignment);
224 
225    vk_foreach_struct (ext, pMemoryRequirements->pNext) {
226       switch (ext->sType) {
227       case VK_STRUCTURE_TYPE_MEMORY_DEDICATED_REQUIREMENTS: {
228          VkMemoryDedicatedRequirements *req = (VkMemoryDedicatedRequirements *)ext;
229          req->requiresDedicatedAllocation = false;
230          req->prefersDedicatedAllocation = req->requiresDedicatedAllocation;
231          break;
232       }
233       default:
234          break;
235       }
236    }
237 }
238 
239 static const VkBufferUsageFlagBits2KHR
radv_get_buffer_usage_flags(const VkBufferCreateInfo * pCreateInfo)240 radv_get_buffer_usage_flags(const VkBufferCreateInfo *pCreateInfo)
241 {
242    const VkBufferUsageFlags2CreateInfoKHR *flags2 =
243       vk_find_struct_const(pCreateInfo->pNext, BUFFER_USAGE_FLAGS_2_CREATE_INFO_KHR);
244    return flags2 ? flags2->usage : pCreateInfo->usage;
245 }
246 
247 VKAPI_ATTR void VKAPI_CALL
radv_GetDeviceBufferMemoryRequirements(VkDevice _device,const VkDeviceBufferMemoryRequirements * pInfo,VkMemoryRequirements2 * pMemoryRequirements)248 radv_GetDeviceBufferMemoryRequirements(VkDevice _device, const VkDeviceBufferMemoryRequirements *pInfo,
249                                        VkMemoryRequirements2 *pMemoryRequirements)
250 {
251    VK_FROM_HANDLE(radv_device, device, _device);
252    const VkBufferUsageFlagBits2KHR usage_flags = radv_get_buffer_usage_flags(pInfo->pCreateInfo);
253 
254    radv_get_buffer_memory_requirements(device, pInfo->pCreateInfo->size, pInfo->pCreateInfo->flags, usage_flags,
255                                        pMemoryRequirements);
256 }
257 
258 VKAPI_ATTR VkDeviceAddress VKAPI_CALL
radv_GetBufferDeviceAddress(VkDevice device,const VkBufferDeviceAddressInfo * pInfo)259 radv_GetBufferDeviceAddress(VkDevice device, const VkBufferDeviceAddressInfo *pInfo)
260 {
261    VK_FROM_HANDLE(radv_buffer, buffer, pInfo->buffer);
262    return radv_buffer_get_va(buffer->bo) + buffer->offset;
263 }
264 
265 VKAPI_ATTR uint64_t VKAPI_CALL
radv_GetBufferOpaqueCaptureAddress(VkDevice device,const VkBufferDeviceAddressInfo * pInfo)266 radv_GetBufferOpaqueCaptureAddress(VkDevice device, const VkBufferDeviceAddressInfo *pInfo)
267 {
268    VK_FROM_HANDLE(radv_buffer, buffer, pInfo->buffer);
269    return buffer->bo ? radv_buffer_get_va(buffer->bo) + buffer->offset : 0;
270 }
271 
272 VkResult
radv_bo_create(struct radv_device * device,struct vk_object_base * object,uint64_t size,unsigned alignment,enum radeon_bo_domain domain,enum radeon_bo_flag flags,unsigned priority,uint64_t address,bool is_internal,struct radeon_winsys_bo ** out_bo)273 radv_bo_create(struct radv_device *device, struct vk_object_base *object, uint64_t size, unsigned alignment,
274                enum radeon_bo_domain domain, enum radeon_bo_flag flags, unsigned priority, uint64_t address,
275                bool is_internal, struct radeon_winsys_bo **out_bo)
276 {
277    struct radv_physical_device *pdev = radv_device_physical(device);
278    struct radv_instance *instance = radv_physical_device_instance(pdev);
279    struct radeon_winsys *ws = device->ws;
280    VkResult result;
281 
282    result = ws->buffer_create(ws, size, alignment, domain, flags, priority, address, out_bo);
283    if (result != VK_SUCCESS)
284       return result;
285 
286    radv_rmv_log_bo_allocate(device, *out_bo, is_internal);
287 
288    vk_address_binding_report(&instance->vk, object ? object : &device->vk.base, radv_buffer_get_va(*out_bo),
289                              (*out_bo)->size, VK_DEVICE_ADDRESS_BINDING_TYPE_BIND_EXT);
290    return VK_SUCCESS;
291 }
292 
293 void
radv_bo_destroy(struct radv_device * device,struct vk_object_base * object,struct radeon_winsys_bo * bo)294 radv_bo_destroy(struct radv_device *device, struct vk_object_base *object, struct radeon_winsys_bo *bo)
295 {
296    struct radv_physical_device *pdev = radv_device_physical(device);
297    struct radv_instance *instance = radv_physical_device_instance(pdev);
298    struct radeon_winsys *ws = device->ws;
299 
300    radv_rmv_log_bo_destroy(device, bo);
301 
302    vk_address_binding_report(&instance->vk, object ? object : &device->vk.base, radv_buffer_get_va(bo), bo->size,
303                              VK_DEVICE_ADDRESS_BINDING_TYPE_UNBIND_EXT);
304 
305    ws->buffer_destroy(ws, bo);
306 }
307 
308 VkResult
radv_bo_virtual_bind(struct radv_device * device,struct vk_object_base * object,struct radeon_winsys_bo * parent,uint64_t offset,uint64_t size,struct radeon_winsys_bo * bo,uint64_t bo_offset)309 radv_bo_virtual_bind(struct radv_device *device, struct vk_object_base *object, struct radeon_winsys_bo *parent,
310                      uint64_t offset, uint64_t size, struct radeon_winsys_bo *bo, uint64_t bo_offset)
311 {
312    struct radv_physical_device *pdev = radv_device_physical(device);
313    struct radv_instance *instance = radv_physical_device_instance(pdev);
314    struct radeon_winsys *ws = device->ws;
315    VkResult result;
316 
317    result = ws->buffer_virtual_bind(ws, parent, offset, size, bo, bo_offset);
318    if (result != VK_SUCCESS)
319       return result;
320 
321    if (bo)
322       radv_rmv_log_sparse_add_residency(device, parent, offset);
323    else
324       radv_rmv_log_sparse_remove_residency(device, parent, offset);
325 
326    vk_address_binding_report(&instance->vk, object, radv_buffer_get_va(parent) + offset, size,
327                              bo ? VK_DEVICE_ADDRESS_BINDING_TYPE_BIND_EXT : VK_DEVICE_ADDRESS_BINDING_TYPE_UNBIND_EXT);
328 
329    return VK_SUCCESS;
330 }
331