vkr: add vkr_device_alloc_queue_sync

Add vkr_device_{alloc,free}_queue_sync.  No real change.

Signed-off-by: Chia-I Wu <olvaffe@gmail.com>
Reviewed-by: Yiwei Zhang <zzyiwei@chromium.org>
Reviewed-by: Louis-Francis Ratté-Boulianne <lfrb@collabora.com>
macos/master
Chia-I Wu 3 years ago
parent 457fe55ebb
commit fe9d9282ad
  1. 82
      src/vkr_renderer.c

@ -1462,6 +1462,50 @@ vkr_dispatch_vkGetPhysicalDeviceExternalFenceProperties(
args->physicalDevice, args->pExternalFenceInfo, args->pExternalFenceProperties); args->physicalDevice, args->pExternalFenceInfo, args->pExternalFenceProperties);
} }
static struct vkr_queue_sync *
vkr_device_alloc_queue_sync(struct vkr_device *dev,
uint32_t fence_flags,
void *fence_cookie)
{
struct vkr_queue_sync *sync;
if (LIST_IS_EMPTY(&dev->free_syncs)) {
sync = malloc(sizeof(*sync));
if (!sync)
return NULL;
const VkExportFenceCreateInfo export_info = {
.sType = VK_STRUCTURE_TYPE_EXPORT_FENCE_CREATE_INFO,
.handleTypes = VK_EXTERNAL_FENCE_HANDLE_TYPE_SYNC_FD_BIT,
};
const struct VkFenceCreateInfo create_info = {
.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO,
.pNext = dev->physical_device->KHR_external_fence_fd ? &export_info : NULL,
};
VkResult result =
vkCreateFence(dev->base.handle.device, &create_info, NULL, &sync->fence);
if (result != VK_SUCCESS) {
free(sync);
return NULL;
}
} else {
sync = LIST_ENTRY(struct vkr_queue_sync, dev->free_syncs.next, head);
list_del(&sync->head);
vkResetFences(dev->base.handle.device, 1, &sync->fence);
}
sync->flags = fence_flags;
sync->fence_cookie = fence_cookie;
return sync;
}
static void
vkr_device_free_queue_sync(struct vkr_device *dev, struct vkr_queue_sync *sync)
{
list_addtail(&sync->head, &dev->free_syncs);
}
static void static void
vkr_queue_retire_syncs(struct vkr_queue *queue, vkr_queue_retire_syncs(struct vkr_queue *queue,
struct list_head *retired_syncs, struct list_head *retired_syncs,
@ -1480,7 +1524,7 @@ vkr_queue_retire_syncs(struct vkr_queue *queue,
!(sync->flags & VIRGL_RENDERER_FENCE_FLAG_MERGEABLE)) !(sync->flags & VIRGL_RENDERER_FENCE_FLAG_MERGEABLE))
list_addtail(&sync->head, retired_syncs); list_addtail(&sync->head, retired_syncs);
else else
list_addtail(&sync->head, &dev->free_syncs); vkr_device_free_queue_sync(dev, sync);
} }
list_inithead(&queue->signaled_syncs); list_inithead(&queue->signaled_syncs);
@ -1498,7 +1542,7 @@ vkr_queue_retire_syncs(struct vkr_queue *queue,
!(sync->flags & VIRGL_RENDERER_FENCE_FLAG_MERGEABLE)) !(sync->flags & VIRGL_RENDERER_FENCE_FLAG_MERGEABLE))
list_addtail(&sync->head, retired_syncs); list_addtail(&sync->head, retired_syncs);
else else
list_addtail(&sync->head, &dev->free_syncs); vkr_device_free_queue_sync(dev, sync);
} }
*queue_empty = LIST_IS_EMPTY(&queue->pending_syncs); *queue_empty = LIST_IS_EMPTY(&queue->pending_syncs);
@ -1560,13 +1604,13 @@ vkr_queue_destroy(struct vkr_context *ctx, struct vkr_queue *queue)
thrd_join(queue->thread, NULL); thrd_join(queue->thread, NULL);
LIST_FOR_EACH_ENTRY_SAFE (sync, tmp, &queue->signaled_syncs, head) LIST_FOR_EACH_ENTRY_SAFE (sync, tmp, &queue->signaled_syncs, head)
list_addtail(&sync->head, &queue->device->free_syncs); vkr_device_free_queue_sync(queue->device, sync);
} else { } else {
assert(LIST_IS_EMPTY(&queue->signaled_syncs)); assert(LIST_IS_EMPTY(&queue->signaled_syncs));
} }
LIST_FOR_EACH_ENTRY_SAFE (sync, tmp, &queue->pending_syncs, head) LIST_FOR_EACH_ENTRY_SAFE (sync, tmp, &queue->pending_syncs, head)
list_addtail(&sync->head, &queue->device->free_syncs); vkr_device_free_queue_sync(queue->device, sync);
mtx_destroy(&queue->mutex); mtx_destroy(&queue->mutex);
cnd_destroy(&queue->cond); cnd_destroy(&queue->cond);
@ -4258,40 +4302,16 @@ vkr_context_submit_fence_locked(struct virgl_context *base,
return -EINVAL; return -EINVAL;
struct vkr_device *dev = queue->device; struct vkr_device *dev = queue->device;
struct vkr_queue_sync *sync; struct vkr_queue_sync *sync = vkr_device_alloc_queue_sync(dev, flags, fence_cookie);
if (LIST_IS_EMPTY(&dev->free_syncs)) {
sync = malloc(sizeof(*sync));
if (!sync) if (!sync)
return -ENOMEM; return -ENOMEM;
const VkExportFenceCreateInfo export_info = {
.sType = VK_STRUCTURE_TYPE_EXPORT_FENCE_CREATE_INFO,
.handleTypes = VK_EXTERNAL_FENCE_HANDLE_TYPE_SYNC_FD_BIT,
};
const struct VkFenceCreateInfo create_info = {
.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO,
.pNext = dev->physical_device->KHR_external_fence_fd ? &export_info : NULL,
};
result = vkCreateFence(dev->base.handle.device, &create_info, NULL, &sync->fence);
if (result != VK_SUCCESS) {
free(sync);
return -ENOMEM;
}
} else {
sync = LIST_ENTRY(struct vkr_queue_sync, dev->free_syncs.next, head);
list_del(&sync->head);
vkResetFences(dev->base.handle.device, 1, &sync->fence);
}
result = vkQueueSubmit(queue->base.handle.queue, 0, NULL, sync->fence); result = vkQueueSubmit(queue->base.handle.queue, 0, NULL, sync->fence);
if (result != VK_SUCCESS) { if (result != VK_SUCCESS) {
list_add(&sync->head, &dev->free_syncs); vkr_device_free_queue_sync(dev, sync);
return -1; return -1;
} }
sync->flags = flags;
sync->fence_cookie = fence_cookie;
if (vkr_renderer_flags & VKR_RENDERER_THREAD_SYNC) { if (vkr_renderer_flags & VKR_RENDERER_THREAD_SYNC) {
mtx_lock(&queue->mutex); mtx_lock(&queue->mutex);
list_addtail(&sync->head, &queue->pending_syncs); list_addtail(&sync->head, &queue->pending_syncs);
@ -4344,7 +4364,7 @@ vkr_context_retire_fences_locked(UNUSED struct virgl_context *base)
LIST_FOR_EACH_ENTRY_SAFE (sync, sync_tmp, &retired_syncs, head) { LIST_FOR_EACH_ENTRY_SAFE (sync, sync_tmp, &retired_syncs, head) {
ctx->base.fence_retire(&ctx->base, queue->base.id, sync->fence_cookie); ctx->base.fence_retire(&ctx->base, queue->base.id, sync->fence_cookie);
list_addtail(&sync->head, &dev->free_syncs); vkr_device_free_queue_sync(dev, sync);
} }
if (queue_empty) if (queue_empty)

Loading…
Cancel
Save