No functional change. Signed-off-by: Chia-I Wu <olvaffe@gmail.com> Reviewed-by: Yiwei Zhang <zzyiwei@chromium.org> Reviewed-by: Ryan Neph <ryanneph@google.com>macos/master
parent
f29550d86c
commit
813607e5bd
@ -0,0 +1,567 @@ |
||||
/*
|
||||
* Copyright 2020 Google LLC |
||||
* SPDX-License-Identifier: MIT |
||||
*/ |
||||
|
||||
#include "vkr_queue.h" |
||||
|
||||
#include "venus-protocol/vn_protocol_renderer_device.h" |
||||
#include "venus-protocol/vn_protocol_renderer_event.h" |
||||
#include "venus-protocol/vn_protocol_renderer_fence.h" |
||||
#include "venus-protocol/vn_protocol_renderer_queue.h" |
||||
#include "venus-protocol/vn_protocol_renderer_semaphore.h" |
||||
|
||||
#include "vkr_context.h" |
||||
#include "vkr_device.h" |
||||
|
||||
struct vkr_queue_sync * |
||||
vkr_device_alloc_queue_sync(struct vkr_device *dev, |
||||
uint32_t fence_flags, |
||||
void *fence_cookie) |
||||
{ |
||||
struct vkr_queue_sync *sync; |
||||
|
||||
if (vkr_renderer_flags & VKR_RENDERER_ASYNC_FENCE_CB) |
||||
mtx_lock(&dev->free_sync_mutex); |
||||
|
||||
if (LIST_IS_EMPTY(&dev->free_syncs)) { |
||||
if (vkr_renderer_flags & VKR_RENDERER_ASYNC_FENCE_CB) |
||||
mtx_unlock(&dev->free_sync_mutex); |
||||
|
||||
sync = malloc(sizeof(*sync)); |
||||
if (!sync) |
||||
return NULL; |
||||
|
||||
const VkExportFenceCreateInfo export_info = { |
||||
.sType = VK_STRUCTURE_TYPE_EXPORT_FENCE_CREATE_INFO, |
||||
.handleTypes = VK_EXTERNAL_FENCE_HANDLE_TYPE_SYNC_FD_BIT, |
||||
}; |
||||
const struct VkFenceCreateInfo create_info = { |
||||
.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO, |
||||
.pNext = dev->physical_device->KHR_external_fence_fd ? &export_info : NULL, |
||||
}; |
||||
VkResult result = |
||||
vkCreateFence(dev->base.handle.device, &create_info, NULL, &sync->fence); |
||||
if (result != VK_SUCCESS) { |
||||
free(sync); |
||||
return NULL; |
||||
} |
||||
} else { |
||||
sync = LIST_ENTRY(struct vkr_queue_sync, dev->free_syncs.next, head); |
||||
list_del(&sync->head); |
||||
|
||||
if (vkr_renderer_flags & VKR_RENDERER_ASYNC_FENCE_CB) |
||||
mtx_unlock(&dev->free_sync_mutex); |
||||
|
||||
vkResetFences(dev->base.handle.device, 1, &sync->fence); |
||||
} |
||||
|
||||
sync->flags = fence_flags; |
||||
sync->fence_cookie = fence_cookie; |
||||
|
||||
return sync; |
||||
} |
||||
|
||||
void |
||||
vkr_device_free_queue_sync(struct vkr_device *dev, struct vkr_queue_sync *sync) |
||||
{ |
||||
if (vkr_renderer_flags & VKR_RENDERER_ASYNC_FENCE_CB) { |
||||
mtx_lock(&dev->free_sync_mutex); |
||||
list_addtail(&sync->head, &dev->free_syncs); |
||||
mtx_unlock(&dev->free_sync_mutex); |
||||
} else { |
||||
list_addtail(&sync->head, &dev->free_syncs); |
||||
} |
||||
} |
||||
|
||||
void |
||||
vkr_queue_retire_syncs(struct vkr_queue *queue, |
||||
struct list_head *retired_syncs, |
||||
bool *queue_empty) |
||||
{ |
||||
struct vkr_device *dev = queue->device; |
||||
struct vkr_queue_sync *sync, *tmp; |
||||
|
||||
assert(!(vkr_renderer_flags & VKR_RENDERER_ASYNC_FENCE_CB)); |
||||
|
||||
list_inithead(retired_syncs); |
||||
|
||||
if (vkr_renderer_flags & VKR_RENDERER_THREAD_SYNC) { |
||||
mtx_lock(&queue->mutex); |
||||
|
||||
LIST_FOR_EACH_ENTRY_SAFE (sync, tmp, &queue->signaled_syncs, head) { |
||||
if (sync->head.next == &queue->signaled_syncs || |
||||
!(sync->flags & VIRGL_RENDERER_FENCE_FLAG_MERGEABLE)) |
||||
list_addtail(&sync->head, retired_syncs); |
||||
else |
||||
vkr_device_free_queue_sync(dev, sync); |
||||
} |
||||
list_inithead(&queue->signaled_syncs); |
||||
|
||||
*queue_empty = LIST_IS_EMPTY(&queue->pending_syncs); |
||||
|
||||
mtx_unlock(&queue->mutex); |
||||
} else { |
||||
LIST_FOR_EACH_ENTRY_SAFE (sync, tmp, &queue->pending_syncs, head) { |
||||
VkResult result = vkGetFenceStatus(dev->base.handle.device, sync->fence); |
||||
if (result == VK_NOT_READY) |
||||
break; |
||||
|
||||
list_del(&sync->head); |
||||
if (sync->head.next == &queue->pending_syncs || |
||||
!(sync->flags & VIRGL_RENDERER_FENCE_FLAG_MERGEABLE)) |
||||
list_addtail(&sync->head, retired_syncs); |
||||
else |
||||
vkr_device_free_queue_sync(dev, sync); |
||||
} |
||||
|
||||
*queue_empty = LIST_IS_EMPTY(&queue->pending_syncs); |
||||
} |
||||
} |
||||
|
||||
void |
||||
vkr_queue_destroy(struct vkr_context *ctx, struct vkr_queue *queue) |
||||
{ |
||||
struct vkr_queue_sync *sync, *tmp; |
||||
|
||||
if (vkr_renderer_flags & VKR_RENDERER_THREAD_SYNC) { |
||||
mtx_lock(&queue->mutex); |
||||
queue->join = true; |
||||
mtx_unlock(&queue->mutex); |
||||
|
||||
cnd_signal(&queue->cond); |
||||
thrd_join(queue->thread, NULL); |
||||
|
||||
LIST_FOR_EACH_ENTRY_SAFE (sync, tmp, &queue->signaled_syncs, head) |
||||
vkr_device_free_queue_sync(queue->device, sync); |
||||
} else { |
||||
assert(LIST_IS_EMPTY(&queue->signaled_syncs)); |
||||
} |
||||
|
||||
LIST_FOR_EACH_ENTRY_SAFE (sync, tmp, &queue->pending_syncs, head) |
||||
vkr_device_free_queue_sync(queue->device, sync); |
||||
|
||||
mtx_destroy(&queue->mutex); |
||||
cnd_destroy(&queue->cond); |
||||
|
||||
list_del(&queue->head); |
||||
list_del(&queue->busy_head); |
||||
|
||||
util_hash_table_remove_u64(ctx->object_table, queue->base.id); |
||||
} |
||||
|
||||
static int |
||||
vkr_queue_thread(void *arg) |
||||
{ |
||||
struct vkr_queue *queue = arg; |
||||
struct vkr_context *ctx = queue->context; |
||||
struct vkr_device *dev = queue->device; |
||||
const uint64_t ns_per_sec = 1000000000llu; |
||||
char thread_name[16]; |
||||
|
||||
snprintf(thread_name, ARRAY_SIZE(thread_name), "vkr-queue-%d", ctx->base.ctx_id); |
||||
pipe_thread_setname(thread_name); |
||||
|
||||
mtx_lock(&queue->mutex); |
||||
while (true) { |
||||
while (LIST_IS_EMPTY(&queue->pending_syncs) && !queue->join) |
||||
cnd_wait(&queue->cond, &queue->mutex); |
||||
|
||||
if (queue->join) |
||||
break; |
||||
|
||||
struct vkr_queue_sync *sync = |
||||
LIST_ENTRY(struct vkr_queue_sync, queue->pending_syncs.next, head); |
||||
|
||||
mtx_unlock(&queue->mutex); |
||||
|
||||
VkResult result = |
||||
vkWaitForFences(dev->base.handle.device, 1, &sync->fence, false, ns_per_sec * 3); |
||||
|
||||
mtx_lock(&queue->mutex); |
||||
|
||||
if (result == VK_TIMEOUT) |
||||
continue; |
||||
|
||||
list_del(&sync->head); |
||||
|
||||
if (vkr_renderer_flags & VKR_RENDERER_ASYNC_FENCE_CB) { |
||||
ctx->base.fence_retire(&ctx->base, queue->base.id, sync->fence_cookie); |
||||
vkr_device_free_queue_sync(queue->device, sync); |
||||
} else { |
||||
list_addtail(&sync->head, &queue->signaled_syncs); |
||||
write_eventfd(queue->eventfd, 1); |
||||
} |
||||
} |
||||
mtx_unlock(&queue->mutex); |
||||
|
||||
return 0; |
||||
} |
||||
|
||||
static struct vkr_queue * |
||||
vkr_queue_create(struct vkr_context *ctx, |
||||
struct vkr_device *dev, |
||||
vkr_object_id id, |
||||
VkQueue handle, |
||||
uint32_t family, |
||||
uint32_t index) |
||||
{ |
||||
struct vkr_queue *queue; |
||||
int ret; |
||||
|
||||
LIST_FOR_EACH_ENTRY (queue, &dev->queues, head) { |
||||
if (queue->family == family && queue->index == index) |
||||
return queue; |
||||
} |
||||
|
||||
queue = calloc(1, sizeof(*queue)); |
||||
if (!queue) |
||||
return NULL; |
||||
|
||||
queue->base.type = VK_OBJECT_TYPE_QUEUE; |
||||
queue->base.id = id; |
||||
queue->base.handle.queue = handle; |
||||
|
||||
queue->context = ctx; |
||||
queue->device = dev; |
||||
queue->family = family; |
||||
queue->index = index; |
||||
|
||||
list_inithead(&queue->pending_syncs); |
||||
list_inithead(&queue->signaled_syncs); |
||||
|
||||
ret = mtx_init(&queue->mutex, mtx_plain); |
||||
if (ret != thrd_success) { |
||||
free(queue); |
||||
return NULL; |
||||
} |
||||
ret = cnd_init(&queue->cond); |
||||
if (ret != thrd_success) { |
||||
mtx_destroy(&queue->mutex); |
||||
free(queue); |
||||
return NULL; |
||||
} |
||||
|
||||
if (vkr_renderer_flags & VKR_RENDERER_THREAD_SYNC) { |
||||
ret = thrd_create(&queue->thread, vkr_queue_thread, queue); |
||||
if (ret != thrd_success) { |
||||
mtx_destroy(&queue->mutex); |
||||
cnd_destroy(&queue->cond); |
||||
free(queue); |
||||
return NULL; |
||||
} |
||||
queue->eventfd = ctx->fence_eventfd; |
||||
} |
||||
|
||||
/* currently queues are not tracked as device objects */ |
||||
list_addtail(&queue->head, &dev->queues); |
||||
list_inithead(&queue->busy_head); |
||||
|
||||
util_hash_table_set_u64(ctx->object_table, queue->base.id, queue); |
||||
|
||||
return queue; |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkGetDeviceQueue(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkGetDeviceQueue *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
|
||||
struct vkr_device *dev = (struct vkr_device *)args->device; |
||||
if (!dev || dev->base.type != VK_OBJECT_TYPE_DEVICE) { |
||||
vkr_cs_decoder_set_fatal(&ctx->decoder); |
||||
return; |
||||
} |
||||
|
||||
const vkr_object_id id = |
||||
vkr_cs_handle_load_id((const void **)args->pQueue, VK_OBJECT_TYPE_QUEUE); |
||||
|
||||
VkQueue handle; |
||||
vn_replace_vkGetDeviceQueue_args_handle(args); |
||||
vkGetDeviceQueue(args->device, args->queueFamilyIndex, args->queueIndex, &handle); |
||||
|
||||
struct vkr_queue *queue = |
||||
vkr_queue_create(ctx, dev, id, handle, args->queueFamilyIndex, args->queueIndex); |
||||
/* TODO create queues with device and deal with failures there */ |
||||
if (!queue) |
||||
vrend_printf("failed to create queue\n"); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkGetDeviceQueue2(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkGetDeviceQueue2 *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
|
||||
struct vkr_device *dev = (struct vkr_device *)args->device; |
||||
if (!dev || dev->base.type != VK_OBJECT_TYPE_DEVICE) { |
||||
vkr_cs_decoder_set_fatal(&ctx->decoder); |
||||
return; |
||||
} |
||||
|
||||
const vkr_object_id id = |
||||
vkr_cs_handle_load_id((const void **)args->pQueue, VK_OBJECT_TYPE_QUEUE); |
||||
|
||||
VkQueue handle; |
||||
vn_replace_vkGetDeviceQueue2_args_handle(args); |
||||
vkGetDeviceQueue2(args->device, args->pQueueInfo, &handle); |
||||
|
||||
/* TODO deal with errors */ |
||||
vkr_queue_create(ctx, dev, id, handle, args->pQueueInfo->queueFamilyIndex, |
||||
args->pQueueInfo->queueIndex); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkQueueSubmit(UNUSED struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkQueueSubmit *args) |
||||
{ |
||||
vn_replace_vkQueueSubmit_args_handle(args); |
||||
args->ret = vkQueueSubmit(args->queue, args->submitCount, args->pSubmits, args->fence); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkQueueBindSparse(UNUSED struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkQueueBindSparse *args) |
||||
{ |
||||
vn_replace_vkQueueBindSparse_args_handle(args); |
||||
args->ret = |
||||
vkQueueBindSparse(args->queue, args->bindInfoCount, args->pBindInfo, args->fence); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkQueueWaitIdle(struct vn_dispatch_context *dispatch, |
||||
UNUSED struct vn_command_vkQueueWaitIdle *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
/* no blocking call */ |
||||
vkr_cs_decoder_set_fatal(&ctx->decoder); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkCreateFence(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkCreateFence *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
|
||||
CREATE_OBJECT(fence, fence, FENCE, vkCreateFence, pFence); |
||||
|
||||
util_hash_table_set_u64(ctx->object_table, fence->base.id, fence); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkDestroyFence(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkDestroyFence *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
|
||||
DESTROY_OBJECT(fence, fence, FENCE, vkDestroyFence, fence); |
||||
|
||||
util_hash_table_remove_u64(ctx->object_table, fence->base.id); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkResetFences(UNUSED struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkResetFences *args) |
||||
{ |
||||
vn_replace_vkResetFences_args_handle(args); |
||||
args->ret = vkResetFences(args->device, args->fenceCount, args->pFences); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkGetFenceStatus(UNUSED struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkGetFenceStatus *args) |
||||
{ |
||||
vn_replace_vkGetFenceStatus_args_handle(args); |
||||
args->ret = vkGetFenceStatus(args->device, args->fence); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkWaitForFences(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkWaitForFences *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
|
||||
/* Being single-threaded, we cannot afford potential blocking calls. It
|
||||
* also leads to GPU lost when the wait never returns and can only be |
||||
* unblocked by a following command (e.g., vkCmdWaitEvents that is |
||||
* unblocked by a following vkSetEvent). |
||||
*/ |
||||
if (args->timeout) { |
||||
vkr_cs_decoder_set_fatal(&ctx->decoder); |
||||
return; |
||||
} |
||||
|
||||
vn_replace_vkWaitForFences_args_handle(args); |
||||
args->ret = vkWaitForFences(args->device, args->fenceCount, args->pFences, |
||||
args->waitAll, args->timeout); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkCreateSemaphore(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkCreateSemaphore *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
|
||||
CREATE_OBJECT(sem, semaphore, SEMAPHORE, vkCreateSemaphore, pSemaphore); |
||||
|
||||
util_hash_table_set_u64(ctx->object_table, sem->base.id, sem); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkDestroySemaphore(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkDestroySemaphore *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
|
||||
DESTROY_OBJECT(sem, semaphore, SEMAPHORE, vkDestroySemaphore, semaphore); |
||||
|
||||
util_hash_table_remove_u64(ctx->object_table, sem->base.id); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkGetSemaphoreCounterValue(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkGetSemaphoreCounterValue *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
struct vkr_device *dev = (struct vkr_device *)args->device; |
||||
if (!dev || dev->base.type != VK_OBJECT_TYPE_DEVICE) { |
||||
vkr_cs_decoder_set_fatal(&ctx->decoder); |
||||
return; |
||||
} |
||||
|
||||
vn_replace_vkGetSemaphoreCounterValue_args_handle(args); |
||||
args->ret = dev->GetSemaphoreCounterValue(args->device, args->semaphore, args->pValue); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkWaitSemaphores(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkWaitSemaphores *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
struct vkr_device *dev = (struct vkr_device *)args->device; |
||||
|
||||
if (!dev || dev->base.type != VK_OBJECT_TYPE_DEVICE) { |
||||
vkr_cs_decoder_set_fatal(&ctx->decoder); |
||||
return; |
||||
} |
||||
|
||||
/* no blocking call */ |
||||
if (args->timeout) { |
||||
vkr_cs_decoder_set_fatal(&ctx->decoder); |
||||
return; |
||||
} |
||||
|
||||
vn_replace_vkWaitSemaphores_args_handle(args); |
||||
args->ret = dev->WaitSemaphores(args->device, args->pWaitInfo, args->timeout); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkSignalSemaphore(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkSignalSemaphore *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
struct vkr_device *dev = (struct vkr_device *)args->device; |
||||
if (!dev || dev->base.type != VK_OBJECT_TYPE_DEVICE) { |
||||
vkr_cs_decoder_set_fatal(&ctx->decoder); |
||||
return; |
||||
} |
||||
|
||||
vn_replace_vkSignalSemaphore_args_handle(args); |
||||
args->ret = dev->SignalSemaphore(args->device, args->pSignalInfo); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkCreateEvent(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkCreateEvent *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
|
||||
CREATE_OBJECT(ev, event, EVENT, vkCreateEvent, pEvent); |
||||
|
||||
util_hash_table_set_u64(ctx->object_table, ev->base.id, ev); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkDestroyEvent(struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkDestroyEvent *args) |
||||
{ |
||||
struct vkr_context *ctx = dispatch->data; |
||||
|
||||
DESTROY_OBJECT(ev, event, EVENT, vkDestroyEvent, event); |
||||
|
||||
util_hash_table_remove_u64(ctx->object_table, ev->base.id); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkGetEventStatus(UNUSED struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkGetEventStatus *args) |
||||
{ |
||||
vn_replace_vkGetEventStatus_args_handle(args); |
||||
args->ret = vkGetEventStatus(args->device, args->event); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkSetEvent(UNUSED struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkSetEvent *args) |
||||
{ |
||||
vn_replace_vkSetEvent_args_handle(args); |
||||
args->ret = vkSetEvent(args->device, args->event); |
||||
} |
||||
|
||||
static void |
||||
vkr_dispatch_vkResetEvent(UNUSED struct vn_dispatch_context *dispatch, |
||||
struct vn_command_vkResetEvent *args) |
||||
{ |
||||
vn_replace_vkResetEvent_args_handle(args); |
||||
args->ret = vkResetEvent(args->device, args->event); |
||||
} |
||||
|
||||
void |
||||
vkr_context_init_queue_dispatch(struct vkr_context *ctx) |
||||
{ |
||||
struct vn_dispatch_context *dispatch = &ctx->dispatch; |
||||
|
||||
dispatch->dispatch_vkGetDeviceQueue = vkr_dispatch_vkGetDeviceQueue; |
||||
dispatch->dispatch_vkGetDeviceQueue2 = vkr_dispatch_vkGetDeviceQueue2; |
||||
dispatch->dispatch_vkQueueSubmit = vkr_dispatch_vkQueueSubmit; |
||||
dispatch->dispatch_vkQueueBindSparse = vkr_dispatch_vkQueueBindSparse; |
||||
dispatch->dispatch_vkQueueWaitIdle = vkr_dispatch_vkQueueWaitIdle; |
||||
} |
||||
|
||||
void |
||||
vkr_context_init_fence_dispatch(struct vkr_context *ctx) |
||||
{ |
||||
struct vn_dispatch_context *dispatch = &ctx->dispatch; |
||||
|
||||
dispatch->dispatch_vkCreateFence = vkr_dispatch_vkCreateFence; |
||||
dispatch->dispatch_vkDestroyFence = vkr_dispatch_vkDestroyFence; |
||||
dispatch->dispatch_vkResetFences = vkr_dispatch_vkResetFences; |
||||
dispatch->dispatch_vkGetFenceStatus = vkr_dispatch_vkGetFenceStatus; |
||||
dispatch->dispatch_vkWaitForFences = vkr_dispatch_vkWaitForFences; |
||||
} |
||||
|
||||
void |
||||
vkr_context_init_semaphore_dispatch(struct vkr_context *ctx) |
||||
{ |
||||
struct vn_dispatch_context *dispatch = &ctx->dispatch; |
||||
|
||||
dispatch->dispatch_vkCreateSemaphore = vkr_dispatch_vkCreateSemaphore; |
||||
dispatch->dispatch_vkDestroySemaphore = vkr_dispatch_vkDestroySemaphore; |
||||
dispatch->dispatch_vkGetSemaphoreCounterValue = |
||||
vkr_dispatch_vkGetSemaphoreCounterValue; |
||||
dispatch->dispatch_vkWaitSemaphores = vkr_dispatch_vkWaitSemaphores; |
||||
dispatch->dispatch_vkSignalSemaphore = vkr_dispatch_vkSignalSemaphore; |
||||
} |
||||
|
||||
void |
||||
vkr_context_init_event_dispatch(struct vkr_context *ctx) |
||||
{ |
||||
struct vn_dispatch_context *dispatch = &ctx->dispatch; |
||||
|
||||
dispatch->dispatch_vkCreateEvent = vkr_dispatch_vkCreateEvent; |
||||
dispatch->dispatch_vkDestroyEvent = vkr_dispatch_vkDestroyEvent; |
||||
dispatch->dispatch_vkGetEventStatus = vkr_dispatch_vkGetEventStatus; |
||||
dispatch->dispatch_vkSetEvent = vkr_dispatch_vkSetEvent; |
||||
dispatch->dispatch_vkResetEvent = vkr_dispatch_vkResetEvent; |
||||
} |
Loading…
Reference in new issue