mesa/src/virtio/vulkan/vn_instance.h
Yiwei Zhang dfcad90240 venus: rename common vk object base member to vk
This way we avoid the confusing base.base or even base.base.base when
venus uses common objects. This also aligns with the naming of the other
drivers.

Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/34083>
2025-03-15 22:14:58 +00:00

129 lines
3.7 KiB
C

/*
* Copyright 2019 Google LLC
* SPDX-License-Identifier: MIT
*
* based in part on anv and radv which are:
* Copyright © 2015 Intel Corporation
* Copyright © 2016 Red Hat.
* Copyright © 2016 Bas Nieuwenhuizen
*/
#ifndef VN_INSTANCE_H
#define VN_INSTANCE_H
#include "vn_common.h"
#include "vn_renderer_util.h"
/* require and request at least Vulkan 1.1 at both instance and device levels
*/
#define VN_MIN_RENDERER_VERSION VK_API_VERSION_1_1
/* max advertised version at both instance and device levels */
#if !defined(ANDROID_STRICT) || ANDROID_API_LEVEL >= 36
#define VN_MAX_API_VERSION VK_MAKE_VERSION(1, 4, VK_HEADER_VERSION)
#elif ANDROID_API_LEVEL >= 33
#define VN_MAX_API_VERSION VK_MAKE_VERSION(1, 3, VK_HEADER_VERSION)
#else
#define VN_MAX_API_VERSION VK_MAKE_VERSION(1, 1, VK_HEADER_VERSION)
#endif
struct vn_instance {
struct vn_instance_base base;
struct driOptionCache dri_options;
struct driOptionCache available_dri_options;
bool enable_wsi_multi_plane_modifiers;
struct vn_renderer *renderer;
/* for VN_CS_ENCODER_STORAGE_SHMEM_POOL */
struct vn_renderer_shmem_pool cs_shmem_pool;
struct vn_renderer_shmem_pool reply_shmem_pool;
mtx_t ring_idx_mutex;
uint64_t ring_idx_used_mask;
struct {
struct vn_ring *ring;
struct list_head tls_rings;
struct vn_watchdog watchdog;
} ring;
/* Between the driver and the app, VN_MAX_API_VERSION is what we advertise
* and base.base.app_info.api_version is what the app requests.
*
* Between the driver and the renderer, renderer_api_version is the api
* version we request internally, which can be higher than
* base.base.app_info.api_version. renderer_version is the instance
* version we can use internally.
*/
uint32_t renderer_api_version;
uint32_t renderer_version;
bool engine_is_zink;
struct {
mtx_t mutex;
bool initialized;
struct vn_physical_device *devices;
uint32_t device_count;
VkPhysicalDeviceGroupProperties *groups;
uint32_t group_count;
} physical_device;
};
VK_DEFINE_HANDLE_CASTS(vn_instance,
base.vk.base,
VkInstance,
VK_OBJECT_TYPE_INSTANCE)
static inline struct vn_renderer_shmem *
vn_instance_cs_shmem_alloc(struct vn_instance *instance,
size_t size,
size_t *out_offset)
{
return vn_renderer_shmem_pool_alloc(
instance->renderer, &instance->cs_shmem_pool, size, out_offset);
}
static inline struct vn_renderer_shmem *
vn_instance_reply_shmem_alloc(struct vn_instance *instance,
size_t size,
size_t *out_offset)
{
return vn_renderer_shmem_pool_alloc(
instance->renderer, &instance->reply_shmem_pool, size, out_offset);
}
static inline int
vn_instance_acquire_ring_idx(struct vn_instance *instance)
{
mtx_lock(&instance->ring_idx_mutex);
int ring_idx = ffsll(~instance->ring_idx_used_mask) - 1;
if (ring_idx >= instance->renderer->info.max_timeline_count)
ring_idx = -1;
if (ring_idx > 0)
instance->ring_idx_used_mask |= (1ULL << (uint32_t)ring_idx);
mtx_unlock(&instance->ring_idx_mutex);
assert(ring_idx); /* never acquire the dedicated CPU ring */
/* returns -1 when no vacant rings */
return ring_idx;
}
static inline void
vn_instance_release_ring_idx(struct vn_instance *instance, uint32_t ring_idx)
{
assert(ring_idx > 0);
mtx_lock(&instance->ring_idx_mutex);
assert(instance->ring_idx_used_mask & (1ULL << ring_idx));
instance->ring_idx_used_mask &= ~(1ULL << ring_idx);
mtx_unlock(&instance->ring_idx_mutex);
}
#endif /* VN_INSTANCE_H */