blob: c6cc99bbe0ea3c0a63dcf07c6b1dd75695ed7387 [file] [log] [blame]
/*
* Copyright 2014, 2015 Red Hat.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* on the rights to use, copy, modify, merge, publish, distribute, sub
* license, and/or sell copies of the Software, and to permit persons to whom
* the Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
* THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
* DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
* OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
* USE OR OTHER DEALINGS IN THE SOFTWARE.
*/
#include <errno.h>
#include <fcntl.h>
#include <limits.h>
#include <stdio.h>
#include <sys/ioctl.h>
#include <sys/stat.h>
#include "os/os_mman.h"
#include "util/os_file.h"
#include "util/os_time.h"
#include "util/u_memory.h"
#include "util/format/u_format.h"
#include "util/u_hash_table.h"
#include "util/u_inlines.h"
#include "util/u_pointer.h"
#include "frontend/drm_driver.h"
#include "virgl/virgl_screen.h"
#include "virgl/virgl_public.h"
#include <xf86drm.h>
#include <libsync.h>
#include "drm-uapi/virtgpu_drm.h"
#include "virgl_drm_winsys.h"
#include "virgl_drm_public.h"
#define VIRGL_DRM_VERSION(major, minor) ((major) << 16 | (minor))
#define VIRGL_DRM_VERSION_FENCE_FD VIRGL_DRM_VERSION(0, 1)
/* Gets a pointer to the virgl_hw_res containing the pointed to cache entry. */
#define cache_entry_container_res(ptr) \
(struct virgl_hw_res*)((char*)ptr - offsetof(struct virgl_hw_res, cache_entry))
static inline boolean can_cache_resource_with_bind(uint32_t bind)
{
return bind == VIRGL_BIND_CONSTANT_BUFFER ||
bind == VIRGL_BIND_INDEX_BUFFER ||
bind == VIRGL_BIND_VERTEX_BUFFER ||
bind == VIRGL_BIND_CUSTOM ||
bind == VIRGL_BIND_STAGING;
}
static void virgl_hw_res_destroy(struct virgl_drm_winsys *qdws,
struct virgl_hw_res *res)
{
struct drm_gem_close args;
mtx_lock(&qdws->bo_handles_mutex);
_mesa_hash_table_remove_key(qdws->bo_handles,
(void *)(uintptr_t)res->bo_handle);
if (res->flink_name)
_mesa_hash_table_remove_key(qdws->bo_names,
(void *)(uintptr_t)res->flink_name);
mtx_unlock(&qdws->bo_handles_mutex);
if (res->ptr)
os_munmap(res->ptr, res->size);
memset(&args, 0, sizeof(args));
args.handle = res->bo_handle;
drmIoctl(qdws->fd, DRM_IOCTL_GEM_CLOSE, &args);
FREE(res);
}
static boolean virgl_drm_resource_is_busy(struct virgl_winsys *vws,
struct virgl_hw_res *res)
{
struct virgl_drm_winsys *vdws = virgl_drm_winsys(vws);
struct drm_virtgpu_3d_wait waitcmd;
int ret;
if (!p_atomic_read(&res->maybe_busy) && !p_atomic_read(&res->external))
return false;
memset(&waitcmd, 0, sizeof(waitcmd));
waitcmd.handle = res->bo_handle;
waitcmd.flags = VIRTGPU_WAIT_NOWAIT;
ret = drmIoctl(vdws->fd, DRM_IOCTL_VIRTGPU_WAIT, &waitcmd);
if (ret && errno == EBUSY)
return TRUE;
p_atomic_set(&res->maybe_busy, false);
return FALSE;
}
static void
virgl_drm_winsys_destroy(struct virgl_winsys *qws)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
virgl_resource_cache_flush(&qdws->cache);
_mesa_hash_table_destroy(qdws->bo_handles, NULL);
_mesa_hash_table_destroy(qdws->bo_names, NULL);
mtx_destroy(&qdws->bo_handles_mutex);
mtx_destroy(&qdws->mutex);
FREE(qdws);
}
static void virgl_drm_resource_reference(struct virgl_winsys *qws,
struct virgl_hw_res **dres,
struct virgl_hw_res *sres)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
struct virgl_hw_res *old = *dres;
if (pipe_reference(&(*dres)->reference, &sres->reference)) {
if (!can_cache_resource_with_bind(old->bind) ||
p_atomic_read(&old->external)) {
virgl_hw_res_destroy(qdws, old);
} else {
mtx_lock(&qdws->mutex);
virgl_resource_cache_add(&qdws->cache, &old->cache_entry);
mtx_unlock(&qdws->mutex);
}
}
*dres = sres;
}
static struct virgl_hw_res *
virgl_drm_winsys_resource_create(struct virgl_winsys *qws,
enum pipe_texture_target target,
uint32_t format,
uint32_t bind,
uint32_t width,
uint32_t height,
uint32_t depth,
uint32_t array_size,
uint32_t last_level,
uint32_t nr_samples,
uint32_t size,
bool for_fencing)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
struct drm_virtgpu_resource_create createcmd;
int ret;
struct virgl_hw_res *res;
uint32_t stride = width * util_format_get_blocksize(format);
res = CALLOC_STRUCT(virgl_hw_res);
if (!res)
return NULL;
memset(&createcmd, 0, sizeof(createcmd));
createcmd.target = target;
createcmd.format = pipe_to_virgl_format(format);
createcmd.bind = bind;
createcmd.width = width;
createcmd.height = height;
createcmd.depth = depth;
createcmd.array_size = array_size;
createcmd.last_level = last_level;
createcmd.nr_samples = nr_samples;
createcmd.stride = stride;
createcmd.size = size;
ret = drmIoctl(qdws->fd, DRM_IOCTL_VIRTGPU_RESOURCE_CREATE, &createcmd);
if (ret != 0) {
FREE(res);
return NULL;
}
res->bind = bind;
res->res_handle = createcmd.res_handle;
res->bo_handle = createcmd.bo_handle;
res->size = size;
pipe_reference_init(&res->reference, 1);
p_atomic_set(&res->external, false);
p_atomic_set(&res->num_cs_references, 0);
/* A newly created resource is considered busy by the kernel until the
* command is retired. But for our purposes, we can consider it idle
* unless it is used for fencing.
*/
p_atomic_set(&res->maybe_busy, for_fencing);
virgl_resource_cache_entry_init(&res->cache_entry, size, bind, format);
return res;
}
static int
virgl_bo_transfer_put(struct virgl_winsys *vws,
struct virgl_hw_res *res,
const struct pipe_box *box,
uint32_t stride, uint32_t layer_stride,
uint32_t buf_offset, uint32_t level)
{
struct virgl_drm_winsys *vdws = virgl_drm_winsys(vws);
struct drm_virtgpu_3d_transfer_to_host tohostcmd;
p_atomic_set(&res->maybe_busy, true);
memset(&tohostcmd, 0, sizeof(tohostcmd));
tohostcmd.bo_handle = res->bo_handle;
tohostcmd.box.x = box->x;
tohostcmd.box.y = box->y;
tohostcmd.box.z = box->z;
tohostcmd.box.w = box->width;
tohostcmd.box.h = box->height;
tohostcmd.box.d = box->depth;
tohostcmd.offset = buf_offset;
tohostcmd.level = level;
// tohostcmd.stride = stride;
// tohostcmd.layer_stride = stride;
return drmIoctl(vdws->fd, DRM_IOCTL_VIRTGPU_TRANSFER_TO_HOST, &tohostcmd);
}
static int
virgl_bo_transfer_get(struct virgl_winsys *vws,
struct virgl_hw_res *res,
const struct pipe_box *box,
uint32_t stride, uint32_t layer_stride,
uint32_t buf_offset, uint32_t level)
{
struct virgl_drm_winsys *vdws = virgl_drm_winsys(vws);
struct drm_virtgpu_3d_transfer_from_host fromhostcmd;
p_atomic_set(&res->maybe_busy, true);
memset(&fromhostcmd, 0, sizeof(fromhostcmd));
fromhostcmd.bo_handle = res->bo_handle;
fromhostcmd.level = level;
fromhostcmd.offset = buf_offset;
// fromhostcmd.stride = stride;
// fromhostcmd.layer_stride = layer_stride;
fromhostcmd.box.x = box->x;
fromhostcmd.box.y = box->y;
fromhostcmd.box.z = box->z;
fromhostcmd.box.w = box->width;
fromhostcmd.box.h = box->height;
fromhostcmd.box.d = box->depth;
return drmIoctl(vdws->fd, DRM_IOCTL_VIRTGPU_TRANSFER_FROM_HOST, &fromhostcmd);
}
static struct virgl_hw_res *
virgl_drm_winsys_resource_cache_create(struct virgl_winsys *qws,
enum pipe_texture_target target,
uint32_t format,
uint32_t bind,
uint32_t width,
uint32_t height,
uint32_t depth,
uint32_t array_size,
uint32_t last_level,
uint32_t nr_samples,
uint32_t size)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
struct virgl_hw_res *res;
struct virgl_resource_cache_entry *entry;
if (!can_cache_resource_with_bind(bind))
goto alloc;
mtx_lock(&qdws->mutex);
entry = virgl_resource_cache_remove_compatible(&qdws->cache, size,
bind, format);
if (entry) {
res = cache_entry_container_res(entry);
mtx_unlock(&qdws->mutex);
pipe_reference_init(&res->reference, 1);
return res;
}
mtx_unlock(&qdws->mutex);
alloc:
res = virgl_drm_winsys_resource_create(qws, target, format, bind,
width, height, depth, array_size,
last_level, nr_samples, size, false);
return res;
}
static struct virgl_hw_res *
virgl_drm_winsys_resource_create_handle(struct virgl_winsys *qws,
struct winsys_handle *whandle,
uint32_t *plane,
uint32_t *stride,
uint32_t *plane_offset,
uint64_t *modifier)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
struct drm_gem_open open_arg = {};
struct drm_virtgpu_resource_info info_arg = {};
struct virgl_hw_res *res = NULL;
uint32_t handle = whandle->handle;
if (whandle->offset != 0 && whandle->type == WINSYS_HANDLE_TYPE_SHARED) {
_debug_printf("attempt to import unsupported winsys offset %u\n",
whandle->offset);
return NULL;
} else if (whandle->type == WINSYS_HANDLE_TYPE_FD) {
*plane = whandle->plane;
*stride = whandle->stride;
*plane_offset = whandle->offset;
*modifier = whandle->modifier;
}
mtx_lock(&qdws->bo_handles_mutex);
/* We must maintain a list of pairs <handle, bo>, so that we always return
* the same BO for one particular handle. If we didn't do that and created
* more than one BO for the same handle and then relocated them in a CS,
* we would hit a deadlock in the kernel.
*
* The list of pairs is guarded by a mutex, of course. */
if (whandle->type == WINSYS_HANDLE_TYPE_SHARED) {
res = util_hash_table_get(qdws->bo_names, (void*)(uintptr_t)handle);
} else if (whandle->type == WINSYS_HANDLE_TYPE_FD) {
int r;
r = drmPrimeFDToHandle(qdws->fd, whandle->handle, &handle);
if (r)
goto done;
res = util_hash_table_get(qdws->bo_handles, (void*)(uintptr_t)handle);
} else {
/* Unknown handle type */
goto done;
}
if (res) {
struct virgl_hw_res *r = NULL;
virgl_drm_resource_reference(&qdws->base, &r, res);
goto done;
}
res = CALLOC_STRUCT(virgl_hw_res);
if (!res)
goto done;
if (whandle->type == WINSYS_HANDLE_TYPE_FD) {
res->bo_handle = handle;
} else {
memset(&open_arg, 0, sizeof(open_arg));
open_arg.name = whandle->handle;
if (drmIoctl(qdws->fd, DRM_IOCTL_GEM_OPEN, &open_arg)) {
FREE(res);
res = NULL;
goto done;
}
res->bo_handle = open_arg.handle;
res->flink_name = whandle->handle;
}
memset(&info_arg, 0, sizeof(info_arg));
info_arg.bo_handle = res->bo_handle;
if (drmIoctl(qdws->fd, DRM_IOCTL_VIRTGPU_RESOURCE_INFO, &info_arg)) {
/* close */
FREE(res);
res = NULL;
goto done;
}
res->res_handle = info_arg.res_handle;
res->size = info_arg.size;
pipe_reference_init(&res->reference, 1);
p_atomic_set(&res->external, true);
res->num_cs_references = 0;
if (res->flink_name)
_mesa_hash_table_insert(qdws->bo_names, (void *)(uintptr_t)res->flink_name, res);
_mesa_hash_table_insert(qdws->bo_handles, (void *)(uintptr_t)res->bo_handle, res);
done:
mtx_unlock(&qdws->bo_handles_mutex);
return res;
}
static boolean virgl_drm_winsys_resource_get_handle(struct virgl_winsys *qws,
struct virgl_hw_res *res,
uint32_t stride,
struct winsys_handle *whandle)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
struct drm_gem_flink flink;
if (!res)
return FALSE;
if (whandle->type == WINSYS_HANDLE_TYPE_SHARED) {
if (!res->flink_name) {
memset(&flink, 0, sizeof(flink));
flink.handle = res->bo_handle;
if (drmIoctl(qdws->fd, DRM_IOCTL_GEM_FLINK, &flink)) {
return FALSE;
}
res->flink_name = flink.name;
mtx_lock(&qdws->bo_handles_mutex);
_mesa_hash_table_insert(qdws->bo_names, (void *)(uintptr_t)res->flink_name, res);
mtx_unlock(&qdws->bo_handles_mutex);
}
whandle->handle = res->flink_name;
} else if (whandle->type == WINSYS_HANDLE_TYPE_KMS) {
whandle->handle = res->bo_handle;
} else if (whandle->type == WINSYS_HANDLE_TYPE_FD) {
if (drmPrimeHandleToFD(qdws->fd, res->bo_handle, DRM_CLOEXEC, (int*)&whandle->handle))
return FALSE;
mtx_lock(&qdws->bo_handles_mutex);
_mesa_hash_table_insert(qdws->bo_handles, (void *)(uintptr_t)res->bo_handle, res);
mtx_unlock(&qdws->bo_handles_mutex);
}
p_atomic_set(&res->external, true);
whandle->stride = stride;
return TRUE;
}
static void *virgl_drm_resource_map(struct virgl_winsys *qws,
struct virgl_hw_res *res)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
struct drm_virtgpu_map mmap_arg;
void *ptr;
if (res->ptr)
return res->ptr;
memset(&mmap_arg, 0, sizeof(mmap_arg));
mmap_arg.handle = res->bo_handle;
if (drmIoctl(qdws->fd, DRM_IOCTL_VIRTGPU_MAP, &mmap_arg))
return NULL;
ptr = os_mmap(0, res->size, PROT_READ|PROT_WRITE, MAP_SHARED,
qdws->fd, mmap_arg.offset);
if (ptr == MAP_FAILED)
return NULL;
res->ptr = ptr;
return ptr;
}
static void virgl_drm_resource_wait(struct virgl_winsys *qws,
struct virgl_hw_res *res)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
struct drm_virtgpu_3d_wait waitcmd;
int ret;
if (!p_atomic_read(&res->maybe_busy) && !p_atomic_read(&res->external))
return;
memset(&waitcmd, 0, sizeof(waitcmd));
waitcmd.handle = res->bo_handle;
ret = drmIoctl(qdws->fd, DRM_IOCTL_VIRTGPU_WAIT, &waitcmd);
if (ret)
_debug_printf("waiting got error - %d, slow gpu or hang?\n", errno);
p_atomic_set(&res->maybe_busy, false);
}
static bool virgl_drm_alloc_res_list(struct virgl_drm_cmd_buf *cbuf,
int initial_size)
{
cbuf->nres = initial_size;
cbuf->cres = 0;
cbuf->res_bo = CALLOC(cbuf->nres, sizeof(struct virgl_hw_buf*));
if (!cbuf->res_bo)
return false;
cbuf->res_hlist = MALLOC(cbuf->nres * sizeof(uint32_t));
if (!cbuf->res_hlist) {
FREE(cbuf->res_bo);
return false;
}
return true;
}
static void virgl_drm_free_res_list(struct virgl_drm_cmd_buf *cbuf)
{
int i;
for (i = 0; i < cbuf->cres; i++) {
p_atomic_dec(&cbuf->res_bo[i]->num_cs_references);
virgl_drm_resource_reference(cbuf->ws, &cbuf->res_bo[i], NULL);
}
FREE(cbuf->res_hlist);
FREE(cbuf->res_bo);
}
static boolean virgl_drm_lookup_res(struct virgl_drm_cmd_buf *cbuf,
struct virgl_hw_res *res)
{
unsigned hash = res->res_handle & (sizeof(cbuf->is_handle_added)-1);
int i;
if (cbuf->is_handle_added[hash]) {
i = cbuf->reloc_indices_hashlist[hash];
if (cbuf->res_bo[i] == res)
return true;
for (i = 0; i < cbuf->cres; i++) {
if (cbuf->res_bo[i] == res) {
cbuf->reloc_indices_hashlist[hash] = i;
return true;
}
}
}
return false;
}
static void virgl_drm_add_res(struct virgl_drm_winsys *qdws,
struct virgl_drm_cmd_buf *cbuf,
struct virgl_hw_res *res)
{
unsigned hash = res->res_handle & (sizeof(cbuf->is_handle_added)-1);
if (cbuf->cres >= cbuf->nres) {
unsigned new_nres = cbuf->nres + 256;
void *new_ptr = REALLOC(cbuf->res_bo,
cbuf->nres * sizeof(struct virgl_hw_buf*),
new_nres * sizeof(struct virgl_hw_buf*));
if (!new_ptr) {
_debug_printf("failure to add relocation %d, %d\n", cbuf->cres, new_nres);
return;
}
cbuf->res_bo = new_ptr;
new_ptr = REALLOC(cbuf->res_hlist,
cbuf->nres * sizeof(uint32_t),
new_nres * sizeof(uint32_t));
if (!new_ptr) {
_debug_printf("failure to add hlist relocation %d, %d\n", cbuf->cres, cbuf->nres);
return;
}
cbuf->res_hlist = new_ptr;
cbuf->nres = new_nres;
}
cbuf->res_bo[cbuf->cres] = NULL;
virgl_drm_resource_reference(&qdws->base, &cbuf->res_bo[cbuf->cres], res);
cbuf->res_hlist[cbuf->cres] = res->bo_handle;
cbuf->is_handle_added[hash] = TRUE;
cbuf->reloc_indices_hashlist[hash] = cbuf->cres;
p_atomic_inc(&res->num_cs_references);
cbuf->cres++;
}
/* This is called after the cbuf is submitted. */
static void virgl_drm_clear_res_list(struct virgl_drm_cmd_buf *cbuf)
{
int i;
for (i = 0; i < cbuf->cres; i++) {
/* mark all BOs busy after submission */
p_atomic_set(&cbuf->res_bo[i]->maybe_busy, true);
p_atomic_dec(&cbuf->res_bo[i]->num_cs_references);
virgl_drm_resource_reference(cbuf->ws, &cbuf->res_bo[i], NULL);
}
cbuf->cres = 0;
memset(cbuf->is_handle_added, 0, sizeof(cbuf->is_handle_added));
}
static void virgl_drm_emit_res(struct virgl_winsys *qws,
struct virgl_cmd_buf *_cbuf,
struct virgl_hw_res *res, boolean write_buf)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
struct virgl_drm_cmd_buf *cbuf = virgl_drm_cmd_buf(_cbuf);
boolean already_in_list = virgl_drm_lookup_res(cbuf, res);
if (write_buf)
cbuf->base.buf[cbuf->base.cdw++] = res->res_handle;
if (!already_in_list)
virgl_drm_add_res(qdws, cbuf, res);
}
static boolean virgl_drm_res_is_ref(struct virgl_winsys *qws,
struct virgl_cmd_buf *_cbuf,
struct virgl_hw_res *res)
{
if (!p_atomic_read(&res->num_cs_references))
return FALSE;
return TRUE;
}
static struct virgl_cmd_buf *virgl_drm_cmd_buf_create(struct virgl_winsys *qws,
uint32_t size)
{
struct virgl_drm_cmd_buf *cbuf;
cbuf = CALLOC_STRUCT(virgl_drm_cmd_buf);
if (!cbuf)
return NULL;
cbuf->ws = qws;
if (!virgl_drm_alloc_res_list(cbuf, 512)) {
FREE(cbuf);
return NULL;
}
cbuf->buf = CALLOC(size, sizeof(uint32_t));
if (!cbuf->buf) {
FREE(cbuf->res_hlist);
FREE(cbuf->res_bo);
FREE(cbuf);
return NULL;
}
cbuf->in_fence_fd = -1;
cbuf->base.buf = cbuf->buf;
return &cbuf->base;
}
static void virgl_drm_cmd_buf_destroy(struct virgl_cmd_buf *_cbuf)
{
struct virgl_drm_cmd_buf *cbuf = virgl_drm_cmd_buf(_cbuf);
virgl_drm_free_res_list(cbuf);
FREE(cbuf->buf);
FREE(cbuf);
}
static struct pipe_fence_handle *
virgl_drm_fence_create(struct virgl_winsys *vws, int fd, bool external)
{
struct virgl_drm_fence *fence;
assert(vws->supports_fences);
if (external) {
fd = os_dupfd_cloexec(fd);
if (fd < 0)
return NULL;
}
fence = CALLOC_STRUCT(virgl_drm_fence);
if (!fence) {
close(fd);
return NULL;
}
fence->fd = fd;
fence->external = external;
pipe_reference_init(&fence->reference, 1);
return (struct pipe_fence_handle *)fence;
}
static struct pipe_fence_handle *
virgl_drm_fence_create_legacy(struct virgl_winsys *vws)
{
struct virgl_drm_fence *fence;
assert(!vws->supports_fences);
fence = CALLOC_STRUCT(virgl_drm_fence);
if (!fence)
return NULL;
fence->fd = -1;
/* Resources for fences should not be from the cache, since we are basing
* the fence status on the resource creation busy status.
*/
fence->hw_res = virgl_drm_winsys_resource_create(vws, PIPE_BUFFER,
PIPE_FORMAT_R8_UNORM, VIRGL_BIND_CUSTOM, 8, 1, 1, 0, 0, 0, 8, true);
if (!fence->hw_res) {
FREE(fence);
return NULL;
}
pipe_reference_init(&fence->reference, 1);
return (struct pipe_fence_handle *)fence;
}
static int virgl_drm_winsys_submit_cmd(struct virgl_winsys *qws,
struct virgl_cmd_buf *_cbuf,
struct pipe_fence_handle **fence)
{
struct virgl_drm_winsys *qdws = virgl_drm_winsys(qws);
struct virgl_drm_cmd_buf *cbuf = virgl_drm_cmd_buf(_cbuf);
struct drm_virtgpu_execbuffer eb;
int ret;
if (cbuf->base.cdw == 0)
return 0;
memset(&eb, 0, sizeof(struct drm_virtgpu_execbuffer));
eb.command = (unsigned long)(void*)cbuf->buf;
eb.size = cbuf->base.cdw * 4;
eb.num_bo_handles = cbuf->cres;
eb.bo_handles = (unsigned long)(void *)cbuf->res_hlist;
eb.fence_fd = -1;
if (qws->supports_fences) {
if (cbuf->in_fence_fd >= 0) {
eb.flags |= VIRTGPU_EXECBUF_FENCE_FD_IN;
eb.fence_fd = cbuf->in_fence_fd;
}
if (fence != NULL)
eb.flags |= VIRTGPU_EXECBUF_FENCE_FD_OUT;
} else {
assert(cbuf->in_fence_fd < 0);
}
ret = drmIoctl(qdws->fd, DRM_IOCTL_VIRTGPU_EXECBUFFER, &eb);
if (ret == -1)
_debug_printf("got error from kernel - expect bad rendering %d\n", errno);
cbuf->base.cdw = 0;
if (qws->supports_fences) {
if (cbuf->in_fence_fd >= 0) {
close(cbuf->in_fence_fd);
cbuf->in_fence_fd = -1;
}
if (fence != NULL && ret == 0)
*fence = virgl_drm_fence_create(qws, eb.fence_fd, false);
} else {
if (fence != NULL && ret == 0)
*fence = virgl_drm_fence_create_legacy(qws);
}
virgl_drm_clear_res_list(cbuf);
return ret;
}
static int virgl_drm_get_caps(struct virgl_winsys *vws,
struct virgl_drm_caps *caps)
{
struct virgl_drm_winsys *vdws = virgl_drm_winsys(vws);
struct drm_virtgpu_get_caps args;
int ret;
virgl_ws_fill_new_caps_defaults(caps);
memset(&args, 0, sizeof(args));
if (vdws->has_capset_query_fix) {
/* if we have the query fix - try and get cap set id 2 first */
args.cap_set_id = 2;
args.size = sizeof(union virgl_caps);
} else {
args.cap_set_id = 1;
args.size = sizeof(struct virgl_caps_v1);
}
args.addr = (unsigned long)&caps->caps;
ret = drmIoctl(vdws->fd, DRM_IOCTL_VIRTGPU_GET_CAPS, &args);
if (ret == -1 && errno == EINVAL) {
/* Fallback to v1 */
args.cap_set_id = 1;
args.size = sizeof(struct virgl_caps_v1);
ret = drmIoctl(vdws->fd, DRM_IOCTL_VIRTGPU_GET_CAPS, &args);
if (ret == -1)
return ret;
}
return ret;
}
static struct pipe_fence_handle *
virgl_cs_create_fence(struct virgl_winsys *vws, int fd)
{
if (!vws->supports_fences)
return NULL;
return virgl_drm_fence_create(vws, fd, true);
}
static bool virgl_fence_wait(struct virgl_winsys *vws,
struct pipe_fence_handle *_fence,
uint64_t timeout)
{
struct virgl_drm_fence *fence = virgl_drm_fence(_fence);
if (vws->supports_fences) {
uint64_t timeout_ms;
int timeout_poll;
if (timeout == 0)
return sync_wait(fence->fd, 0) == 0;
timeout_ms = timeout / 1000000;
/* round up */
if (timeout_ms * 1000000 < timeout)
timeout_ms++;
timeout_poll = timeout_ms <= INT_MAX ? (int) timeout_ms : -1;
return sync_wait(fence->fd, timeout_poll) == 0;
}
if (timeout == 0)
return !virgl_drm_resource_is_busy(vws, fence->hw_res);
if (timeout != PIPE_TIMEOUT_INFINITE) {
int64_t start_time = os_time_get();
timeout /= 1000;
while (virgl_drm_resource_is_busy(vws, fence->hw_res)) {
if (os_time_get() - start_time >= timeout)
return FALSE;
os_time_sleep(10);
}
return TRUE;
}
virgl_drm_resource_wait(vws, fence->hw_res);
return TRUE;
}
static void virgl_fence_reference(struct virgl_winsys *vws,
struct pipe_fence_handle **dst,
struct pipe_fence_handle *src)
{
struct virgl_drm_fence *dfence = virgl_drm_fence(*dst);
struct virgl_drm_fence *sfence = virgl_drm_fence(src);
if (pipe_reference(&dfence->reference, &sfence->reference)) {
if (vws->supports_fences) {
close(dfence->fd);
} else {
struct virgl_drm_winsys *vdws = virgl_drm_winsys(vws);
virgl_hw_res_destroy(vdws, dfence->hw_res);
}
FREE(dfence);
}
*dst = src;
}
static void virgl_fence_server_sync(struct virgl_winsys *vws,
struct virgl_cmd_buf *_cbuf,
struct pipe_fence_handle *_fence)
{
struct virgl_drm_cmd_buf *cbuf = virgl_drm_cmd_buf(_cbuf);
struct virgl_drm_fence *fence = virgl_drm_fence(_fence);
if (!vws->supports_fences)
return;
/* if not an external fence, then nothing more to do without preemption: */
if (!fence->external)
return;
sync_accumulate("virgl", &cbuf->in_fence_fd, fence->fd);
}
static int virgl_fence_get_fd(struct virgl_winsys *vws,
struct pipe_fence_handle *_fence)
{
struct virgl_drm_fence *fence = virgl_drm_fence(_fence);
if (!vws->supports_fences)
return -1;
return os_dupfd_cloexec(fence->fd);
}
static int virgl_drm_get_version(int fd)
{
int ret;
drmVersionPtr version;
version = drmGetVersion(fd);
if (!version)
ret = -EFAULT;
else if (version->version_major != 0)
ret = -EINVAL;
else
ret = VIRGL_DRM_VERSION(0, version->version_minor);
drmFreeVersion(version);
return ret;
}
static bool
virgl_drm_resource_cache_entry_is_busy(struct virgl_resource_cache_entry *entry,
void *user_data)
{
struct virgl_drm_winsys *qdws = user_data;
struct virgl_hw_res *res = cache_entry_container_res(entry);
return virgl_drm_resource_is_busy(&qdws->base, res);
}
static void
virgl_drm_resource_cache_entry_release(struct virgl_resource_cache_entry *entry,
void *user_data)
{
struct virgl_drm_winsys *qdws = user_data;
struct virgl_hw_res *res = cache_entry_container_res(entry);
virgl_hw_res_destroy(qdws, res);
}
static struct virgl_winsys *
virgl_drm_winsys_create(int drmFD)
{
static const unsigned CACHE_TIMEOUT_USEC = 1000000;
struct virgl_drm_winsys *qdws;
int drm_version;
int ret;
int gl = 0;
struct drm_virtgpu_getparam getparam = {0};
getparam.param = VIRTGPU_PARAM_3D_FEATURES;
getparam.value = (uint64_t)(uintptr_t)&gl;
ret = drmIoctl(drmFD, DRM_IOCTL_VIRTGPU_GETPARAM, &getparam);
if (ret < 0 || !gl)
return NULL;
drm_version = virgl_drm_get_version(drmFD);
if (drm_version < 0)
return NULL;
qdws = CALLOC_STRUCT(virgl_drm_winsys);
if (!qdws)
return NULL;
qdws->fd = drmFD;
virgl_resource_cache_init(&qdws->cache, CACHE_TIMEOUT_USEC,
virgl_drm_resource_cache_entry_is_busy,
virgl_drm_resource_cache_entry_release,
qdws);
(void) mtx_init(&qdws->mutex, mtx_plain);
(void) mtx_init(&qdws->bo_handles_mutex, mtx_plain);
qdws->bo_handles = util_hash_table_create_ptr_keys();
qdws->bo_names = util_hash_table_create_ptr_keys();
qdws->base.destroy = virgl_drm_winsys_destroy;
qdws->base.transfer_put = virgl_bo_transfer_put;
qdws->base.transfer_get = virgl_bo_transfer_get;
qdws->base.resource_create = virgl_drm_winsys_resource_cache_create;
qdws->base.resource_reference = virgl_drm_resource_reference;
qdws->base.resource_create_from_handle = virgl_drm_winsys_resource_create_handle;
qdws->base.resource_get_handle = virgl_drm_winsys_resource_get_handle;
qdws->base.resource_map = virgl_drm_resource_map;
qdws->base.resource_wait = virgl_drm_resource_wait;
qdws->base.resource_is_busy = virgl_drm_resource_is_busy;
qdws->base.cmd_buf_create = virgl_drm_cmd_buf_create;
qdws->base.cmd_buf_destroy = virgl_drm_cmd_buf_destroy;
qdws->base.submit_cmd = virgl_drm_winsys_submit_cmd;
qdws->base.emit_res = virgl_drm_emit_res;
qdws->base.res_is_referenced = virgl_drm_res_is_ref;
qdws->base.cs_create_fence = virgl_cs_create_fence;
qdws->base.fence_wait = virgl_fence_wait;
qdws->base.fence_reference = virgl_fence_reference;
qdws->base.fence_server_sync = virgl_fence_server_sync;
qdws->base.fence_get_fd = virgl_fence_get_fd;
qdws->base.supports_fences = drm_version >= VIRGL_DRM_VERSION_FENCE_FD;
qdws->base.supports_encoded_transfers = 1;
qdws->base.get_caps = virgl_drm_get_caps;
uint32_t value = 0;
getparam.param = VIRTGPU_PARAM_CAPSET_QUERY_FIX;
getparam.value = (uint64_t)(uintptr_t)&value;
ret = drmIoctl(qdws->fd, DRM_IOCTL_VIRTGPU_GETPARAM, &getparam);
if (ret == 0) {
if (value == 1)
qdws->has_capset_query_fix = true;
}
return &qdws->base;
}
static struct hash_table *fd_tab = NULL;
static mtx_t virgl_screen_mutex = _MTX_INITIALIZER_NP;
static void
virgl_drm_screen_destroy(struct pipe_screen *pscreen)
{
struct virgl_screen *screen = virgl_screen(pscreen);
boolean destroy;
mtx_lock(&virgl_screen_mutex);
destroy = --screen->refcnt == 0;
if (destroy) {
int fd = virgl_drm_winsys(screen->vws)->fd;
_mesa_hash_table_remove_key(fd_tab, intptr_to_pointer(fd));
close(fd);
}
mtx_unlock(&virgl_screen_mutex);
if (destroy) {
pscreen->destroy = screen->winsys_priv;
pscreen->destroy(pscreen);
}
}
struct pipe_screen *
virgl_drm_screen_create(int fd, const struct pipe_screen_config *config)
{
struct pipe_screen *pscreen = NULL;
mtx_lock(&virgl_screen_mutex);
if (!fd_tab) {
fd_tab = util_hash_table_create_fd_keys();
if (!fd_tab)
goto unlock;
}
pscreen = util_hash_table_get(fd_tab, intptr_to_pointer(fd));
if (pscreen) {
virgl_screen(pscreen)->refcnt++;
} else {
struct virgl_winsys *vws;
int dup_fd = os_dupfd_cloexec(fd);
vws = virgl_drm_winsys_create(dup_fd);
if (!vws) {
close(dup_fd);
goto unlock;
}
pscreen = virgl_create_screen(vws, config);
if (pscreen) {
_mesa_hash_table_insert(fd_tab, intptr_to_pointer(dup_fd), pscreen);
/* Bit of a hack, to avoid circular linkage dependency,
* ie. pipe driver having to call in to winsys, we
* override the pipe drivers screen->destroy():
*/
virgl_screen(pscreen)->winsys_priv = pscreen->destroy;
pscreen->destroy = virgl_drm_screen_destroy;
}
}
unlock:
mtx_unlock(&virgl_screen_mutex);
return pscreen;
}