blob: 5c4e7b71254b366f3fcea770a758fa529b52d633 [file] [log] [blame]
/*
* Copyright © 2007, 2011, 2013, 2014 Intel Corporation
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
* IN THE SOFTWARE.
*
* Authors:
* Eric Anholt <eric@anholt.net>
* Daniel Vetter <daniel.vetter@ffwll.ch>
*
*/
#ifdef HAVE_LIBGEN_H
#include <libgen.h>
#endif
#include <stdio.h>
#include <fcntl.h>
#include <sys/stat.h>
#include <sys/ioctl.h>
#include <string.h>
#include <sys/mman.h>
#include <signal.h>
#include <pciaccess.h>
#include <getopt.h>
#include <stdlib.h>
#include <unistd.h>
#include <sys/wait.h>
#include <sys/types.h>
#include <sys/syscall.h>
#include <sys/utsname.h>
#include <termios.h>
#include <errno.h>
#include "drmtest.h"
#include "i915_drm.h"
#include "intel_batchbuffer.h"
#include "intel_chipset.h"
#include "intel_io.h"
#include "igt_debugfs.h"
#include "igt_sysfs.h"
#include "config.h"
#ifdef HAVE_VALGRIND
#include <valgrind/valgrind.h>
#include <valgrind/memcheck.h>
#define VG(x) x
#else
#define VG(x) do {} while (0)
#endif
#include "ioctl_wrappers.h"
/**
* SECTION:ioctl_wrappers
* @short_description: ioctl wrappers and related functions
* @title: ioctl wrappers
* @include: igt.h
*
* This helper library contains simple functions to wrap the raw drm/i915 kernel
* ioctls. The normal versions never pass any error codes to the caller and use
* igt_assert() to check for error conditions instead. For some ioctls raw
* wrappers which do pass on error codes are available. These raw wrappers have
* a __ prefix.
*
* For wrappers which check for feature bits there can also be two versions: The
* normal one simply returns a boolean to the caller. But when skipping the
* testcase entirely is the right action then it's better to use igt_skip()
* directly in the wrapper. Such functions have _require_ in their name to
* distinguish them.
*/
int (*igt_ioctl)(int fd, unsigned long request, void *arg) = drmIoctl;
/**
* gem_handle_to_libdrm_bo:
* @bufmgr: libdrm buffer manager instance
* @fd: open i915 drm file descriptor
* @name: buffer name in libdrm
* @handle: gem buffer object handle
*
* This helper function imports a raw gem buffer handle into the libdrm buffer
* manager.
*
* Returns: The imported libdrm buffer manager object.
*/
drm_intel_bo *
gem_handle_to_libdrm_bo(drm_intel_bufmgr *bufmgr, int fd, const char *name, uint32_t handle)
{
struct drm_gem_flink flink;
int ret;
drm_intel_bo *bo;
memset(&flink, 0, sizeof(handle));
flink.handle = handle;
ret = ioctl(fd, DRM_IOCTL_GEM_FLINK, &flink);
igt_assert(ret == 0);
errno = 0;
bo = drm_intel_bo_gem_create_from_name(bufmgr, name, flink.name);
igt_assert(bo);
return bo;
}
static int
__gem_get_tiling(int fd, struct drm_i915_gem_get_tiling *arg)
{
int err;
err = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_GET_TILING, arg))
err = -errno;
errno = 0;
return err;
}
/**
* gem_get_tiling:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @tiling: (out) tiling mode of the gem buffer
* @swizzle: (out) bit 6 swizzle mode
*
* This wraps the GET_TILING ioctl.
*
* Returns whether the actual physical tiling matches the reported tiling.
*/
bool
gem_get_tiling(int fd, uint32_t handle, uint32_t *tiling, uint32_t *swizzle)
{
struct drm_i915_gem_get_tiling get_tiling;
memset(&get_tiling, 0, sizeof(get_tiling));
get_tiling.handle = handle;
igt_assert_eq(__gem_get_tiling(fd, &get_tiling), 0);
*tiling = get_tiling.tiling_mode;
*swizzle = get_tiling.swizzle_mode;
return get_tiling.phys_swizzle_mode == get_tiling.swizzle_mode;
}
int __gem_set_tiling(int fd, uint32_t handle, uint32_t tiling, uint32_t stride)
{
struct drm_i915_gem_set_tiling st;
int ret;
/* The kernel doesn't know about these tiling modes, expects NONE */
if (tiling == I915_TILING_Yf || tiling == I915_TILING_Ys)
tiling = I915_TILING_NONE;
memset(&st, 0, sizeof(st));
do {
st.handle = handle;
st.tiling_mode = tiling;
st.stride = tiling ? stride : 0;
ret = ioctl(fd, DRM_IOCTL_I915_GEM_SET_TILING, &st);
} while (ret == -1 && (errno == EINTR || errno == EAGAIN));
if (ret != 0)
return -errno;
errno = 0;
igt_assert(st.tiling_mode == tiling);
return 0;
}
/**
* gem_set_tiling:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @tiling: tiling mode bits
* @stride: stride of the buffer when using a tiled mode, otherwise must be 0
*
* This wraps the SET_TILING ioctl.
*/
void gem_set_tiling(int fd, uint32_t handle, uint32_t tiling, uint32_t stride)
{
igt_assert(__gem_set_tiling(fd, handle, tiling, stride) == 0);
}
int __gem_set_caching(int fd, uint32_t handle, uint32_t caching)
{
struct drm_i915_gem_caching arg;
int err;
memset(&arg, 0, sizeof(arg));
arg.handle = handle;
arg.caching = caching;
err = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_SET_CACHING, &arg))
err = -errno;
errno = 0;
return err;
}
/**
* gem_set_caching:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @caching: caching mode bits
*
* This wraps the SET_CACHING ioctl. Note that this function internally calls
* igt_require() when SET_CACHING isn't available, hence automatically skips the
* test. Therefore always extract test logic which uses this into its own
* subtest.
*/
void gem_set_caching(int fd, uint32_t handle, uint32_t caching)
{
igt_require(__gem_set_caching(fd, handle, caching) == 0);
}
/**
* gem_get_caching:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This wraps the GET_CACHING ioctl.
*
* Returns: The current caching mode bits.
*/
uint32_t gem_get_caching(int fd, uint32_t handle)
{
struct drm_i915_gem_caching arg;
int ret;
memset(&arg, 0, sizeof(arg));
arg.handle = handle;
ret = ioctl(fd, DRM_IOCTL_I915_GEM_GET_CACHING, &arg);
igt_assert(ret == 0);
errno = 0;
return arg.caching;
}
/**
* gem_open:
* @fd: open i915 drm file descriptor
* @name: flink buffer name
*
* This wraps the GEM_OPEN ioctl, which is used to import an flink name.
*
* Returns: gem file-private buffer handle of the open object.
*/
uint32_t gem_open(int fd, uint32_t name)
{
struct drm_gem_open open_struct;
int ret;
memset(&open_struct, 0, sizeof(open_struct));
open_struct.name = name;
ret = ioctl(fd, DRM_IOCTL_GEM_OPEN, &open_struct);
igt_assert(ret == 0);
igt_assert(open_struct.handle != 0);
errno = 0;
return open_struct.handle;
}
/**
* gem_flink:
* @fd: open i915 drm file descriptor
* @handle: file-private gem buffer object handle
*
* This wraps the GEM_FLINK ioctl, which is used to export a gem buffer object
* into the device-global flink namespace. See gem_open() for opening such a
* buffer name on a different i915 drm file descriptor.
*
* Returns: The created flink buffer name.
*/
uint32_t gem_flink(int fd, uint32_t handle)
{
struct drm_gem_flink flink;
int ret;
memset(&flink, 0, sizeof(flink));
flink.handle = handle;
ret = ioctl(fd, DRM_IOCTL_GEM_FLINK, &flink);
igt_assert(ret == 0);
errno = 0;
return flink.name;
}
/**
* gem_close:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This wraps the GEM_CLOSE ioctl, which to release a file-private gem buffer
* handle.
*/
void gem_close(int fd, uint32_t handle)
{
struct drm_gem_close close_bo;
igt_assert_neq(handle, 0);
memset(&close_bo, 0, sizeof(close_bo));
close_bo.handle = handle;
do_ioctl(fd, DRM_IOCTL_GEM_CLOSE, &close_bo);
}
int __gem_write(int fd, uint32_t handle, uint64_t offset, const void *buf, uint64_t length)
{
struct drm_i915_gem_pwrite gem_pwrite;
int err;
memset(&gem_pwrite, 0, sizeof(gem_pwrite));
gem_pwrite.handle = handle;
gem_pwrite.offset = offset;
gem_pwrite.size = length;
gem_pwrite.data_ptr = to_user_pointer(buf);
err = 0;
if (drmIoctl(fd, DRM_IOCTL_I915_GEM_PWRITE, &gem_pwrite))
err = -errno;
return err;
}
/**
* gem_write:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @offset: offset within the buffer of the subrange
* @buf: pointer to the data to write into the buffer
* @length: size of the subrange
*
* This wraps the PWRITE ioctl, which is to upload a linear data to a subrange
* of a gem buffer object.
*/
void gem_write(int fd, uint32_t handle, uint64_t offset, const void *buf, uint64_t length)
{
igt_assert_eq(__gem_write(fd, handle, offset, buf, length), 0);
}
static int __gem_read(int fd, uint32_t handle, uint64_t offset, void *buf, uint64_t length)
{
struct drm_i915_gem_pread gem_pread;
int err;
memset(&gem_pread, 0, sizeof(gem_pread));
gem_pread.handle = handle;
gem_pread.offset = offset;
gem_pread.size = length;
gem_pread.data_ptr = to_user_pointer(buf);
err = 0;
if (drmIoctl(fd, DRM_IOCTL_I915_GEM_PREAD, &gem_pread))
err = -errno;
return err;
}
/**
* gem_read:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @offset: offset within the buffer of the subrange
* @buf: pointer to the data to read into
* @length: size of the subrange
*
* This wraps the PREAD ioctl, which is to download a linear data to a subrange
* of a gem buffer object.
*/
void gem_read(int fd, uint32_t handle, uint64_t offset, void *buf, uint64_t length)
{
igt_assert_eq(__gem_read(fd, handle, offset, buf, length), 0);
}
int __gem_set_domain(int fd, uint32_t handle, uint32_t read, uint32_t write)
{
struct drm_i915_gem_set_domain set_domain;
int err;
memset(&set_domain, 0, sizeof(set_domain));
set_domain.handle = handle;
set_domain.read_domains = read;
set_domain.write_domain = write;
err = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_SET_DOMAIN, &set_domain))
err = -errno;
return err;
}
/**
* gem_set_domain:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @read: gem domain bits for read access
* @write: gem domain bit for write access
*
* This wraps the SET_DOMAIN ioctl, which is used to control the coherency of
* the gem buffer object between the cpu and gtt mappings. It is also use to
* synchronize with outstanding rendering in general, but for that use-case
* please have a look at gem_sync().
*/
void gem_set_domain(int fd, uint32_t handle, uint32_t read, uint32_t write)
{
igt_assert_eq(__gem_set_domain(fd, handle, read, write), 0);
}
/**
* __gem_wait:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @timeout_ns: [in] time to wait, [out] remaining time (in nanoseconds)
*
* This functions waits for outstanding rendering to complete, upto
* the timeout_ns. If no timeout_ns is provided, the wait is indefinite and
* only returns upon an error or when the rendering is complete.
*/
int gem_wait(int fd, uint32_t handle, int64_t *timeout_ns)
{
struct drm_i915_gem_wait wait;
int ret;
memset(&wait, 0, sizeof(wait));
wait.bo_handle = handle;
wait.timeout_ns = timeout_ns ? *timeout_ns : -1;
wait.flags = 0;
ret = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_WAIT, &wait))
ret = -errno;
if (timeout_ns)
*timeout_ns = wait.timeout_ns;
return ret;
}
/**
* gem_sync:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This functions waits for outstanding rendering to complete.
*/
void gem_sync(int fd, uint32_t handle)
{
if (gem_wait(fd, handle, NULL))
gem_set_domain(fd, handle,
I915_GEM_DOMAIN_GTT,
I915_GEM_DOMAIN_GTT);
errno = 0;
}
bool gem_create__has_stolen_support(int fd)
{
static int has_stolen_support = -1;
struct drm_i915_getparam gp;
int val = -1;
if (has_stolen_support < 0) {
memset(&gp, 0, sizeof(gp));
gp.param = 38; /* CREATE_VERSION */
gp.value = &val;
/* Do we have the extended gem_create_ioctl? */
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp);
has_stolen_support = val >= 2;
}
return has_stolen_support;
}
struct local_i915_gem_create_v2 {
uint64_t size;
uint32_t handle;
uint32_t pad;
#define I915_CREATE_PLACEMENT_STOLEN (1<<0)
uint32_t flags;
};
#define LOCAL_IOCTL_I915_GEM_CREATE DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_CREATE, struct local_i915_gem_create_v2)
uint32_t __gem_create_stolen(int fd, uint64_t size)
{
struct local_i915_gem_create_v2 create;
int ret;
memset(&create, 0, sizeof(create));
create.handle = 0;
create.size = size;
create.flags = I915_CREATE_PLACEMENT_STOLEN;
ret = igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_CREATE, &create);
if (ret < 0)
return 0;
errno = 0;
return create.handle;
}
/**
* gem_create_stolen:
* @fd: open i915 drm file descriptor
* @size: desired size of the buffer
*
* This wraps the new GEM_CREATE ioctl, which allocates a new gem buffer
* object of @size and placement in stolen memory region.
*
* Returns: The file-private handle of the created buffer object
*/
uint32_t gem_create_stolen(int fd, uint64_t size)
{
struct local_i915_gem_create_v2 create;
memset(&create, 0, sizeof(create));
create.handle = 0;
create.size = size;
create.flags = I915_CREATE_PLACEMENT_STOLEN;
do_ioctl(fd, LOCAL_IOCTL_I915_GEM_CREATE, &create);
igt_assert(create.handle);
return create.handle;
}
int __gem_create(int fd, uint64_t size, uint32_t *handle)
{
struct drm_i915_gem_create create = {
.size = size,
};
int err = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_CREATE, &create) == 0) {
*handle = create.handle;
} else {
err = -errno;
igt_assume(err != 0);
}
errno = 0;
return err;
}
/**
* gem_create:
* @fd: open i915 drm file descriptor
* @size: desired size of the buffer
*
* This wraps the GEM_CREATE ioctl, which allocates a new gem buffer object of
* @size.
*
* Returns: The file-private handle of the created buffer object
*/
uint32_t gem_create(int fd, uint64_t size)
{
uint32_t handle;
igt_assert_eq(__gem_create(fd, size, &handle), 0);
return handle;
}
/**
* __gem_execbuf:
* @fd: open i915 drm file descriptor
* @execbuf: execbuffer data structure
*
* This wraps the EXECBUFFER2 ioctl, which submits a batchbuffer for the gpu to
* run. This is allowed to fail, with -errno returned.
*/
int __gem_execbuf(int fd, struct drm_i915_gem_execbuffer2 *execbuf)
{
int err = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_EXECBUFFER2, execbuf)) {
err = -errno;
igt_assume(err != 0);
}
errno = 0;
return err;
}
/**
* gem_execbuf:
* @fd: open i915 drm file descriptor
* @execbuf: execbuffer data structure
*
* This wraps the EXECBUFFER2 ioctl, which submits a batchbuffer for the gpu to
* run.
*/
void gem_execbuf(int fd, struct drm_i915_gem_execbuffer2 *execbuf)
{
igt_assert_eq(__gem_execbuf(fd, execbuf), 0);
}
/**
* __gem_execbuf_wr:
* @fd: open i915 drm file descriptor
* @execbuf: execbuffer data structure
*
* This wraps the EXECBUFFER2_WR ioctl, which submits a batchbuffer for the gpu to
* run. This is allowed to fail, with -errno returned.
*/
int __gem_execbuf_wr(int fd, struct drm_i915_gem_execbuffer2 *execbuf)
{
int err = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_EXECBUFFER2_WR, execbuf)) {
err = -errno;
igt_assume(err != 0);
}
errno = 0;
return err;
}
/**
* gem_execbuf_wr:
* @fd: open i915 drm file descriptor
* @execbuf: execbuffer data structure
*
* This wraps the EXECBUFFER2_WR ioctl, which submits a batchbuffer for the gpu to
* run.
*/
void gem_execbuf_wr(int fd, struct drm_i915_gem_execbuffer2 *execbuf)
{
igt_assert_eq(__gem_execbuf_wr(fd, execbuf), 0);
}
/**
* gem_madvise:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @state: desired madvise state
*
* This wraps the MADVISE ioctl, which is used in libdrm to implement
* opportunistic buffer object caching. Objects in the cache are set to DONTNEED
* (internally in the kernel tracked as purgeable objects). When such a cached
* object is in need again it must be set back to WILLNEED before first use.
*
* Returns: When setting the madvise state to WILLNEED this returns whether the
* backing storage was still available or not.
*/
int gem_madvise(int fd, uint32_t handle, int state)
{
struct drm_i915_gem_madvise madv;
memset(&madv, 0, sizeof(madv));
madv.handle = handle;
madv.madv = state;
madv.retained = 1;
do_ioctl(fd, DRM_IOCTL_I915_GEM_MADVISE, &madv);
return madv.retained;
}
int __gem_userptr(int fd, void *ptr, uint64_t size, int read_only, uint32_t flags, uint32_t *handle)
{
struct drm_i915_gem_userptr userptr;
memset(&userptr, 0, sizeof(userptr));
userptr.user_ptr = to_user_pointer(ptr);
userptr.user_size = size;
userptr.flags = flags;
if (read_only)
userptr.flags |= I915_USERPTR_READ_ONLY;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_USERPTR, &userptr))
return -errno;
*handle = userptr.handle;
return 0;
}
/**
* gem_userptr:
* @fd: open i915 drm file descriptor
* @ptr: userptr pointer to be passed
* @size: desired size of the buffer
* @read_only: specify whether userptr is opened read only
* @flags: other userptr flags
* @handle: returned handle for the object
*
* Returns userptr handle for the GEM object.
*/
void gem_userptr(int fd, void *ptr, uint64_t size, int read_only, uint32_t flags, uint32_t *handle)
{
igt_assert_eq(__gem_userptr(fd, ptr, size, read_only, flags, handle), 0);
}
/**
* gem_sw_finish:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This wraps the SW_FINISH ioctl, which is used to flush out frontbuffer
* rendering done through the direct cpu memory mappings. Shipping userspace
* does _not_ call this after frontbuffer rendering through gtt memory mappings.
*/
void gem_sw_finish(int fd, uint32_t handle)
{
struct drm_i915_gem_sw_finish finish;
memset(&finish, 0, sizeof(finish));
finish.handle = handle;
do_ioctl(fd, DRM_IOCTL_I915_GEM_SW_FINISH, &finish);
}
/**
* gem_bo_busy:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This wraps the BUSY ioctl, which tells whether a buffer object is still
* actively used by the gpu in a execbuffer.
*
* Returns: The busy state of the buffer object.
*/
bool gem_bo_busy(int fd, uint32_t handle)
{
struct drm_i915_gem_busy busy;
memset(&busy, 0, sizeof(busy));
busy.handle = handle;
do_ioctl(fd, DRM_IOCTL_I915_GEM_BUSY, &busy);
return !!busy.busy;
}
/* feature test helpers */
/**
* gem_gtt_type:
* @fd: open i915 drm file descriptor
*
* Feature test macro to check what type of gtt is being used by the kernel:
* 0 - global gtt
* 1 - aliasing ppgtt
* 2 - full ppgtt
*
* Returns: Type of gtt being used.
*/
static int gem_gtt_type(int fd)
{
struct drm_i915_getparam gp;
int val = 0;
memset(&gp, 0, sizeof(gp));
gp.param = I915_PARAM_HAS_ALIASING_PPGTT;
gp.value = &val;
if (ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp, sizeof(gp)))
return 0;
errno = 0;
return val;
}
/**
* gem_uses_ppgtt:
* @fd: open i915 drm file descriptor
*
* Feature test macro to check whether the kernel internally uses ppgtt to
* execute batches. Note that this is also true when we're using full ppgtt.
*
* Returns: Whether batches are run through ppgtt.
*/
bool gem_uses_ppgtt(int fd)
{
return gem_gtt_type(fd) > 0;
}
/**
* gem_uses_full_ppgtt:
* @fd: open i915 drm file descriptor
*
* Feature test macro to check whether the kernel internally uses full
* per-process gtt to execute batches. Note that this is also true when we're
* using full 64b ppgtt.
*
* Returns: Whether batches are run through full ppgtt.
*/
bool gem_uses_full_ppgtt(int fd)
{
return gem_gtt_type(fd) > 1;
}
/**
* gem_gpu_reset_type:
* @fd: open i915 drm file descriptor
*
* Query whether reset-engine (2), global-reset (1) or reset-disable (0)
* is available.
*
* Returns: GPU reset type available
*/
int gem_gpu_reset_type(int fd)
{
struct drm_i915_getparam gp;
int gpu_reset_type = -1;
memset(&gp, 0, sizeof(gp));
gp.param = I915_PARAM_HAS_GPU_RESET;
gp.value = &gpu_reset_type;
drmIoctl(fd, DRM_IOCTL_I915_GETPARAM, &gp);
return gpu_reset_type;
}
/**
* gem_gpu_reset_enabled:
* @fd: open i915 drm file descriptor
*
* Feature test macro to check whether the kernel internally uses hangchecks
* and can reset the GPU upon hang detection. Note that this is also true when
* reset-engine (the lightweight, single engine reset) is available.
*
* Returns: Whether the driver will detect hangs and perform a reset.
*/
bool gem_gpu_reset_enabled(int fd)
{
return gem_gpu_reset_type(fd) > 0;
}
/**
* gem_engine_reset_enabled:
* @fd: open i915 drm file descriptor
*
* Feature test macro to check whether the kernel internally uses hangchecks
* and can reset individual engines upon hang detection.
*
* Returns: Whether the driver will detect hangs and perform an engine reset.
*/
bool gem_engine_reset_enabled(int fd)
{
return gem_gpu_reset_type(fd) > 1;
}
/**
* gem_available_fences:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the number of available fences
* usable in a batchbuffer. Only relevant for pre-gen4.
*
* Returns: The number of available fences.
*/
int gem_available_fences(int fd)
{
static int num_fences = -1;
if (num_fences < 0) {
struct drm_i915_getparam gp;
memset(&gp, 0, sizeof(gp));
gp.param = I915_PARAM_NUM_FENCES_AVAIL;
gp.value = &num_fences;
num_fences = 0;
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp, sizeof(gp));
errno = 0;
}
return num_fences;
}
bool gem_has_llc(int fd)
{
static int has_llc = -1;
if (has_llc < 0) {
struct drm_i915_getparam gp;
memset(&gp, 0, sizeof(gp));
gp.param = I915_PARAM_HAS_LLC;
gp.value = &has_llc;
has_llc = 0;
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp, sizeof(gp));
errno = 0;
}
return has_llc;
}
static bool has_param(int fd, int param)
{
drm_i915_getparam_t gp;
int tmp = 0;
memset(&gp, 0, sizeof(gp));
gp.value = &tmp;
gp.param = param;
if (igt_ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp))
return false;
errno = 0;
return tmp > 0;
}
/**
* gem_has_bsd:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the BSD ring is available.
*
* Note that recent Bspec calls this the VCS ring for Video Command Submission.
*
* Returns: Whether the BSD ring is available or not.
*/
bool gem_has_bsd(int fd)
{
static int has_bsd = -1;
if (has_bsd < 0)
has_bsd = has_param(fd, I915_PARAM_HAS_BSD);
return has_bsd;
}
/**
* gem_has_blt:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the blitter ring is available.
*
* Note that recent Bspec calls this the BCS ring for Blitter Command Submission.
*
* Returns: Whether the blitter ring is available or not.
*/
bool gem_has_blt(int fd)
{
static int has_blt = -1;
if (has_blt < 0)
has_blt = has_param(fd, I915_PARAM_HAS_BLT);
return has_blt;
}
/**
* gem_has_vebox:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the vebox ring is available.
*
* Note that recent Bspec calls this the VECS ring for Video Enhancement Command
* Submission.
*
* Returns: Whether the vebox ring is available or not.
*/
bool gem_has_vebox(int fd)
{
static int has_vebox = -1;
if (has_vebox < 0)
has_vebox = has_param(fd, I915_PARAM_HAS_VEBOX);
return has_vebox;
}
#define I915_PARAM_HAS_BSD2 31
/**
* gem_has_bsd2:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the BSD2 ring is available.
*
* Note that recent Bspec calls this the VCS ring for Video Command Submission.
*
* Returns: Whether the BSD ring is avaible or not.
*/
bool gem_has_bsd2(int fd)
{
static int has_bsd2 = -1;
if (has_bsd2 < 0)
has_bsd2 = has_param(fd, I915_PARAM_HAS_BSD2);
return has_bsd2;
}
struct local_i915_gem_get_aperture {
__u64 aper_size;
__u64 aper_available_size;
__u64 version;
__u64 map_total_size;
__u64 stolen_total_size;
};
#define DRM_I915_GEM_GET_APERTURE 0x23
#define LOCAL_IOCTL_I915_GEM_GET_APERTURE DRM_IOR (DRM_COMMAND_BASE + DRM_I915_GEM_GET_APERTURE, struct local_i915_gem_get_aperture)
/**
* gem_total_mappable_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the total mappable size.
*
* Returns: Total mappable address space size.
*/
uint64_t gem_total_mappable_size(int fd)
{
struct local_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
do_ioctl(fd, LOCAL_IOCTL_I915_GEM_GET_APERTURE, &aperture);
return aperture.map_total_size;
}
/**
* gem_total_stolen_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the total stolen size.
*
* Returns: Total stolen memory.
*/
uint64_t gem_total_stolen_size(int fd)
{
struct local_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
do_ioctl(fd, LOCAL_IOCTL_I915_GEM_GET_APERTURE, &aperture);
return aperture.stolen_total_size;
}
/**
* gem_available_aperture_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the available gpu aperture size
* usable in a batchbuffer.
*
* Returns: The available gtt address space size.
*/
uint64_t gem_available_aperture_size(int fd)
{
struct drm_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
aperture.aper_size = 256*1024*1024;
do_ioctl(fd, DRM_IOCTL_I915_GEM_GET_APERTURE, &aperture);
return aperture.aper_available_size;
}
/**
* gem_aperture_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the total gpu aperture size.
*
* Returns: The total gtt address space size.
*/
uint64_t gem_aperture_size(int fd)
{
static uint64_t aperture_size = 0;
if (aperture_size == 0) {
struct drm_i915_gem_context_param p;
memset(&p, 0, sizeof(p));
p.param = 0x3;
if (__gem_context_get_param(fd, &p) == 0) {
aperture_size = p.value;
} else {
struct drm_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
aperture.aper_size = 256*1024*1024;
do_ioctl(fd, DRM_IOCTL_I915_GEM_GET_APERTURE, &aperture);
aperture_size = aperture.aper_size;
}
}
return aperture_size;
}
/**
* gem_mappable_aperture_size:
*
* Feature test macro to query the kernel for the mappable gpu aperture size.
* This is the area available for GTT memory mappings.
*
* Returns: The mappable gtt address space size.
*/
uint64_t gem_mappable_aperture_size(void)
{
#if defined(USE_INTEL)
struct pci_device *pci_dev = intel_get_pci_device();
int bar;
if (intel_gen(pci_dev->device_id) < 3)
bar = 0;
else
bar = 2;
return pci_dev->regions[bar].size;
#else
return 0;
#endif
}
/**
* gem_global_aperture_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the global gpu aperture size.
* This is the area available for the kernel to perform address translations.
*
* Returns: The mappable gtt address space size.
*/
uint64_t gem_global_aperture_size(int fd)
{
struct drm_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
aperture.aper_size = 256*1024*1024;
do_ioctl(fd, DRM_IOCTL_I915_GEM_GET_APERTURE, &aperture);
return aperture.aper_size;
}
/**
* gem_has_softpin:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the softpinning functionality is
* supported.
*
* Returns: Whether softpin support is available
*/
bool gem_has_softpin(int fd)
{
static int has_softpin = -1;
if (has_softpin < 0) {
struct drm_i915_getparam gp;
memset(&gp, 0, sizeof(gp));
gp.param = I915_PARAM_HAS_EXEC_SOFTPIN;
gp.value = &has_softpin;
has_softpin = 0;
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp, sizeof(gp));
errno = 0;
}
return has_softpin;
}
/**
* gem_has_exec_fence:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether in/out fence support in execbuffer is
* available.
*
* Returns: Whether fence support is available
*/
bool gem_has_exec_fence(int fd)
{
static int has_exec_fence = -1;
if (has_exec_fence < 0) {
struct drm_i915_getparam gp;
memset(&gp, 0, sizeof(gp));
gp.param = I915_PARAM_HAS_EXEC_FENCE;
gp.value = &has_exec_fence;
has_exec_fence = 0;
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp, sizeof(gp));
errno = 0;
}
return has_exec_fence;
}
/**
* gem_require_caching:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether buffer object caching control is
* available. Automatically skips through igt_require() if not.
*/
void gem_require_caching(int fd)
{
uint32_t handle;
handle = gem_create(fd, 4096);
gem_set_caching(fd, handle, 0);
gem_close(fd, handle);
errno = 0;
}
static void reset_device(int fd)
{
int dir;
dir = igt_debugfs_dir(fd);
igt_require(dir >= 0);
if (ioctl(fd, DRM_IOCTL_I915_GEM_THROTTLE)) {
igt_info("Found wedged device, trying to reset and continue\n");
igt_sysfs_set(dir, "i915_wedged", "-1");
}
igt_sysfs_set(dir, "i915_next_seqno", "1");
close(dir);
}
void igt_require_gem(int fd)
{
char path[256];
int err;
igt_require_intel(fd);
/*
* We only want to use the throttle-ioctl for its -EIO reporting
* of a wedged device, not for actually waiting on outstanding
* requests! So create a new drm_file for the device that is clean.
*/
snprintf(path, sizeof(path), "/proc/self/fd/%d", fd);
fd = open(path, O_RDWR);
igt_assert_lte(0, fd);
/*
* Reset the global seqno at the start of each test. This ensures that
* the test will not wrap unless it explicitly sets up seqno wrapping
* itself, which avoids accidentally hanging when setting up long
* sequences of batches.
*/
reset_device(fd);
err = 0;
if (ioctl(fd, DRM_IOCTL_I915_GEM_THROTTLE))
err = -errno;
close(fd);
igt_require_f(err == 0, "Unresponsive i915/GEM device\n");
}
/**
* gem_require_ring:
* @fd: open i915 drm file descriptor
* @ring: ring flag bit as used in gem_execbuf()
*
* Feature test macro to query whether a specific ring is available.
* This automagically skips if the ring isn't available by
* calling igt_require().
*/
void gem_require_ring(int fd, unsigned ring)
{
igt_require(gem_has_ring(fd, ring));
}
/**
* gem_has_mocs_registers:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the device has MOCS registers.
* These exist gen 9+.
*/
bool gem_has_mocs_registers(int fd)
{
return intel_gen(intel_get_drm_devid(fd)) >= 9;
}
/**
* gem_require_mocs_registers:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the device has MOCS registers.
* These exist gen 9+.
*/
void gem_require_mocs_registers(int fd)
{
igt_require(gem_has_mocs_registers(fd));
}
/* prime */
/**
* prime_handle_to_fd:
* @fd: open i915 drm file descriptor
* @handle: file-private gem buffer object handle
*
* This wraps the PRIME_HANDLE_TO_FD ioctl, which is used to export a gem buffer
* object into a global (i.e. potentially cross-device) dma-buf file-descriptor
* handle.
*
* Returns: The created dma-buf fd handle.
*/
int prime_handle_to_fd(int fd, uint32_t handle)
{
struct drm_prime_handle args;
memset(&args, 0, sizeof(args));
args.handle = handle;
args.flags = DRM_CLOEXEC;
args.fd = -1;
do_ioctl(fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &args);
return args.fd;
}
/**
* prime_handle_to_fd_for_mmap:
* @fd: open i915 drm file descriptor
* @handle: file-private gem buffer object handle
*
* Same as prime_handle_to_fd above but with DRM_RDWR capabilities, which can
* be useful for writing into the mmap'ed dma-buf file-descriptor.
*
* Returns: The created dma-buf fd handle or -1 if the ioctl fails.
*/
int prime_handle_to_fd_for_mmap(int fd, uint32_t handle)
{
struct drm_prime_handle args;
memset(&args, 0, sizeof(args));
args.handle = handle;
args.flags = DRM_CLOEXEC | DRM_RDWR;
args.fd = -1;
if (igt_ioctl(fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &args) != 0)
return -1;
return args.fd;
}
/**
* prime_fd_to_handle:
* @fd: open i915 drm file descriptor
* @dma_buf_fd: dma-buf fd handle
*
* This wraps the PRIME_FD_TO_HANDLE ioctl, which is used to import a dma-buf
* file-descriptor into a gem buffer object.
*
* Returns: The created gem buffer object handle.
*/
uint32_t prime_fd_to_handle(int fd, int dma_buf_fd)
{
struct drm_prime_handle args;
memset(&args, 0, sizeof(args));
args.fd = dma_buf_fd;
args.flags = 0;
args.handle = 0;
do_ioctl(fd, DRM_IOCTL_PRIME_FD_TO_HANDLE, &args);
return args.handle;
}
/**
* prime_get_size:
* @dma_buf_fd: dma-buf fd handle
*
* This wraps the lseek() protocol used to query the invariant size of a
* dma-buf. Not all kernels support this, which is check with igt_require() and
* so will result in automagic test skipping.
*
* Returns: The lifetime-invariant size of the dma-buf object.
*/
off_t prime_get_size(int dma_buf_fd)
{
off_t ret;
ret = lseek(dma_buf_fd, 0, SEEK_END);
igt_assert(ret >= 0 || errno == ESPIPE);
igt_require(ret >= 0);
errno = 0;
return ret;
}
/**
* prime_sync_start
* @dma_buf_fd: dma-buf fd handle
* @write: read/write or read-only access
*
* Must be called before starting CPU mmap access to a dma-buf.
*/
void prime_sync_start(int dma_buf_fd, bool write)
{
struct local_dma_buf_sync sync_start;
memset(&sync_start, 0, sizeof(sync_start));
sync_start.flags = LOCAL_DMA_BUF_SYNC_START;
sync_start.flags |= LOCAL_DMA_BUF_SYNC_READ;
if (write)
sync_start.flags |= LOCAL_DMA_BUF_SYNC_WRITE;
do_ioctl(dma_buf_fd, LOCAL_DMA_BUF_IOCTL_SYNC, &sync_start);
}
/**
* prime_sync_end
* @dma_buf_fd: dma-buf fd handle
* @write: read/write or read-only access
*
* Must be called after finishing CPU mmap access to a dma-buf.
*/
void prime_sync_end(int dma_buf_fd, bool write)
{
struct local_dma_buf_sync sync_end;
memset(&sync_end, 0, sizeof(sync_end));
sync_end.flags = LOCAL_DMA_BUF_SYNC_END;
sync_end.flags |= LOCAL_DMA_BUF_SYNC_READ;
if (write)
sync_end.flags |= LOCAL_DMA_BUF_SYNC_WRITE;
do_ioctl(dma_buf_fd, LOCAL_DMA_BUF_IOCTL_SYNC, &sync_end);
}
bool igt_has_fb_modifiers(int fd)
{
static bool has_modifiers, cap_modifiers_tested;
if (!cap_modifiers_tested) {
uint64_t cap_modifiers;
int ret;
ret = drmGetCap(fd, DRM_CAP_ADDFB2_MODIFIERS, &cap_modifiers);
igt_assert(ret == 0 || errno == EINVAL || errno == EOPNOTSUPP);
has_modifiers = ret == 0 && cap_modifiers == 1;
cap_modifiers_tested = true;
}
return has_modifiers;
}
/**
* igt_require_fb_modifiers:
* @fd: Open DRM file descriptor.
*
* Requires presence of DRM_CAP_ADDFB2_MODIFIERS.
*/
void igt_require_fb_modifiers(int fd)
{
igt_require(igt_has_fb_modifiers(fd));
}
int __kms_addfb(int fd, uint32_t handle,
uint32_t width, uint32_t height,
uint32_t pixel_format, uint64_t modifier,
uint32_t strides[4], uint32_t offsets[4],
int num_planes, uint32_t flags, uint32_t *buf_id)
{
struct drm_mode_fb_cmd2 f;
int ret, i;
if (flags & DRM_MODE_FB_MODIFIERS)
igt_require_fb_modifiers(fd);
memset(&f, 0, sizeof(f));
f.width = width;
f.height = height;
f.pixel_format = pixel_format;
f.flags = flags;
for (i = 0; i < num_planes; i++) {
f.handles[i] = handle;
f.modifier[i] = modifier;
f.pitches[i] = strides[i];
f.offsets[i] = offsets[i];
}
ret = igt_ioctl(fd, DRM_IOCTL_MODE_ADDFB2, &f);
*buf_id = f.fb_id;
return ret < 0 ? -errno : ret;
}