blob: 7ed10ee559060675a28367d933e2de0179476c38 [file] [log] [blame]
/*
* Copyright 2014 The Chromium OS Authors. All rights reserved.
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#include <assert.h>
#include <errno.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <sys/mman.h>
#include <sys/types.h>
#include <unistd.h>
#include <xf86drm.h>
#include "drv_priv.h"
#include "helpers.h"
#include "util.h"
struct planar_layout {
size_t num_planes;
int horizontal_subsampling[DRV_MAX_PLANES];
int vertical_subsampling[DRV_MAX_PLANES];
int bytes_per_pixel[DRV_MAX_PLANES];
};
// clang-format off
static const struct planar_layout packed_1bpp_layout = {
.num_planes = 1,
.horizontal_subsampling = { 1 },
.vertical_subsampling = { 1 },
.bytes_per_pixel = { 1 }
};
static const struct planar_layout packed_2bpp_layout = {
.num_planes = 1,
.horizontal_subsampling = { 1 },
.vertical_subsampling = { 1 },
.bytes_per_pixel = { 2 }
};
static const struct planar_layout packed_3bpp_layout = {
.num_planes = 1,
.horizontal_subsampling = { 1 },
.vertical_subsampling = { 1 },
.bytes_per_pixel = { 3 }
};
static const struct planar_layout packed_4bpp_layout = {
.num_planes = 1,
.horizontal_subsampling = { 1 },
.vertical_subsampling = { 1 },
.bytes_per_pixel = { 4 }
};
static const struct planar_layout packed_8bpp_layout = {
.num_planes = 1,
.horizontal_subsampling = { 1 },
.vertical_subsampling = { 1 },
.bytes_per_pixel = { 8 }
};
static const struct planar_layout biplanar_yuv_420_layout = {
.num_planes = 2,
.horizontal_subsampling = { 1, 2 },
.vertical_subsampling = { 1, 2 },
.bytes_per_pixel = { 1, 2 }
};
static const struct planar_layout triplanar_yuv_420_layout = {
.num_planes = 3,
.horizontal_subsampling = { 1, 2, 2 },
.vertical_subsampling = { 1, 2, 2 },
.bytes_per_pixel = { 1, 1, 1 }
};
static const struct planar_layout biplanar_yuv_p010_layout = {
.num_planes = 2,
.horizontal_subsampling = { 1, 2 },
.vertical_subsampling = { 1, 2 },
.bytes_per_pixel = { 2, 4 }
};
// clang-format on
static const struct planar_layout *layout_from_format(uint32_t format)
{
switch (format) {
case DRM_FORMAT_BGR233:
case DRM_FORMAT_C8:
case DRM_FORMAT_R8:
case DRM_FORMAT_RGB332:
return &packed_1bpp_layout;
case DRM_FORMAT_R16:
return &packed_2bpp_layout;
case DRM_FORMAT_YVU420:
case DRM_FORMAT_YVU420_ANDROID:
return &triplanar_yuv_420_layout;
case DRM_FORMAT_NV12:
case DRM_FORMAT_NV21:
return &biplanar_yuv_420_layout;
case DRM_FORMAT_P010:
return &biplanar_yuv_p010_layout;
case DRM_FORMAT_ABGR1555:
case DRM_FORMAT_ABGR4444:
case DRM_FORMAT_ARGB1555:
case DRM_FORMAT_ARGB4444:
case DRM_FORMAT_BGR565:
case DRM_FORMAT_BGRA4444:
case DRM_FORMAT_BGRA5551:
case DRM_FORMAT_BGRX4444:
case DRM_FORMAT_BGRX5551:
case DRM_FORMAT_GR88:
case DRM_FORMAT_RG88:
case DRM_FORMAT_RGB565:
case DRM_FORMAT_RGBA4444:
case DRM_FORMAT_RGBA5551:
case DRM_FORMAT_RGBX4444:
case DRM_FORMAT_RGBX5551:
case DRM_FORMAT_UYVY:
case DRM_FORMAT_VYUY:
case DRM_FORMAT_XBGR1555:
case DRM_FORMAT_XBGR4444:
case DRM_FORMAT_XRGB1555:
case DRM_FORMAT_XRGB4444:
case DRM_FORMAT_YUYV:
case DRM_FORMAT_YVYU:
case DRM_FORMAT_MTISP_SXYZW10:
return &packed_2bpp_layout;
case DRM_FORMAT_BGR888:
case DRM_FORMAT_RGB888:
return &packed_3bpp_layout;
case DRM_FORMAT_ABGR2101010:
case DRM_FORMAT_ABGR8888:
case DRM_FORMAT_ARGB2101010:
case DRM_FORMAT_ARGB8888:
case DRM_FORMAT_AYUV:
case DRM_FORMAT_BGRA1010102:
case DRM_FORMAT_BGRA8888:
case DRM_FORMAT_BGRX1010102:
case DRM_FORMAT_BGRX8888:
case DRM_FORMAT_RGBA1010102:
case DRM_FORMAT_RGBA8888:
case DRM_FORMAT_RGBX1010102:
case DRM_FORMAT_RGBX8888:
case DRM_FORMAT_XBGR2101010:
case DRM_FORMAT_XBGR8888:
case DRM_FORMAT_XRGB2101010:
case DRM_FORMAT_XRGB8888:
return &packed_4bpp_layout;
case DRM_FORMAT_ABGR16161616F:
return &packed_8bpp_layout;
default:
drv_log("UNKNOWN FORMAT %d\n", format);
return NULL;
}
}
size_t drv_num_planes_from_format(uint32_t format)
{
const struct planar_layout *layout = layout_from_format(format);
/*
* drv_bo_new calls this function early to query number of planes and
* considers 0 planes to mean unknown format, so we have to support
* that. All other layout_from_format() queries can assume that the
* format is supported and that the return value is non-NULL.
*/
return layout ? layout->num_planes : 0;
}
size_t drv_num_planes_from_modifier(struct driver *drv, uint32_t format, uint64_t modifier)
{
size_t planes = drv_num_planes_from_format(format);
/* Disallow unsupported formats. */
if (!planes)
return 0;
if (drv->backend->num_planes_from_modifier && modifier != DRM_FORMAT_MOD_INVALID &&
modifier != DRM_FORMAT_MOD_LINEAR)
return drv->backend->num_planes_from_modifier(drv, format, modifier);
return planes;
}
uint32_t drv_height_from_format(uint32_t format, uint32_t height, size_t plane)
{
const struct planar_layout *layout = layout_from_format(format);
assert(plane < layout->num_planes);
return DIV_ROUND_UP(height, layout->vertical_subsampling[plane]);
}
uint32_t drv_vertical_subsampling_from_format(uint32_t format, size_t plane)
{
const struct planar_layout *layout = layout_from_format(format);
assert(plane < layout->num_planes);
return layout->vertical_subsampling[plane];
}
uint32_t drv_bytes_per_pixel_from_format(uint32_t format, size_t plane)
{
const struct planar_layout *layout = layout_from_format(format);
assert(plane < layout->num_planes);
return layout->bytes_per_pixel[plane];
}
/*
* This function returns the stride for a given format, width and plane.
*/
uint32_t drv_stride_from_format(uint32_t format, uint32_t width, size_t plane)
{
const struct planar_layout *layout = layout_from_format(format);
assert(plane < layout->num_planes);
uint32_t plane_width = DIV_ROUND_UP(width, layout->horizontal_subsampling[plane]);
uint32_t stride = plane_width * layout->bytes_per_pixel[plane];
/*
* The stride of Android YV12 buffers is required to be aligned to 16 bytes
* (see <system/graphics.h>).
*/
if (format == DRM_FORMAT_YVU420_ANDROID)
stride = (plane == 0) ? ALIGN(stride, 32) : ALIGN(stride, 16);
return stride;
}
uint32_t drv_size_from_format(uint32_t format, uint32_t stride, uint32_t height, size_t plane)
{
return stride * drv_height_from_format(format, height, plane);
}
static uint32_t subsample_stride(uint32_t stride, uint32_t format, size_t plane)
{
if (plane != 0) {
switch (format) {
case DRM_FORMAT_YVU420:
case DRM_FORMAT_YVU420_ANDROID:
stride = DIV_ROUND_UP(stride, 2);
break;
}
}
return stride;
}
/*
* This function fills in the buffer object given the driver aligned stride of
* the first plane, height and a format. This function assumes there is just
* one kernel buffer per buffer object.
*/
int drv_bo_from_format(struct bo *bo, uint32_t stride, uint32_t aligned_height, uint32_t format)
{
uint32_t padding[DRV_MAX_PLANES] = { 0 };
return drv_bo_from_format_and_padding(bo, stride, aligned_height, format, padding);
}
int drv_bo_from_format_and_padding(struct bo *bo, uint32_t stride, uint32_t aligned_height,
uint32_t format, uint32_t padding[DRV_MAX_PLANES])
{
size_t p, num_planes;
uint32_t offset = 0;
num_planes = drv_num_planes_from_format(format);
assert(num_planes);
/*
* HAL_PIXEL_FORMAT_YV12 requires that (see <system/graphics.h>):
* - the aligned height is same as the buffer's height.
* - the chroma stride is 16 bytes aligned, i.e., the luma's strides
* is 32 bytes aligned.
*/
if (format == DRM_FORMAT_YVU420_ANDROID) {
assert(aligned_height == bo->meta.height);
assert(stride == ALIGN(stride, 32));
}
for (p = 0; p < num_planes; p++) {
bo->meta.strides[p] = subsample_stride(stride, format, p);
bo->meta.sizes[p] =
drv_size_from_format(format, bo->meta.strides[p], aligned_height, p) +
padding[p];
bo->meta.offsets[p] = offset;
offset += bo->meta.sizes[p];
}
bo->meta.total_size = offset;
return 0;
}
int drv_dumb_bo_create_ex(struct bo *bo, uint32_t width, uint32_t height, uint32_t format,
uint64_t use_flags, uint64_t quirks)
{
int ret;
size_t plane;
uint32_t aligned_width, aligned_height;
struct drm_mode_create_dumb create_dumb = { 0 };
aligned_width = width;
aligned_height = height;
switch (format) {
case DRM_FORMAT_R16:
/* HAL_PIXEL_FORMAT_Y16 requires that the buffer's width be 16 pixel
* aligned. See hardware/interfaces/graphics/common/1.0/types.hal. */
aligned_width = ALIGN(width, 16);
break;
case DRM_FORMAT_YVU420_ANDROID:
/* HAL_PIXEL_FORMAT_YV12 requires that the buffer's height not
* be aligned. Update 'height' so that drv_bo_from_format below
* uses the non-aligned height. */
height = bo->meta.height;
/* Align width to 32 pixels, so chroma strides are 16 bytes as
* Android requires. */
aligned_width = ALIGN(width, 32);
/* Adjust the height to include room for chroma planes. */
aligned_height = 3 * DIV_ROUND_UP(height, 2);
break;
case DRM_FORMAT_YVU420:
case DRM_FORMAT_NV12:
case DRM_FORMAT_NV21:
/* Adjust the height to include room for chroma planes */
aligned_height = 3 * DIV_ROUND_UP(height, 2);
break;
default:
break;
}
if (quirks & BO_QUIRK_DUMB32BPP) {
aligned_width =
DIV_ROUND_UP(aligned_width * layout_from_format(format)->bytes_per_pixel[0], 4);
create_dumb.bpp = 32;
} else {
create_dumb.bpp = layout_from_format(format)->bytes_per_pixel[0] * 8;
}
create_dumb.width = aligned_width;
create_dumb.height = aligned_height;
create_dumb.flags = 0;
ret = drmIoctl(bo->drv->fd, DRM_IOCTL_MODE_CREATE_DUMB, &create_dumb);
if (ret) {
drv_log("DRM_IOCTL_MODE_CREATE_DUMB failed (%d, %d)\n", bo->drv->fd, errno);
return -errno;
}
drv_bo_from_format(bo, create_dumb.pitch, height, format);
for (plane = 0; plane < bo->meta.num_planes; plane++)
bo->handles[plane].u32 = create_dumb.handle;
bo->meta.total_size = create_dumb.size;
return 0;
}
int drv_dumb_bo_create(struct bo *bo, uint32_t width, uint32_t height, uint32_t format,
uint64_t use_flags)
{
return drv_dumb_bo_create_ex(bo, width, height, format, use_flags, BO_QUIRK_NONE);
}
int drv_dumb_bo_destroy(struct bo *bo)
{
int ret;
struct drm_mode_destroy_dumb destroy_dumb = { 0 };
destroy_dumb.handle = bo->handles[0].u32;
ret = drmIoctl(bo->drv->fd, DRM_IOCTL_MODE_DESTROY_DUMB, &destroy_dumb);
if (ret) {
drv_log("DRM_IOCTL_MODE_DESTROY_DUMB failed (handle=%x)\n", bo->handles[0].u32);
return -errno;
}
return 0;
}
int drv_gem_bo_destroy(struct bo *bo)
{
struct drm_gem_close gem_close;
int ret, error = 0;
size_t plane, i;
for (plane = 0; plane < bo->meta.num_planes; plane++) {
for (i = 0; i < plane; i++)
if (bo->handles[i].u32 == bo->handles[plane].u32)
break;
/* Make sure close hasn't already been called on this handle */
if (i != plane)
continue;
memset(&gem_close, 0, sizeof(gem_close));
gem_close.handle = bo->handles[plane].u32;
ret = drmIoctl(bo->drv->fd, DRM_IOCTL_GEM_CLOSE, &gem_close);
if (ret) {
drv_log("DRM_IOCTL_GEM_CLOSE failed (handle=%x) error %d\n",
bo->handles[plane].u32, ret);
error = -errno;
}
}
return error;
}
int drv_prime_bo_import(struct bo *bo, struct drv_import_fd_data *data)
{
int ret;
size_t plane;
struct drm_prime_handle prime_handle;
for (plane = 0; plane < bo->meta.num_planes; plane++) {
memset(&prime_handle, 0, sizeof(prime_handle));
prime_handle.fd = data->fds[plane];
ret = drmIoctl(bo->drv->fd, DRM_IOCTL_PRIME_FD_TO_HANDLE, &prime_handle);
if (ret) {
drv_log("DRM_IOCTL_PRIME_FD_TO_HANDLE failed (fd=%u)\n", prime_handle.fd);
/*
* Need to call GEM close on planes that were opened,
* if any. Adjust the num_planes variable to be the
* plane that failed, so GEM close will be called on
* planes before that plane.
*/
bo->meta.num_planes = plane;
drv_gem_bo_destroy(bo);
return -errno;
}
bo->handles[plane].u32 = prime_handle.handle;
}
bo->meta.tiling = data->tiling;
return 0;
}
void *drv_dumb_bo_map(struct bo *bo, struct vma *vma, size_t plane, uint32_t map_flags)
{
int ret;
size_t i;
struct drm_mode_map_dumb map_dumb;
memset(&map_dumb, 0, sizeof(map_dumb));
map_dumb.handle = bo->handles[plane].u32;
ret = drmIoctl(bo->drv->fd, DRM_IOCTL_MODE_MAP_DUMB, &map_dumb);
if (ret) {
drv_log("DRM_IOCTL_MODE_MAP_DUMB failed\n");
return MAP_FAILED;
}
for (i = 0; i < bo->meta.num_planes; i++)
if (bo->handles[i].u32 == bo->handles[plane].u32)
vma->length += bo->meta.sizes[i];
return mmap(0, vma->length, drv_get_prot(map_flags), MAP_SHARED, bo->drv->fd,
map_dumb.offset);
}
int drv_bo_munmap(struct bo *bo, struct vma *vma)
{
return munmap(vma->addr, vma->length);
}
int drv_mapping_destroy(struct bo *bo)
{
int ret;
size_t plane;
struct mapping *mapping;
uint32_t idx;
/*
* This function is called right before the buffer is destroyed. It will free any mappings
* associated with the buffer.
*/
idx = 0;
for (plane = 0; plane < bo->meta.num_planes; plane++) {
while (idx < drv_array_size(bo->drv->mappings)) {
mapping = (struct mapping *)drv_array_at_idx(bo->drv->mappings, idx);
if (mapping->vma->handle != bo->handles[plane].u32) {
idx++;
continue;
}
if (!--mapping->vma->refcount) {
ret = bo->drv->backend->bo_unmap(bo, mapping->vma);
if (ret) {
drv_log("munmap failed\n");
return ret;
}
free(mapping->vma);
}
/* This shrinks and shifts the array, so don't increment idx. */
drv_array_remove(bo->drv->mappings, idx);
}
}
return 0;
}
int drv_get_prot(uint32_t map_flags)
{
return (BO_MAP_WRITE & map_flags) ? PROT_WRITE | PROT_READ : PROT_READ;
}
uintptr_t drv_get_reference_count(struct driver *drv, struct bo *bo, size_t plane)
{
void *count;
uintptr_t num = 0;
if (!drmHashLookup(drv->buffer_table, bo->handles[plane].u32, &count))
num = (uintptr_t)(count);
return num;
}
void drv_increment_reference_count(struct driver *drv, struct bo *bo, size_t plane)
{
uintptr_t num = drv_get_reference_count(drv, bo, plane);
/* If a value isn't in the table, drmHashDelete is a no-op */
drmHashDelete(drv->buffer_table, bo->handles[plane].u32);
drmHashInsert(drv->buffer_table, bo->handles[plane].u32, (void *)(num + 1));
}
void drv_decrement_reference_count(struct driver *drv, struct bo *bo, size_t plane)
{
uintptr_t num = drv_get_reference_count(drv, bo, plane);
drmHashDelete(drv->buffer_table, bo->handles[plane].u32);
if (num > 0)
drmHashInsert(drv->buffer_table, bo->handles[plane].u32, (void *)(num - 1));
}
void drv_add_combination(struct driver *drv, const uint32_t format,
struct format_metadata *metadata, uint64_t use_flags)
{
struct combination combo = { .format = format,
.metadata = *metadata,
.use_flags = use_flags };
drv_array_append(drv->combos, &combo);
}
void drv_add_combinations(struct driver *drv, const uint32_t *formats, uint32_t num_formats,
struct format_metadata *metadata, uint64_t use_flags)
{
uint32_t i;
for (i = 0; i < num_formats; i++) {
struct combination combo = { .format = formats[i],
.metadata = *metadata,
.use_flags = use_flags };
drv_array_append(drv->combos, &combo);
}
}
void drv_modify_combination(struct driver *drv, uint32_t format, struct format_metadata *metadata,
uint64_t use_flags)
{
uint32_t i;
struct combination *combo;
/* Attempts to add the specified flags to an existing combination. */
for (i = 0; i < drv_array_size(drv->combos); i++) {
combo = (struct combination *)drv_array_at_idx(drv->combos, i);
if (combo->format == format && combo->metadata.tiling == metadata->tiling &&
combo->metadata.modifier == metadata->modifier)
combo->use_flags |= use_flags;
}
}
int drv_modify_linear_combinations(struct driver *drv)
{
/*
* All current drivers can scanout linear XRGB8888/ARGB8888 as a primary
* plane and as a cursor.
*/
drv_modify_combination(drv, DRM_FORMAT_XRGB8888, &LINEAR_METADATA,
BO_USE_CURSOR | BO_USE_SCANOUT);
drv_modify_combination(drv, DRM_FORMAT_ARGB8888, &LINEAR_METADATA,
BO_USE_CURSOR | BO_USE_SCANOUT);
return 0;
}
/*
* Pick the best modifier from modifiers, according to the ordering
* given by modifier_order.
*/
uint64_t drv_pick_modifier(const uint64_t *modifiers, uint32_t count,
const uint64_t *modifier_order, uint32_t order_count)
{
uint32_t i, j;
for (i = 0; i < order_count; i++) {
for (j = 0; j < count; j++) {
if (modifiers[j] == modifier_order[i]) {
return modifiers[j];
}
}
}
return DRM_FORMAT_MOD_LINEAR;
}
/*
* Search a list of modifiers to see if a given modifier is present
*/
bool drv_has_modifier(const uint64_t *list, uint32_t count, uint64_t modifier)
{
uint32_t i;
for (i = 0; i < count; i++)
if (list[i] == modifier)
return true;
return false;
}