blob: 9a1315d01ecd3403e0c91dfe7ce66469b64e84c2 [file] [log] [blame]
/**************************************************************************
*
* Copyright 2007 VMware, Inc.
* All Rights Reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the
* "Software"), to deal in the Software without restriction, including
* without limitation the rights to use, copy, modify, merge, publish,
* distribute, sub license, and/or sell copies of the Software, and to
* permit persons to whom the Software is furnished to do so, subject to
* the following conditions:
*
* The above copyright notice and this permission notice (including the
* next paragraph) shall be included in all copies or substantial portions
* of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
* OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
* IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*
**************************************************************************/
#ifndef U_INLINES_H
#define U_INLINES_H
#include "pipe/p_context.h"
#include "pipe/p_defines.h"
#include "pipe/p_shader_tokens.h"
#include "pipe/p_state.h"
#include "pipe/p_screen.h"
#include "util/u_debug.h"
#include "util/u_debug_describe.h"
#include "util/u_debug_refcnt.h"
#include "util/u_atomic.h"
#include "util/u_box.h"
#include "util/u_math.h"
#ifdef __cplusplus
extern "C" {
#endif
/*
* Reference counting helper functions.
*/
static inline void
pipe_reference_init(struct pipe_reference *dst, unsigned count)
{
p_atomic_set(&dst->count, count);
}
static inline boolean
pipe_is_referenced(struct pipe_reference *src)
{
return p_atomic_read(&src->count) != 0;
}
/**
* Update reference counting.
* The old thing pointed to, if any, will be unreferenced.
* Both 'dst' and 'src' may be NULL.
* \return TRUE if the object's refcount hits zero and should be destroyed.
*/
static inline boolean
pipe_reference_described(struct pipe_reference *dst,
struct pipe_reference *src,
debug_reference_descriptor get_desc)
{
if (dst != src) {
/* bump the src.count first */
if (src) {
ASSERTED int count = p_atomic_inc_return(&src->count);
assert(count != 1); /* src had to be referenced */
debug_reference(src, get_desc, 1);
}
if (dst) {
int count = p_atomic_dec_return(&dst->count);
assert(count != -1); /* dst had to be referenced */
debug_reference(dst, get_desc, -1);
if (!count)
return true;
}
}
return false;
}
static inline boolean
pipe_reference(struct pipe_reference *dst, struct pipe_reference *src)
{
return pipe_reference_described(dst, src,
(debug_reference_descriptor)
debug_describe_reference);
}
static inline void
pipe_surface_reference(struct pipe_surface **dst, struct pipe_surface *src)
{
struct pipe_surface *old_dst = *dst;
if (pipe_reference_described(old_dst ? &old_dst->reference : NULL,
src ? &src->reference : NULL,
(debug_reference_descriptor)
debug_describe_surface))
old_dst->context->surface_destroy(old_dst->context, old_dst);
*dst = src;
}
/**
* Similar to pipe_surface_reference() but always set the pointer to NULL
* and pass in an explicit context. The explicit context avoids the problem
* of using a deleted context's surface_destroy() method when freeing a surface
* that's shared by multiple contexts.
*/
static inline void
pipe_surface_release(struct pipe_context *pipe, struct pipe_surface **ptr)
{
struct pipe_surface *old = *ptr;
if (pipe_reference_described(&old->reference, NULL,
(debug_reference_descriptor)
debug_describe_surface))
pipe->surface_destroy(pipe, old);
*ptr = NULL;
}
static inline void
pipe_resource_reference(struct pipe_resource **dst, struct pipe_resource *src)
{
struct pipe_resource *old_dst = *dst;
if (pipe_reference_described(old_dst ? &old_dst->reference : NULL,
src ? &src->reference : NULL,
(debug_reference_descriptor)
debug_describe_resource)) {
/* Avoid recursion, which would prevent inlining this function */
do {
struct pipe_resource *next = old_dst->next;
old_dst->screen->resource_destroy(old_dst->screen, old_dst);
old_dst = next;
} while (pipe_reference_described(old_dst ? &old_dst->reference : NULL,
NULL,
(debug_reference_descriptor)
debug_describe_resource));
}
*dst = src;
}
/**
* Same as pipe_surface_release, but used when pipe_context doesn't exist
* anymore.
*/
static inline void
pipe_surface_release_no_context(struct pipe_surface **ptr)
{
struct pipe_surface *surf = *ptr;
if (pipe_reference_described(&surf->reference, NULL,
(debug_reference_descriptor)
debug_describe_surface)) {
/* trivially destroy pipe_surface */
pipe_resource_reference(&surf->texture, NULL);
free(surf);
}
*ptr = NULL;
}
/**
* Set *dst to \p src with proper reference counting.
*
* The caller must guarantee that \p src and *dst were created in
* the same context (if they exist), and that this must be the current context.
*/
static inline void
pipe_sampler_view_reference(struct pipe_sampler_view **dst,
struct pipe_sampler_view *src)
{
struct pipe_sampler_view *old_dst = *dst;
if (pipe_reference_described(old_dst ? &old_dst->reference : NULL,
src ? &src->reference : NULL,
(debug_reference_descriptor)
debug_describe_sampler_view))
old_dst->context->sampler_view_destroy(old_dst->context, old_dst);
*dst = src;
}
static inline void
pipe_so_target_reference(struct pipe_stream_output_target **dst,
struct pipe_stream_output_target *src)
{
struct pipe_stream_output_target *old_dst = *dst;
if (pipe_reference_described(old_dst ? &old_dst->reference : NULL,
src ? &src->reference : NULL,
(debug_reference_descriptor)debug_describe_so_target))
old_dst->context->stream_output_target_destroy(old_dst->context, old_dst);
*dst = src;
}
static inline void
pipe_vertex_buffer_unreference(struct pipe_vertex_buffer *dst)
{
if (dst->is_user_buffer)
dst->buffer.user = NULL;
else
pipe_resource_reference(&dst->buffer.resource, NULL);
}
static inline void
pipe_vertex_buffer_reference(struct pipe_vertex_buffer *dst,
const struct pipe_vertex_buffer *src)
{
pipe_vertex_buffer_unreference(dst);
if (!src->is_user_buffer)
pipe_resource_reference(&dst->buffer.resource, src->buffer.resource);
memcpy(dst, src, sizeof(*src));
}
static inline void
pipe_surface_reset(struct pipe_context *ctx, struct pipe_surface* ps,
struct pipe_resource *pt, unsigned level, unsigned layer)
{
pipe_resource_reference(&ps->texture, pt);
ps->format = pt->format;
ps->width = u_minify(pt->width0, level);
ps->height = u_minify(pt->height0, level);
ps->u.tex.level = level;
ps->u.tex.first_layer = ps->u.tex.last_layer = layer;
ps->context = ctx;
}
static inline void
pipe_surface_init(struct pipe_context *ctx, struct pipe_surface* ps,
struct pipe_resource *pt, unsigned level, unsigned layer)
{
ps->texture = 0;
pipe_reference_init(&ps->reference, 1);
pipe_surface_reset(ctx, ps, pt, level, layer);
}
/* Return true if the surfaces are equal. */
static inline boolean
pipe_surface_equal(struct pipe_surface *s1, struct pipe_surface *s2)
{
return s1->texture == s2->texture &&
s1->format == s2->format &&
(s1->texture->target != PIPE_BUFFER ||
(s1->u.buf.first_element == s2->u.buf.first_element &&
s1->u.buf.last_element == s2->u.buf.last_element)) &&
(s1->texture->target == PIPE_BUFFER ||
(s1->u.tex.level == s2->u.tex.level &&
s1->u.tex.first_layer == s2->u.tex.first_layer &&
s1->u.tex.last_layer == s2->u.tex.last_layer));
}
/*
* Convenience wrappers for screen buffer functions.
*/
/**
* Create a new resource.
* \param bind bitmask of PIPE_BIND_x flags
* \param usage a PIPE_USAGE_x value
*/
static inline struct pipe_resource *
pipe_buffer_create(struct pipe_screen *screen,
unsigned bind,
enum pipe_resource_usage usage,
unsigned size)
{
struct pipe_resource buffer;
memset(&buffer, 0, sizeof buffer);
buffer.target = PIPE_BUFFER;
buffer.format = PIPE_FORMAT_R8_UNORM; /* want TYPELESS or similar */
buffer.bind = bind;
buffer.usage = usage;
buffer.flags = 0;
buffer.width0 = size;
buffer.height0 = 1;
buffer.depth0 = 1;
buffer.array_size = 1;
return screen->resource_create(screen, &buffer);
}
static inline struct pipe_resource *
pipe_buffer_create_const0(struct pipe_screen *screen,
unsigned bind,
enum pipe_resource_usage usage,
unsigned size)
{
struct pipe_resource buffer;
memset(&buffer, 0, sizeof buffer);
buffer.target = PIPE_BUFFER;
buffer.format = PIPE_FORMAT_R8_UNORM;
buffer.bind = bind;
buffer.usage = usage;
buffer.flags = screen->get_param(screen, PIPE_CAP_CONSTBUF0_FLAGS);
buffer.width0 = size;
buffer.height0 = 1;
buffer.depth0 = 1;
buffer.array_size = 1;
return screen->resource_create(screen, &buffer);
}
/**
* Map a range of a resource.
* \param offset start of region, in bytes
* \param length size of region, in bytes
* \param access bitmask of PIPE_TRANSFER_x flags
* \param transfer returns a transfer object
*/
static inline void *
pipe_buffer_map_range(struct pipe_context *pipe,
struct pipe_resource *buffer,
unsigned offset,
unsigned length,
unsigned access,
struct pipe_transfer **transfer)
{
struct pipe_box box;
void *map;
assert(offset < buffer->width0);
assert(offset + length <= buffer->width0);
assert(length);
u_box_1d(offset, length, &box);
map = pipe->transfer_map(pipe, buffer, 0, access, &box, transfer);
if (!map) {
return NULL;
}
return map;
}
/**
* Map whole resource.
* \param access bitmask of PIPE_TRANSFER_x flags
* \param transfer returns a transfer object
*/
static inline void *
pipe_buffer_map(struct pipe_context *pipe,
struct pipe_resource *buffer,
unsigned access,
struct pipe_transfer **transfer)
{
return pipe_buffer_map_range(pipe, buffer, 0, buffer->width0,
access, transfer);
}
static inline void
pipe_buffer_unmap(struct pipe_context *pipe,
struct pipe_transfer *transfer)
{
pipe->transfer_unmap(pipe, transfer);
}
static inline void
pipe_buffer_flush_mapped_range(struct pipe_context *pipe,
struct pipe_transfer *transfer,
unsigned offset,
unsigned length)
{
struct pipe_box box;
int transfer_offset;
assert(length);
assert(transfer->box.x <= (int) offset);
assert((int) (offset + length) <= transfer->box.x + transfer->box.width);
/* Match old screen->buffer_flush_mapped_range() behaviour, where
* offset parameter is relative to the start of the buffer, not the
* mapped range.
*/
transfer_offset = offset - transfer->box.x;
u_box_1d(transfer_offset, length, &box);
pipe->transfer_flush_region(pipe, transfer, &box);
}
static inline void
pipe_buffer_write(struct pipe_context *pipe,
struct pipe_resource *buf,
unsigned offset,
unsigned size,
const void *data)
{
/* Don't set any other usage bits. Drivers should derive them. */
pipe->buffer_subdata(pipe, buf, PIPE_TRANSFER_WRITE, offset, size, data);
}
/**
* Special case for writing non-overlapping ranges.
*
* We can avoid GPU/CPU synchronization when writing range that has never
* been written before.
*/
static inline void
pipe_buffer_write_nooverlap(struct pipe_context *pipe,
struct pipe_resource *buf,
unsigned offset, unsigned size,
const void *data)
{
pipe->buffer_subdata(pipe, buf,
(PIPE_TRANSFER_WRITE |
PIPE_TRANSFER_UNSYNCHRONIZED),
offset, size, data);
}
/**
* Create a new resource and immediately put data into it
* \param bind bitmask of PIPE_BIND_x flags
* \param usage bitmask of PIPE_USAGE_x flags
*/
static inline struct pipe_resource *
pipe_buffer_create_with_data(struct pipe_context *pipe,
unsigned bind,
enum pipe_resource_usage usage,
unsigned size,
const void *ptr)
{
struct pipe_resource *res = pipe_buffer_create(pipe->screen,
bind, usage, size);
pipe_buffer_write_nooverlap(pipe, res, 0, size, ptr);
return res;
}
static inline void
pipe_buffer_read(struct pipe_context *pipe,
struct pipe_resource *buf,
unsigned offset,
unsigned size,
void *data)
{
struct pipe_transfer *src_transfer;
ubyte *map;
map = (ubyte *) pipe_buffer_map_range(pipe,
buf,
offset, size,
PIPE_TRANSFER_READ,
&src_transfer);
if (!map)
return;
memcpy(data, map, size);
pipe_buffer_unmap(pipe, src_transfer);
}
/**
* Map a resource for reading/writing.
* \param access bitmask of PIPE_TRANSFER_x flags
*/
static inline void *
pipe_transfer_map(struct pipe_context *context,
struct pipe_resource *resource,
unsigned level, unsigned layer,
unsigned access,
unsigned x, unsigned y,
unsigned w, unsigned h,
struct pipe_transfer **transfer)
{
struct pipe_box box;
u_box_2d_zslice(x, y, layer, w, h, &box);
return context->transfer_map(context,
resource,
level,
access,
&box, transfer);
}
/**
* Map a 3D (texture) resource for reading/writing.
* \param access bitmask of PIPE_TRANSFER_x flags
*/
static inline void *
pipe_transfer_map_3d(struct pipe_context *context,
struct pipe_resource *resource,
unsigned level,
unsigned access,
unsigned x, unsigned y, unsigned z,
unsigned w, unsigned h, unsigned d,
struct pipe_transfer **transfer)
{
struct pipe_box box;
u_box_3d(x, y, z, w, h, d, &box);
return context->transfer_map(context,
resource,
level,
access,
&box, transfer);
}
static inline void
pipe_transfer_unmap(struct pipe_context *context,
struct pipe_transfer *transfer)
{
context->transfer_unmap(context, transfer);
}
static inline void
pipe_set_constant_buffer(struct pipe_context *pipe,
enum pipe_shader_type shader, uint index,
struct pipe_resource *buf)
{
if (buf) {
struct pipe_constant_buffer cb;
cb.buffer = buf;
cb.buffer_offset = 0;
cb.buffer_size = buf->width0;
cb.user_buffer = NULL;
pipe->set_constant_buffer(pipe, shader, index, &cb);
} else {
pipe->set_constant_buffer(pipe, shader, index, NULL);
}
}
/**
* Get the polygon offset enable/disable flag for the given polygon fill mode.
* \param fill_mode one of PIPE_POLYGON_MODE_POINT/LINE/FILL
*/
static inline boolean
util_get_offset(const struct pipe_rasterizer_state *templ,
unsigned fill_mode)
{
switch(fill_mode) {
case PIPE_POLYGON_MODE_POINT:
return templ->offset_point;
case PIPE_POLYGON_MODE_LINE:
return templ->offset_line;
case PIPE_POLYGON_MODE_FILL:
return templ->offset_tri;
default:
assert(0);
return FALSE;
}
}
static inline float
util_get_min_point_size(const struct pipe_rasterizer_state *state)
{
/* The point size should be clamped to this value at the rasterizer stage.
*/
return !state->point_quad_rasterization &&
!state->point_smooth &&
!state->multisample ? 1.0f : 0.0f;
}
static inline void
util_query_clear_result(union pipe_query_result *result, unsigned type)
{
switch (type) {
case PIPE_QUERY_OCCLUSION_PREDICATE:
case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
case PIPE_QUERY_SO_OVERFLOW_PREDICATE:
case PIPE_QUERY_SO_OVERFLOW_ANY_PREDICATE:
case PIPE_QUERY_GPU_FINISHED:
result->b = FALSE;
break;
case PIPE_QUERY_OCCLUSION_COUNTER:
case PIPE_QUERY_TIMESTAMP:
case PIPE_QUERY_TIME_ELAPSED:
case PIPE_QUERY_PRIMITIVES_GENERATED:
case PIPE_QUERY_PRIMITIVES_EMITTED:
result->u64 = 0;
break;
case PIPE_QUERY_SO_STATISTICS:
memset(&result->so_statistics, 0, sizeof(result->so_statistics));
break;
case PIPE_QUERY_TIMESTAMP_DISJOINT:
memset(&result->timestamp_disjoint, 0, sizeof(result->timestamp_disjoint));
break;
case PIPE_QUERY_PIPELINE_STATISTICS:
memset(&result->pipeline_statistics, 0, sizeof(result->pipeline_statistics));
break;
default:
memset(result, 0, sizeof(*result));
}
}
/** Convert PIPE_TEXTURE_x to TGSI_TEXTURE_x */
static inline enum tgsi_texture_type
util_pipe_tex_to_tgsi_tex(enum pipe_texture_target pipe_tex_target,
unsigned nr_samples)
{
switch (pipe_tex_target) {
case PIPE_BUFFER:
return TGSI_TEXTURE_BUFFER;
case PIPE_TEXTURE_1D:
assert(nr_samples <= 1);
return TGSI_TEXTURE_1D;
case PIPE_TEXTURE_2D:
return nr_samples > 1 ? TGSI_TEXTURE_2D_MSAA : TGSI_TEXTURE_2D;
case PIPE_TEXTURE_RECT:
assert(nr_samples <= 1);
return TGSI_TEXTURE_RECT;
case PIPE_TEXTURE_3D:
assert(nr_samples <= 1);
return TGSI_TEXTURE_3D;
case PIPE_TEXTURE_CUBE:
assert(nr_samples <= 1);
return TGSI_TEXTURE_CUBE;
case PIPE_TEXTURE_1D_ARRAY:
assert(nr_samples <= 1);
return TGSI_TEXTURE_1D_ARRAY;
case PIPE_TEXTURE_2D_ARRAY:
return nr_samples > 1 ? TGSI_TEXTURE_2D_ARRAY_MSAA :
TGSI_TEXTURE_2D_ARRAY;
case PIPE_TEXTURE_CUBE_ARRAY:
return TGSI_TEXTURE_CUBE_ARRAY;
default:
assert(0 && "unexpected texture target");
return TGSI_TEXTURE_UNKNOWN;
}
}
static inline void
util_copy_constant_buffer(struct pipe_constant_buffer *dst,
const struct pipe_constant_buffer *src)
{
if (src) {
pipe_resource_reference(&dst->buffer, src->buffer);
dst->buffer_offset = src->buffer_offset;
dst->buffer_size = src->buffer_size;
dst->user_buffer = src->user_buffer;
}
else {
pipe_resource_reference(&dst->buffer, NULL);
dst->buffer_offset = 0;
dst->buffer_size = 0;
dst->user_buffer = NULL;
}
}
static inline void
util_copy_shader_buffer(struct pipe_shader_buffer *dst,
const struct pipe_shader_buffer *src)
{
if (src) {
pipe_resource_reference(&dst->buffer, src->buffer);
dst->buffer_offset = src->buffer_offset;
dst->buffer_size = src->buffer_size;
}
else {
pipe_resource_reference(&dst->buffer, NULL);
dst->buffer_offset = 0;
dst->buffer_size = 0;
}
}
static inline void
util_copy_image_view(struct pipe_image_view *dst,
const struct pipe_image_view *src)
{
if (src) {
pipe_resource_reference(&dst->resource, src->resource);
dst->format = src->format;
dst->access = src->access;
dst->shader_access = src->shader_access;
dst->u = src->u;
} else {
pipe_resource_reference(&dst->resource, NULL);
dst->format = PIPE_FORMAT_NONE;
dst->access = 0;
dst->shader_access = 0;
memset(&dst->u, 0, sizeof(dst->u));
}
}
static inline unsigned
util_max_layer(const struct pipe_resource *r, unsigned level)
{
switch (r->target) {
case PIPE_TEXTURE_3D:
return u_minify(r->depth0, level) - 1;
case PIPE_TEXTURE_CUBE:
assert(r->array_size == 6);
/* fall-through */
case PIPE_TEXTURE_1D_ARRAY:
case PIPE_TEXTURE_2D_ARRAY:
case PIPE_TEXTURE_CUBE_ARRAY:
return r->array_size - 1;
default:
return 0;
}
}
static inline unsigned
util_num_layers(const struct pipe_resource *r, unsigned level)
{
return util_max_layer(r, level) + 1;
}
static inline bool
util_texrange_covers_whole_level(const struct pipe_resource *tex,
unsigned level, unsigned x, unsigned y,
unsigned z, unsigned width,
unsigned height, unsigned depth)
{
return x == 0 && y == 0 && z == 0 &&
width == u_minify(tex->width0, level) &&
height == u_minify(tex->height0, level) &&
depth == util_num_layers(tex, level);
}
static inline bool
util_logicop_reads_dest(enum pipe_logicop op)
{
switch (op) {
case PIPE_LOGICOP_NOR:
case PIPE_LOGICOP_AND_INVERTED:
case PIPE_LOGICOP_AND_REVERSE:
case PIPE_LOGICOP_INVERT:
case PIPE_LOGICOP_XOR:
case PIPE_LOGICOP_NAND:
case PIPE_LOGICOP_AND:
case PIPE_LOGICOP_EQUIV:
case PIPE_LOGICOP_NOOP:
case PIPE_LOGICOP_OR_INVERTED:
case PIPE_LOGICOP_OR_REVERSE:
case PIPE_LOGICOP_OR:
return true;
case PIPE_LOGICOP_CLEAR:
case PIPE_LOGICOP_COPY_INVERTED:
case PIPE_LOGICOP_COPY:
case PIPE_LOGICOP_SET:
return false;
}
unreachable("bad logicop");
}
static inline struct pipe_context *
pipe_create_multimedia_context(struct pipe_screen *screen)
{
unsigned flags = 0;
if (!screen->get_param(screen, PIPE_CAP_GRAPHICS))
flags |= PIPE_CONTEXT_COMPUTE_ONLY;
return screen->context_create(screen, NULL, flags);
}
static inline unsigned util_res_sample_count(struct pipe_resource *res)
{
return res->nr_samples > 0 ? res->nr_samples : 1;
}
#ifdef __cplusplus
}
#endif
#endif /* U_INLINES_H */