| /* |
| Copyright (C) Intel Corp. 2006. All Rights Reserved. |
| Intel funded Tungsten Graphics (http://www.tungstengraphics.com) to |
| develop this 3D driver. |
| |
| Permission is hereby granted, free of charge, to any person obtaining |
| a copy of this software and associated documentation files (the |
| "Software"), to deal in the Software without restriction, including |
| without limitation the rights to use, copy, modify, merge, publish, |
| distribute, sublicense, and/or sell copies of the Software, and to |
| permit persons to whom the Software is furnished to do so, subject to |
| the following conditions: |
| |
| The above copyright notice and this permission notice (including the |
| next paragraph) shall be included in all copies or substantial |
| portions of the Software. |
| |
| THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
| EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. |
| IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE |
| LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION |
| OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION |
| WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. |
| |
| **********************************************************************/ |
| /* |
| * Authors: |
| * Keith Whitwell <keith@tungstengraphics.com> |
| */ |
| |
| #include "brw_state.h" |
| #include "intel_batchbuffer.h" |
| #include "main/imports.h" |
| #include "glsl/ralloc.h" |
| |
| static void |
| brw_track_state_batch(struct brw_context *brw, |
| enum state_struct_type type, |
| uint32_t offset, |
| int size) |
| { |
| struct intel_batchbuffer *batch = &brw->intel.batch; |
| |
| if (!brw->state_batch_list) { |
| /* Our structs are always aligned to at least 32 bytes, so |
| * our array doesn't need to be any larger |
| */ |
| brw->state_batch_list = ralloc_size(brw, sizeof(*brw->state_batch_list) * |
| batch->bo->size / 32); |
| } |
| |
| brw->state_batch_list[brw->state_batch_count].offset = offset; |
| brw->state_batch_list[brw->state_batch_count].size = size; |
| brw->state_batch_list[brw->state_batch_count].type = type; |
| brw->state_batch_count++; |
| } |
| |
| /** |
| * Convenience function to populate a single drm_intel_aub_annotation data |
| * structure. |
| */ |
| static inline void |
| make_annotation(drm_intel_aub_annotation *annotation, uint32_t type, |
| uint32_t subtype, uint32_t ending_offset) |
| { |
| annotation->type = type; |
| annotation->subtype = subtype; |
| annotation->ending_offset = ending_offset; |
| } |
| |
| /** |
| * Generate a set of aub file annotations for the current batch buffer, and |
| * deliver them to DRM. |
| * |
| * The "used" section of the batch buffer (the portion containing batch |
| * commands) is annotated with AUB_TRACE_TYPE_BATCH. The remainder of the |
| * batch buffer (which contains data structures pointed to by batch commands) |
| * is annotated according to the type of each data structure. |
| */ |
| void |
| brw_annotate_aub(struct intel_context *intel) |
| { |
| struct brw_context *brw = brw_context(&intel->ctx); |
| |
| unsigned annotation_count = 2 * brw->state_batch_count + 1; |
| drm_intel_aub_annotation annotations[annotation_count]; |
| int a = 0; |
| make_annotation(&annotations[a++], AUB_TRACE_TYPE_BATCH, 0, |
| 4*intel->batch.used); |
| for (int i = brw->state_batch_count; i-- > 0; ) { |
| uint32_t type = brw->state_batch_list[i].type; |
| uint32_t start_offset = brw->state_batch_list[i].offset; |
| uint32_t end_offset = start_offset + brw->state_batch_list[i].size; |
| make_annotation(&annotations[a++], AUB_TRACE_TYPE_NOTYPE, 0, |
| start_offset); |
| make_annotation(&annotations[a++], AUB_TRACE_TYPE(type), |
| AUB_TRACE_SUBTYPE(type), end_offset); |
| } |
| assert(a == annotation_count); |
| drm_intel_bufmgr_gem_set_aub_annotations(intel->batch.bo, annotations, |
| annotation_count); |
| } |
| |
| /** |
| * Allocates a block of space in the batchbuffer for indirect state. |
| * |
| * We don't want to allocate separate BOs for every bit of indirect |
| * state in the driver. It means overallocating by a significant |
| * margin (4096 bytes, even if the object is just a 20-byte surface |
| * state), and more buffers to walk and count for aperture size checking. |
| * |
| * However, due to the restrictions inposed by the aperture size |
| * checking performance hacks, we can't have the batch point at a |
| * separate indirect state buffer, because once the batch points at |
| * it, no more relocations can be added to it. So, we sneak these |
| * buffers in at the top of the batchbuffer. |
| */ |
| void * |
| brw_state_batch(struct brw_context *brw, |
| enum state_struct_type type, |
| int size, |
| int alignment, |
| uint32_t *out_offset) |
| { |
| struct intel_batchbuffer *batch = &brw->intel.batch; |
| uint32_t offset; |
| |
| assert(size < batch->bo->size); |
| offset = ROUND_DOWN_TO(batch->state_batch_offset - size, alignment); |
| |
| /* If allocating from the top would wrap below the batchbuffer, or |
| * if the batch's used space (plus the reserved pad) collides with our |
| * space, then flush and try again. |
| */ |
| if (batch->state_batch_offset < size || |
| offset < 4*batch->used + batch->reserved_space) { |
| intel_batchbuffer_flush(&brw->intel); |
| offset = ROUND_DOWN_TO(batch->state_batch_offset - size, alignment); |
| } |
| |
| batch->state_batch_offset = offset; |
| |
| if (unlikely(INTEL_DEBUG & (DEBUG_BATCH | DEBUG_AUB))) |
| brw_track_state_batch(brw, type, offset, size); |
| |
| *out_offset = offset; |
| return batch->map + (offset>>2); |
| } |