v3dv: implement vkCmdCopyImageToBuffer

For now we only support the TLB path, which limits us to copying
regions that start at offset (0,0). In the future, we will need to add
a fallback path that uses blitting to copy regions with an offset.

Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/6766>
diff --git a/src/broadcom/vulkan/v3dv_meta_copy.c b/src/broadcom/vulkan/v3dv_meta_copy.c
index 98e6f0b..a7728a2 100644
--- a/src/broadcom/vulkan/v3dv_meta_copy.c
+++ b/src/broadcom/vulkan/v3dv_meta_copy.c
@@ -23,6 +23,379 @@
 
 #include "v3dv_private.h"
 
+#include "broadcom/cle/v3dx_pack.h"
+#include "vk_format_info.h"
+
+static void
+emit_image_loads(struct v3dv_cmd_buffer *cmd_buffer,
+                 struct v3dv_cl *cl,
+                 struct v3dv_image *image,
+                 uint32_t layer,
+                 uint32_t mip_level)
+{
+   uint32_t layer_offset = v3dv_layer_offset(image, mip_level, layer);
+
+   const struct util_format_description *desc =
+      vk_format_description(image->vk_format);
+   bool swap_rb = desc->swizzle[0] == PIPE_SWIZZLE_Z &&
+                  image->vk_format != VK_FORMAT_B5G6R5_UNORM_PACK16;
+
+   cl_emit(cl, LOAD_TILE_BUFFER_GENERAL, load) {
+      load.buffer_to_load = RENDER_TARGET_0;
+      load.address = v3dv_cl_address(image->mem->bo, layer_offset);
+
+      load.input_image_format = image->format->rt_type;
+      load.r_b_swap = swap_rb;
+      load.memory_format = image->slices[0].tiling;
+
+      const struct v3d_resource_slice *slice = &image->slices[0];
+      if (slice->tiling == VC5_TILING_UIF_NO_XOR ||
+          slice->tiling == VC5_TILING_UIF_XOR) {
+         load.height_in_ub_or_stride =
+            slice->padded_height_of_output_image_in_uif_blocks;
+      } else if (slice->tiling == VC5_TILING_RASTER) {
+         load.height_in_ub_or_stride = slice->stride;
+      }
+
+      if (image->samples > VK_SAMPLE_COUNT_1_BIT)
+         load.decimate_mode = V3D_DECIMATE_MODE_ALL_SAMPLES;
+      else
+         load.decimate_mode = V3D_DECIMATE_MODE_SAMPLE_0;
+   }
+
+   cl_emit(cl, END_OF_LOADS, end);
+}
+
+static void
+emit_buffer_stores(struct v3dv_cmd_buffer *cmd_buffer,
+                   struct v3dv_cl *cl,
+                   struct v3dv_buffer *buffer,
+                   struct v3dv_image *image,
+                   uint32_t buffer_offset,
+                   uint32_t buffer_stride)
+{
+   cl_emit(cl, STORE_TILE_BUFFER_GENERAL, store) {
+      store.buffer_to_store = RENDER_TARGET_0;
+      store.address = v3dv_cl_address(buffer->mem->bo, buffer_offset);
+      store.clear_buffer_being_stored = false;
+
+      store.output_image_format = image->format->rt_type;
+      store.r_b_swap = false;
+      store.memory_format = VC5_TILING_RASTER;
+      store.height_in_ub_or_stride = buffer_stride;
+
+      if (image->samples > VK_SAMPLE_COUNT_1_BIT)
+         store.decimate_mode = V3D_DECIMATE_MODE_ALL_SAMPLES;
+      else
+         store.decimate_mode = V3D_DECIMATE_MODE_SAMPLE_0;
+   }
+}
+
+static void
+emit_copy_layer_to_buffer_per_tile_list(struct v3dv_cmd_buffer *cmd_buffer,
+                                        struct v3dv_buffer *buffer,
+                                        struct v3dv_image *image,
+                                        uint32_t layer,
+                                        const VkBufferImageCopy *region)
+{
+   struct v3dv_cl *cl = &cmd_buffer->indirect;
+   v3dv_cl_ensure_space(cl, 200, 1);
+   struct v3dv_cl_reloc tile_list_start = v3dv_cl_get_address(cl);
+
+   cl_emit(cl, TILE_COORDINATES_IMPLICIT, coords);
+
+   const VkImageSubresourceLayers *imgrsc = &region->imageSubresource;
+   assert(layer < imgrsc->layerCount);
+
+   /* Load image to TLB */
+   emit_image_loads(cmd_buffer, cl, image,
+                    imgrsc->baseArrayLayer + layer, imgrsc->mipLevel);
+
+   cl_emit(cl, PRIM_LIST_FORMAT, fmt) {
+      fmt.primitive_type = LIST_TRIANGLES;
+   }
+
+   cl_emit(cl, BRANCH_TO_IMPLICIT_TILE_LIST, branch);
+
+   /* Store TLB to buffer */
+   uint32_t width, height;
+   if (region->bufferRowLength == 0 || region->bufferImageHeight == 0) {
+      width = image->extent.width;
+      height = image->extent.height;
+   } else {
+      width = region->bufferRowLength;
+      height = region->bufferImageHeight;
+   }
+
+   uint32_t buffer_stride = width * image->cpp;
+   uint32_t buffer_offset =
+      region->bufferOffset + height * buffer_stride * layer;
+   emit_buffer_stores(cmd_buffer, cl, buffer, image,
+                      buffer_offset, buffer_stride);
+
+   cl_emit(cl, END_OF_TILE_MARKER, end);
+
+   cl_emit(cl, RETURN_FROM_SUB_LIST, ret);
+
+   cl_emit(&cmd_buffer->rcl, START_ADDRESS_OF_GENERIC_TILE_LIST, branch) {
+      branch.start = tile_list_start;
+      branch.end = v3dv_cl_get_address(cl);
+   }
+}
+
+static void
+emit_copy_layer_to_buffer(struct v3dv_cmd_buffer *cmd_buffer,
+                          uint32_t min_x_supertile,
+                          uint32_t min_y_supertile,
+                          uint32_t max_x_supertile,
+                          uint32_t max_y_supertile,
+                          struct v3dv_buffer *buffer,
+                          struct v3dv_image *image,
+                          struct v3dv_framebuffer *framebuffer,
+                          uint32_t layer,
+                          const VkBufferImageCopy *region)
+{
+   struct v3dv_cl *rcl = &cmd_buffer->rcl;
+
+   const uint32_t tile_alloc_offset =
+      64 * layer * framebuffer->draw_tiles_x * framebuffer->draw_tiles_y;
+   cl_emit(rcl, MULTICORE_RENDERING_TILE_LIST_SET_BASE, list) {
+      list.address = v3dv_cl_address(cmd_buffer->tile_alloc, tile_alloc_offset);
+   }
+
+   cl_emit(rcl, MULTICORE_RENDERING_SUPERTILE_CFG, config) {
+      config.number_of_bin_tile_lists = 1;
+      config.total_frame_width_in_tiles = framebuffer->draw_tiles_x;
+      config.total_frame_height_in_tiles = framebuffer->draw_tiles_y;
+
+      config.supertile_width_in_tiles = framebuffer->supertile_width;
+      config.supertile_height_in_tiles = framebuffer->supertile_height;
+
+      config.total_frame_width_in_supertiles =
+         framebuffer->frame_width_in_supertiles;
+      config.total_frame_height_in_supertiles =
+         framebuffer->frame_height_in_supertiles;
+   }
+
+   /* GFXH-1742 workaround */
+   for (int i = 0; i < 2; i++) {
+      cl_emit(rcl, TILE_COORDINATES, coords);
+      cl_emit(rcl, END_OF_LOADS, end);
+      cl_emit(rcl, STORE_TILE_BUFFER_GENERAL, store) {
+         store.buffer_to_store = NONE;
+      }
+      cl_emit(rcl, END_OF_TILE_MARKER, end);
+   }
+
+   cl_emit(rcl, FLUSH_VCD_CACHE, flush);
+
+   emit_copy_layer_to_buffer_per_tile_list(cmd_buffer, buffer, image,
+                                           layer, region);
+
+   for (int y = min_y_supertile; y <= max_y_supertile; y++) {
+      for (int x = min_x_supertile; x <= max_x_supertile; x++) {
+         cl_emit(rcl, SUPERTILE_COORDINATES, coords) {
+            coords.column_number_in_supertiles = x;
+            coords.row_number_in_supertiles = y;
+         }
+      }
+   }
+}
+
+static void
+emit_copy_image_to_buffer_rcl(struct v3dv_cmd_buffer *cmd_buffer,
+                              struct v3dv_buffer *buffer,
+                              struct v3dv_image *image,
+                              struct v3dv_framebuffer *framebuffer,
+                              uint32_t internal_type,
+                              const VkBufferImageCopy *region)
+{
+   const VkImageSubresourceLayers *imgrsc = &region->imageSubresource;
+
+   struct v3dv_cl *rcl = &cmd_buffer->rcl;
+   v3dv_cl_ensure_space_with_branch(rcl, 200 +
+                                    imgrsc->layerCount * 256 *
+                                    cl_packet_length(SUPERTILE_COORDINATES));
+
+   uint32_t level_width = u_minify(image->extent.width, imgrsc->mipLevel);
+   uint32_t level_height = u_minify(image->extent.height, imgrsc->mipLevel);
+   cl_emit(rcl, TILE_RENDERING_MODE_CFG_COMMON, config) {
+      config.early_z_disable = true;
+      config.image_width_pixels = level_width;
+      config.image_height_pixels = level_height;
+      config.number_of_render_targets = 1;
+      config.multisample_mode_4x = false; /* FIXME */
+      config.maximum_bpp_of_all_render_targets = framebuffer->internal_bpp;
+   }
+
+   cl_emit(rcl, TILE_RENDERING_MODE_CFG_COLOR, rt) {
+      rt.render_target_0_internal_bpp = framebuffer->internal_bpp;
+      rt.render_target_0_internal_type = internal_type;
+      rt.render_target_0_clamp = V3D_RENDER_TARGET_CLAMP_NONE;
+   }
+
+   /* We always need to emit this, since it signals the end of the RCL config */
+   cl_emit(rcl, TILE_RENDERING_MODE_CFG_ZS_CLEAR_VALUES, clear) {
+      clear.z_clear_value = 0;
+      clear.stencil_clear_value = 0;
+   };
+
+   cl_emit(rcl, TILE_LIST_INITIAL_BLOCK_SIZE, init) {
+      init.use_auto_chained_tile_lists = true;
+      init.size_of_first_block_in_chained_tile_lists =
+         TILE_ALLOCATION_BLOCK_SIZE_64B;
+   }
+
+   uint32_t supertile_w_in_pixels =
+      framebuffer->tile_width * framebuffer->supertile_width;
+   uint32_t supertile_h_in_pixels =
+      framebuffer->tile_height * framebuffer->supertile_height;
+   const uint32_t min_x_supertile =
+      region->imageOffset.x / supertile_w_in_pixels;
+   const uint32_t min_y_supertile =
+      region->imageOffset.y / supertile_h_in_pixels;
+
+   const uint32_t max_render_x =
+      region->imageOffset.x + region->imageExtent.width - 1;
+   const uint32_t max_render_y =
+      region->imageOffset.y + region->imageExtent.height - 1;
+   const uint32_t max_x_supertile = max_render_x / supertile_w_in_pixels;
+   const uint32_t max_y_supertile = max_render_y / supertile_h_in_pixels;
+
+   for (int layer = 0; layer < imgrsc->layerCount; layer++) {
+      emit_copy_layer_to_buffer(cmd_buffer,
+                                min_x_supertile, min_y_supertile,
+                                max_x_supertile, max_y_supertile,
+                                buffer, image, framebuffer,
+                                layer,
+                                region);
+   }
+
+   cl_emit(rcl, END_OF_RENDERING, end);
+}
+
+static void
+emit_copy_image_to_buffer_bcl(struct v3dv_cmd_buffer *cmd_buffer,
+                              struct v3dv_framebuffer *framebuffer,
+                              const VkBufferImageCopy *region)
+{
+   v3dv_cl_ensure_space_with_branch(&cmd_buffer->bcl, 256);
+
+   cl_emit(&cmd_buffer->bcl, NUMBER_OF_LAYERS, config) {
+      config.number_of_layers = framebuffer->layers;
+   }
+
+   cl_emit(&cmd_buffer->bcl, TILE_BINNING_MODE_CFG, config) {
+      config.width_in_pixels = framebuffer->width;
+      config.height_in_pixels = framebuffer->height;
+      config.number_of_render_targets = 1;
+      config.multisample_mode_4x = false; /* FIXME */
+      config.maximum_bpp_of_all_render_targets = framebuffer->internal_bpp;
+   }
+
+   cl_emit(&cmd_buffer->bcl, FLUSH_VCD_CACHE, bin);
+
+   cl_emit(&cmd_buffer->bcl, OCCLUSION_QUERY_COUNTER, counter);
+
+   cl_emit(&cmd_buffer->bcl, START_TILE_BINNING, bin);
+
+   cl_emit(&cmd_buffer->bcl, CLIP_WINDOW, clip) {
+      clip.clip_window_left_pixel_coordinate = region->imageOffset.x;
+      clip.clip_window_bottom_pixel_coordinate = region->imageOffset.y;
+      clip.clip_window_width_in_pixels = region->imageExtent.width;
+      clip.clip_window_height_in_pixels = region->imageExtent.height;
+   }
+
+   cl_emit(&cmd_buffer->bcl, FLUSH, flush);
+}
+
+/* Sets framebuffer dimensions and computes tile size parameters based on the
+ * maximum internal bpp provided.
+ */
+static void
+setup_framebuffer_params(struct v3dv_framebuffer *fb,
+                         struct v3dv_image *image,
+                         uint32_t layer_count,
+                         uint32_t internal_bpp)
+{
+   fb->width  = image->extent.width;
+   fb->height = image->extent.height;
+   fb->layers = layer_count;
+   fb->internal_bpp = MAX2(RENDER_TARGET_MAXIMUM_32BPP, internal_bpp);
+
+   /* We are only interested in the framebufer description required to compute
+    * the tiling setup parameters below, so we don't need real attachments,
+    * only the framebuffer size and the internal bpp.
+    */
+   fb->attachment_count = 0;
+
+   v3dv_framebuffer_compute_tiling_params(fb);
+}
+
+static inline bool
+can_use_tlb_copy_for_image_region(const VkBufferImageCopy *region)
+{
+   return region->imageOffset.x == 0 && region->imageOffset.y == 0;
+}
+
+/* Implements a copy using the TLB.
+ *
+ * This only works if we are copying from offset (0,0), since a TLB store for
+ * tile (x,y) will be written at the same tile offset into the destination.
+ * When this requirement is not met, we need to use a blit instead.
+ */
+static void
+copy_image_to_buffer_tlb(struct v3dv_cmd_buffer *cmd_buffer,
+                         struct v3dv_buffer *buffer,
+                         struct v3dv_image *image,
+                         const VkBufferImageCopy *region)
+{
+      assert(can_use_tlb_copy_for_image_region(region));
+
+      /* FIXME: pre-compute this at image creation time? */
+      uint32_t internal_type;
+      uint32_t internal_bpp;
+      v3dv_get_internal_type_bpp_for_output_format(image->format->rt_type,
+                                                   &internal_type,
+                                                   &internal_bpp);
+
+      uint32_t num_layers = region->imageSubresource.layerCount;
+      assert(num_layers > 0);
+
+      struct v3dv_framebuffer framebuffer;
+      setup_framebuffer_params(&framebuffer, image, num_layers, internal_bpp);
+
+      /* FIXME: here we assume that we have a valid tile alloc/state setup,
+       *        which is usually the case for copy after render scenarios. The
+       *        code below simply checks and asserts this requirement,
+       *        however, a proper implementation should allocate new tile
+       *        alloc/state if we don't have one (for example if we haven't
+       *        recorded a render pass yet) or the one we have isn't large
+       *        enough. We still need to figure out how we want to handle
+       *        varying tile alloc/state requirements in a command buffer.
+       */
+      uint32_t tile_alloc_size = 64 * num_layers *
+                                 framebuffer.draw_tiles_x *
+                                 framebuffer.draw_tiles_y;
+      tile_alloc_size = align(tile_alloc_size, 4096);
+      tile_alloc_size += 8192;
+      tile_alloc_size += 512 * 1024;
+      assert(cmd_buffer->tile_alloc &&
+             cmd_buffer->tile_alloc->size >= tile_alloc_size);
+
+      const uint32_t tsda_per_tile_size = 256;
+      const uint32_t tile_state_size = num_layers *
+                                       framebuffer.draw_tiles_x *
+                                       framebuffer.draw_tiles_y *
+                                       tsda_per_tile_size;
+      assert(cmd_buffer->tile_state &&
+             cmd_buffer->tile_state->size >= tile_state_size);
+
+      emit_copy_image_to_buffer_bcl(cmd_buffer, &framebuffer, region);
+      emit_copy_image_to_buffer_rcl(cmd_buffer, buffer, image,
+                                    &framebuffer, internal_type, region);
+}
+
 void
 v3dv_CmdCopyImageToBuffer(VkCommandBuffer commandBuffer,
                           VkImage srcImage,
@@ -31,5 +404,13 @@
                           uint32_t regionCount,
                           const VkBufferImageCopy *pRegions)
 {
+   V3DV_FROM_HANDLE(v3dv_cmd_buffer, cmd_buffer, commandBuffer);
+   V3DV_FROM_HANDLE(v3dv_image, image, srcImage);
+   V3DV_FROM_HANDLE(v3dv_buffer, buffer, destBuffer);
+
+   for (uint32_t i = 0; i < regionCount; i++) {
+      if (can_use_tlb_copy_for_image_region(&pRegions[i]))
+         copy_image_to_buffer_tlb(cmd_buffer, buffer, image, &pRegions[i]);
+   }
 }