| // |
| // Copyright 2012 Francisco Jerez |
| // |
| // Permission is hereby granted, free of charge, to any person obtaining a |
| // copy of this software and associated documentation files (the "Software"), |
| // to deal in the Software without restriction, including without limitation |
| // the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| // and/or sell copies of the Software, and to permit persons to whom the |
| // Software is furnished to do so, subject to the following conditions: |
| // |
| // The above copyright notice and this permission notice shall be included in |
| // all copies or substantial portions of the Software. |
| // |
| // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| // THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR |
| // OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, |
| // ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR |
| // OTHER DEALINGS IN THE SOFTWARE. |
| // |
| |
| #include <cstring> |
| |
| #include "api/util.hpp" |
| #include "core/event.hpp" |
| #include "core/memory.hpp" |
| |
| using namespace clover; |
| |
| namespace { |
| typedef resource::vector vector_t; |
| |
| vector_t |
| vector(const size_t *p) { |
| return range(p, 3); |
| } |
| |
| vector_t |
| pitch(const vector_t ®ion, vector_t pitch) { |
| for (auto x : zip(tail(pitch), |
| map(multiplies(), region, pitch))) { |
| // The spec defines a value of zero as the natural pitch, |
| // i.e. the unaligned size of the previous dimension. |
| if (std::get<0>(x) == 0) |
| std::get<0>(x) = std::get<1>(x); |
| } |
| |
| return pitch; |
| } |
| |
| /// |
| /// Size of a region in bytes. |
| /// |
| size_t |
| size(const vector_t &pitch, const vector_t ®ion) { |
| if (any_of(is_zero(), region)) |
| return 0; |
| else |
| return dot(pitch, region - vector_t{ 0, 1, 1 }); |
| } |
| |
| /// |
| /// Common argument checking shared by memory transfer commands. |
| /// |
| void |
| validate_common(command_queue &q, |
| const ref_vector<event> &deps) { |
| if (any_of([&](const event &ev) { |
| return ev.context() != q.context(); |
| }, deps)) |
| throw error(CL_INVALID_CONTEXT); |
| } |
| |
| /// |
| /// Common error checking for a buffer object argument. |
| /// |
| void |
| validate_object(command_queue &q, buffer &mem, const vector_t &origin, |
| const vector_t &pitch, const vector_t ®ion) { |
| if (mem.context() != q.context()) |
| throw error(CL_INVALID_CONTEXT); |
| |
| // The region must fit within the specified pitch, |
| if (any_of(greater(), map(multiplies(), pitch, region), tail(pitch))) |
| throw error(CL_INVALID_VALUE); |
| |
| // ...and within the specified object. |
| if (dot(pitch, origin) + size(pitch, region) > mem.size()) |
| throw error(CL_INVALID_VALUE); |
| |
| if (any_of(is_zero(), region)) |
| throw error(CL_INVALID_VALUE); |
| } |
| |
| /// |
| /// Common error checking for an image argument. |
| /// |
| void |
| validate_object(command_queue &q, image &img, |
| const vector_t &orig, const vector_t ®ion) { |
| vector_t size = { img.width(), img.height(), img.depth() }; |
| |
| if (!q.device().image_support()) |
| throw error(CL_INVALID_OPERATION); |
| |
| if (img.context() != q.context()) |
| throw error(CL_INVALID_CONTEXT); |
| |
| if (any_of(greater(), orig + region, size)) |
| throw error(CL_INVALID_VALUE); |
| |
| if (any_of(is_zero(), region)) |
| throw error(CL_INVALID_VALUE); |
| } |
| |
| /// |
| /// Common error checking for a host pointer argument. |
| /// |
| void |
| validate_object(command_queue &q, const void *ptr, const vector_t &orig, |
| const vector_t &pitch, const vector_t ®ion) { |
| if (!ptr) |
| throw error(CL_INVALID_VALUE); |
| |
| // The region must fit within the specified pitch. |
| if (any_of(greater(), map(multiplies(), pitch, region), tail(pitch))) |
| throw error(CL_INVALID_VALUE); |
| } |
| |
| /// |
| /// Common argument checking for a copy between two buffer objects. |
| /// |
| void |
| validate_copy(command_queue &q, buffer &dst_mem, |
| const vector_t &dst_orig, const vector_t &dst_pitch, |
| buffer &src_mem, |
| const vector_t &src_orig, const vector_t &src_pitch, |
| const vector_t ®ion) { |
| if (dst_mem == src_mem) { |
| auto dst_offset = dot(dst_pitch, dst_orig); |
| auto src_offset = dot(src_pitch, src_orig); |
| |
| if (interval_overlaps()( |
| dst_offset, dst_offset + size(dst_pitch, region), |
| src_offset, src_offset + size(src_pitch, region))) |
| throw error(CL_MEM_COPY_OVERLAP); |
| } |
| } |
| |
| /// |
| /// Common argument checking for a copy between two image objects. |
| /// |
| void |
| validate_copy(command_queue &q, |
| image &dst_img, const vector_t &dst_orig, |
| image &src_img, const vector_t &src_orig, |
| const vector_t ®ion) { |
| if (dst_img.format() != src_img.format()) |
| throw error(CL_IMAGE_FORMAT_MISMATCH); |
| |
| if (dst_img == src_img) { |
| if (all_of(interval_overlaps(), |
| dst_orig, dst_orig + region, |
| src_orig, src_orig + region)) |
| throw error(CL_MEM_COPY_OVERLAP); |
| } |
| } |
| |
| /// |
| /// Checks that the host access flags of the memory object are |
| /// within the allowed set \a flags. |
| /// |
| void |
| validate_object_access(const memory_obj &mem, const cl_mem_flags flags) { |
| if (mem.flags() & ~flags & |
| (CL_MEM_HOST_READ_ONLY | CL_MEM_HOST_WRITE_ONLY | |
| CL_MEM_HOST_NO_ACCESS)) |
| throw error(CL_INVALID_OPERATION); |
| } |
| |
| /// |
| /// Checks that the mapping flags are correct. |
| /// |
| void |
| validate_map_flags(const memory_obj &mem, const cl_map_flags flags) { |
| if ((flags & (CL_MAP_WRITE | CL_MAP_READ)) && |
| (flags & CL_MAP_WRITE_INVALIDATE_REGION)) |
| throw error(CL_INVALID_VALUE); |
| |
| if (flags & CL_MAP_READ) |
| validate_object_access(mem, CL_MEM_HOST_READ_ONLY); |
| |
| if (flags & (CL_MAP_WRITE | CL_MAP_WRITE_INVALIDATE_REGION)) |
| validate_object_access(mem, CL_MEM_HOST_WRITE_ONLY); |
| } |
| |
| /// |
| /// Class that encapsulates the task of mapping an object of type |
| /// \a T. The return value of get() should be implicitly |
| /// convertible to \a void *. |
| /// |
| template<typename T> |
| struct _map { |
| static mapping |
| get(command_queue &q, T obj, cl_map_flags flags, |
| size_t offset, size_t size) { |
| return { q, obj->resource(q), flags, true, |
| {{ offset }}, {{ size, 1, 1 }} }; |
| } |
| }; |
| |
| template<> |
| struct _map<void *> { |
| static void * |
| get(command_queue &q, void *obj, cl_map_flags flags, |
| size_t offset, size_t size) { |
| return (char *)obj + offset; |
| } |
| }; |
| |
| template<> |
| struct _map<const void *> { |
| static const void * |
| get(command_queue &q, const void *obj, cl_map_flags flags, |
| size_t offset, size_t size) { |
| return (const char *)obj + offset; |
| } |
| }; |
| |
| /// |
| /// Software copy from \a src_obj to \a dst_obj. They can be |
| /// either pointers or memory objects. |
| /// |
| template<typename T, typename S> |
| std::function<void (event &)> |
| soft_copy_op(command_queue &q, |
| T dst_obj, const vector_t &dst_orig, const vector_t &dst_pitch, |
| S src_obj, const vector_t &src_orig, const vector_t &src_pitch, |
| const vector_t ®ion) { |
| return [=, &q](event &) { |
| auto dst = _map<T>::get(q, dst_obj, CL_MAP_WRITE, |
| dot(dst_pitch, dst_orig), |
| size(dst_pitch, region)); |
| auto src = _map<S>::get(q, src_obj, CL_MAP_READ, |
| dot(src_pitch, src_orig), |
| size(src_pitch, region)); |
| vector_t v = {}; |
| |
| for (v[2] = 0; v[2] < region[2]; ++v[2]) { |
| for (v[1] = 0; v[1] < region[1]; ++v[1]) { |
| std::memcpy( |
| static_cast<char *>(dst) + dot(dst_pitch, v), |
| static_cast<const char *>(src) + dot(src_pitch, v), |
| src_pitch[0] * region[0]); |
| } |
| } |
| }; |
| } |
| |
| /// |
| /// Hardware copy from \a src_obj to \a dst_obj. |
| /// |
| template<typename T, typename S> |
| std::function<void (event &)> |
| hard_copy_op(command_queue &q, T dst_obj, const vector_t &dst_orig, |
| S src_obj, const vector_t &src_orig, const vector_t ®ion) { |
| return [=, &q](event &) { |
| dst_obj->resource(q).copy(q, dst_orig, region, |
| src_obj->resource(q), src_orig); |
| }; |
| } |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueReadBuffer(cl_command_queue d_q, cl_mem d_mem, cl_bool blocking, |
| size_t offset, size_t size, void *ptr, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &mem = obj<buffer>(d_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| vector_t region = { size, 1, 1 }; |
| vector_t obj_origin = { offset }; |
| auto obj_pitch = pitch(region, {{ 1 }}); |
| |
| validate_common(q, deps); |
| validate_object(q, ptr, {}, obj_pitch, region); |
| validate_object(q, mem, obj_origin, obj_pitch, region); |
| validate_object_access(mem, CL_MEM_HOST_READ_ONLY); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_READ_BUFFER, deps, |
| soft_copy_op(q, ptr, {}, obj_pitch, |
| &mem, obj_origin, obj_pitch, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueWriteBuffer(cl_command_queue d_q, cl_mem d_mem, cl_bool blocking, |
| size_t offset, size_t size, const void *ptr, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &mem = obj<buffer>(d_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| vector_t region = { size, 1, 1 }; |
| vector_t obj_origin = { offset }; |
| auto obj_pitch = pitch(region, {{ 1 }}); |
| |
| validate_common(q, deps); |
| validate_object(q, mem, obj_origin, obj_pitch, region); |
| validate_object(q, ptr, {}, obj_pitch, region); |
| validate_object_access(mem, CL_MEM_HOST_WRITE_ONLY); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_WRITE_BUFFER, deps, |
| soft_copy_op(q, &mem, obj_origin, obj_pitch, |
| ptr, {}, obj_pitch, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueReadBufferRect(cl_command_queue d_q, cl_mem d_mem, cl_bool blocking, |
| const size_t *p_obj_origin, |
| const size_t *p_host_origin, |
| const size_t *p_region, |
| size_t obj_row_pitch, size_t obj_slice_pitch, |
| size_t host_row_pitch, size_t host_slice_pitch, |
| void *ptr, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &mem = obj<buffer>(d_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| auto region = vector(p_region); |
| auto obj_origin = vector(p_obj_origin); |
| auto obj_pitch = pitch(region, {{ 1, obj_row_pitch, obj_slice_pitch }}); |
| auto host_origin = vector(p_host_origin); |
| auto host_pitch = pitch(region, {{ 1, host_row_pitch, host_slice_pitch }}); |
| |
| validate_common(q, deps); |
| validate_object(q, ptr, host_origin, host_pitch, region); |
| validate_object(q, mem, obj_origin, obj_pitch, region); |
| validate_object_access(mem, CL_MEM_HOST_READ_ONLY); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_READ_BUFFER_RECT, deps, |
| soft_copy_op(q, ptr, host_origin, host_pitch, |
| &mem, obj_origin, obj_pitch, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueWriteBufferRect(cl_command_queue d_q, cl_mem d_mem, cl_bool blocking, |
| const size_t *p_obj_origin, |
| const size_t *p_host_origin, |
| const size_t *p_region, |
| size_t obj_row_pitch, size_t obj_slice_pitch, |
| size_t host_row_pitch, size_t host_slice_pitch, |
| const void *ptr, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &mem = obj<buffer>(d_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| auto region = vector(p_region); |
| auto obj_origin = vector(p_obj_origin); |
| auto obj_pitch = pitch(region, {{ 1, obj_row_pitch, obj_slice_pitch }}); |
| auto host_origin = vector(p_host_origin); |
| auto host_pitch = pitch(region, {{ 1, host_row_pitch, host_slice_pitch }}); |
| |
| validate_common(q, deps); |
| validate_object(q, mem, obj_origin, obj_pitch, region); |
| validate_object(q, ptr, host_origin, host_pitch, region); |
| validate_object_access(mem, CL_MEM_HOST_WRITE_ONLY); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_WRITE_BUFFER_RECT, deps, |
| soft_copy_op(q, &mem, obj_origin, obj_pitch, |
| ptr, host_origin, host_pitch, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueCopyBuffer(cl_command_queue d_q, cl_mem d_src_mem, cl_mem d_dst_mem, |
| size_t src_offset, size_t dst_offset, size_t size, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &src_mem = obj<buffer>(d_src_mem); |
| auto &dst_mem = obj<buffer>(d_dst_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| vector_t region = { size, 1, 1 }; |
| vector_t dst_origin = { dst_offset }; |
| auto dst_pitch = pitch(region, {{ 1 }}); |
| vector_t src_origin = { src_offset }; |
| auto src_pitch = pitch(region, {{ 1 }}); |
| |
| validate_common(q, deps); |
| validate_object(q, dst_mem, dst_origin, dst_pitch, region); |
| validate_object(q, src_mem, src_origin, src_pitch, region); |
| validate_copy(q, dst_mem, dst_origin, dst_pitch, |
| src_mem, src_origin, src_pitch, region); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_COPY_BUFFER, deps, |
| hard_copy_op(q, &dst_mem, dst_origin, |
| &src_mem, src_origin, region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueCopyBufferRect(cl_command_queue d_q, cl_mem d_src_mem, |
| cl_mem d_dst_mem, |
| const size_t *p_src_origin, const size_t *p_dst_origin, |
| const size_t *p_region, |
| size_t src_row_pitch, size_t src_slice_pitch, |
| size_t dst_row_pitch, size_t dst_slice_pitch, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &src_mem = obj<buffer>(d_src_mem); |
| auto &dst_mem = obj<buffer>(d_dst_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| auto region = vector(p_region); |
| auto dst_origin = vector(p_dst_origin); |
| auto dst_pitch = pitch(region, {{ 1, dst_row_pitch, dst_slice_pitch }}); |
| auto src_origin = vector(p_src_origin); |
| auto src_pitch = pitch(region, {{ 1, src_row_pitch, src_slice_pitch }}); |
| |
| validate_common(q, deps); |
| validate_object(q, dst_mem, dst_origin, dst_pitch, region); |
| validate_object(q, src_mem, src_origin, src_pitch, region); |
| validate_copy(q, dst_mem, dst_origin, dst_pitch, |
| src_mem, src_origin, src_pitch, region); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_COPY_BUFFER_RECT, deps, |
| soft_copy_op(q, &dst_mem, dst_origin, dst_pitch, |
| &src_mem, src_origin, src_pitch, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueReadImage(cl_command_queue d_q, cl_mem d_mem, cl_bool blocking, |
| const size_t *p_origin, const size_t *p_region, |
| size_t row_pitch, size_t slice_pitch, void *ptr, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &img = obj<image>(d_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| auto region = vector(p_region); |
| auto dst_pitch = pitch(region, {{ img.pixel_size(), |
| row_pitch, slice_pitch }}); |
| auto src_origin = vector(p_origin); |
| auto src_pitch = pitch(region, {{ img.pixel_size(), |
| img.row_pitch(), img.slice_pitch() }}); |
| |
| validate_common(q, deps); |
| validate_object(q, ptr, {}, dst_pitch, region); |
| validate_object(q, img, src_origin, region); |
| validate_object_access(img, CL_MEM_HOST_READ_ONLY); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_READ_IMAGE, deps, |
| soft_copy_op(q, ptr, {}, dst_pitch, |
| &img, src_origin, src_pitch, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueWriteImage(cl_command_queue d_q, cl_mem d_mem, cl_bool blocking, |
| const size_t *p_origin, const size_t *p_region, |
| size_t row_pitch, size_t slice_pitch, const void *ptr, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &img = obj<image>(d_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| auto region = vector(p_region); |
| auto dst_origin = vector(p_origin); |
| auto dst_pitch = pitch(region, {{ img.pixel_size(), |
| img.row_pitch(), img.slice_pitch() }}); |
| auto src_pitch = pitch(region, {{ img.pixel_size(), |
| row_pitch, slice_pitch }}); |
| |
| validate_common(q, deps); |
| validate_object(q, img, dst_origin, region); |
| validate_object(q, ptr, {}, src_pitch, region); |
| validate_object_access(img, CL_MEM_HOST_WRITE_ONLY); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_WRITE_IMAGE, deps, |
| soft_copy_op(q, &img, dst_origin, dst_pitch, |
| ptr, {}, src_pitch, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueCopyImage(cl_command_queue d_q, cl_mem d_src_mem, cl_mem d_dst_mem, |
| const size_t *p_src_origin, const size_t *p_dst_origin, |
| const size_t *p_region, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &src_img = obj<image>(d_src_mem); |
| auto &dst_img = obj<image>(d_dst_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| auto region = vector(p_region); |
| auto dst_origin = vector(p_dst_origin); |
| auto src_origin = vector(p_src_origin); |
| |
| validate_common(q, deps); |
| validate_object(q, dst_img, dst_origin, region); |
| validate_object(q, src_img, src_origin, region); |
| validate_copy(q, dst_img, dst_origin, src_img, src_origin, region); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_COPY_IMAGE, deps, |
| hard_copy_op(q, &dst_img, dst_origin, |
| &src_img, src_origin, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueCopyImageToBuffer(cl_command_queue d_q, |
| cl_mem d_src_mem, cl_mem d_dst_mem, |
| const size_t *p_src_origin, const size_t *p_region, |
| size_t dst_offset, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &src_img = obj<image>(d_src_mem); |
| auto &dst_mem = obj<buffer>(d_dst_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| auto region = vector(p_region); |
| vector_t dst_origin = { dst_offset }; |
| auto dst_pitch = pitch(region, {{ src_img.pixel_size() }}); |
| auto src_origin = vector(p_src_origin); |
| auto src_pitch = pitch(region, {{ src_img.pixel_size(), |
| src_img.row_pitch(), |
| src_img.slice_pitch() }}); |
| |
| validate_common(q, deps); |
| validate_object(q, dst_mem, dst_origin, dst_pitch, region); |
| validate_object(q, src_img, src_origin, region); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_COPY_IMAGE_TO_BUFFER, deps, |
| soft_copy_op(q, &dst_mem, dst_origin, dst_pitch, |
| &src_img, src_origin, src_pitch, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueCopyBufferToImage(cl_command_queue d_q, |
| cl_mem d_src_mem, cl_mem d_dst_mem, |
| size_t src_offset, |
| const size_t *p_dst_origin, const size_t *p_region, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &src_mem = obj<buffer>(d_src_mem); |
| auto &dst_img = obj<image>(d_dst_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| auto region = vector(p_region); |
| auto dst_origin = vector(p_dst_origin); |
| auto dst_pitch = pitch(region, {{ dst_img.pixel_size(), |
| dst_img.row_pitch(), |
| dst_img.slice_pitch() }}); |
| vector_t src_origin = { src_offset }; |
| auto src_pitch = pitch(region, {{ dst_img.pixel_size() }}); |
| |
| validate_common(q, deps); |
| validate_object(q, dst_img, dst_origin, region); |
| validate_object(q, src_mem, src_origin, src_pitch, region); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_COPY_BUFFER_TO_IMAGE, deps, |
| soft_copy_op(q, &dst_img, dst_origin, dst_pitch, |
| &src_mem, src_origin, src_pitch, |
| region)); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API void * |
| clEnqueueMapBuffer(cl_command_queue d_q, cl_mem d_mem, cl_bool blocking, |
| cl_map_flags flags, size_t offset, size_t size, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev, cl_int *r_errcode) try { |
| auto &q = obj(d_q); |
| auto &mem = obj<buffer>(d_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| vector_t region = { size, 1, 1 }; |
| vector_t obj_origin = { offset }; |
| auto obj_pitch = pitch(region, {{ 1 }}); |
| |
| validate_common(q, deps); |
| validate_object(q, mem, obj_origin, obj_pitch, region); |
| validate_map_flags(mem, flags); |
| |
| void *map = mem.resource(q).add_map(q, flags, blocking, obj_origin, region); |
| |
| ret_object(rd_ev, create<hard_event>(q, CL_COMMAND_MAP_BUFFER, deps)); |
| ret_error(r_errcode, CL_SUCCESS); |
| return map; |
| |
| } catch (error &e) { |
| ret_error(r_errcode, e); |
| return NULL; |
| } |
| |
| CLOVER_API void * |
| clEnqueueMapImage(cl_command_queue d_q, cl_mem d_mem, cl_bool blocking, |
| cl_map_flags flags, |
| const size_t *p_origin, const size_t *p_region, |
| size_t *row_pitch, size_t *slice_pitch, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev, cl_int *r_errcode) try { |
| auto &q = obj(d_q); |
| auto &img = obj<image>(d_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| auto region = vector(p_region); |
| auto origin = vector(p_origin); |
| |
| validate_common(q, deps); |
| validate_object(q, img, origin, region); |
| validate_map_flags(img, flags); |
| |
| void *map = img.resource(q).add_map(q, flags, blocking, origin, region); |
| |
| ret_object(rd_ev, create<hard_event>(q, CL_COMMAND_MAP_IMAGE, deps)); |
| ret_error(r_errcode, CL_SUCCESS); |
| return map; |
| |
| } catch (error &e) { |
| ret_error(r_errcode, e); |
| return NULL; |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueUnmapMemObject(cl_command_queue d_q, cl_mem d_mem, void *ptr, |
| cl_uint num_deps, const cl_event *d_deps, |
| cl_event *rd_ev) try { |
| auto &q = obj(d_q); |
| auto &mem = obj(d_mem); |
| auto deps = objs<wait_list_tag>(d_deps, num_deps); |
| |
| validate_common(q, deps); |
| |
| auto hev = create<hard_event>( |
| q, CL_COMMAND_UNMAP_MEM_OBJECT, deps, |
| [=, &q, &mem](event &) { |
| mem.resource(q).del_map(ptr); |
| }); |
| |
| ret_object(rd_ev, hev); |
| return CL_SUCCESS; |
| |
| } catch (error &e) { |
| return e.get(); |
| } |
| |
| CLOVER_API cl_int |
| clEnqueueMigrateMemObjects(cl_command_queue command_queue, |
| cl_uint num_mem_objects, |
| const cl_mem *mem_objects, |
| cl_mem_migration_flags flags, |
| cl_uint num_events_in_wait_list, |
| const cl_event *event_wait_list, |
| cl_event *event) { |
| CLOVER_NOT_SUPPORTED_UNTIL("1.2"); |
| return CL_INVALID_VALUE; |
| } |