blob: e36a4c2e4139e76d5a85b8fd3259af717e1c0564 [file] [log] [blame]
/*
* Copyright (c) Meta Platforms, Inc. and affiliates.
* All rights reserved.
*
* This source code is licensed under the BSD-style license found in the
* LICENSE file in the root directory of this source tree.
*/
#include <executorch/kernels/portable/cpu/util/matmul_ops_util.h>
#include <executorch/kernels/portable/cpu/vec_ops.h>
#include <executorch/runtime/kernel/kernel_includes.h>
namespace torch {
namespace executor {
namespace native {
using Tensor = exec_aten::Tensor;
Tensor& bmm_out(
RuntimeContext& ctx,
const Tensor& in,
const Tensor& mat2,
Tensor& out) {
ET_KERNEL_CHECK(ctx, check_bmm_args(in, mat2, out), InvalidArgument, out);
size_t output_ndim = 0;
exec_aten::SizesType output_sizes[kTensorDimensionLimit];
get_bmm_out_target_size(in, mat2, output_sizes, &output_ndim);
ET_KERNEL_CHECK(
ctx,
resize_tensor(out, {output_sizes, output_ndim}) == Error::Ok,
InvalidArgument,
out);
ET_SWITCH_REAL_TYPES_AND(
Half, in.scalar_type(), ctx, "bmm.out", CTYPE, [&]() {
const CTYPE* in_data = in.const_data_ptr<CTYPE>();
const CTYPE* mat2_data = mat2.const_data_ptr<CTYPE>();
CTYPE* out_data = out.mutable_data_ptr<CTYPE>();
int64_t batch_size = in.size(0);
int64_t m = in.size(1);
int64_t n = in.size(2);
int64_t p = mat2.size(2);
for (int i = 0; i < batch_size; ++i) {
const CTYPE* in_data_offset = in_data + i * m * n;
const CTYPE* mat2_data_offset = mat2_data + i * n * p;
CTYPE* out_data_offset = out_data + i * m * p;
vec_matmul<CTYPE>(
out_data_offset, in_data_offset, mat2_data_offset, m, n, p);
}
});
return out;
}
} // namespace native
} // namespace executor
} // namespace torch