| // |
| // Copyright © 2017 Arm Ltd. All rights reserved. |
| // SPDX-License-Identifier: MIT |
| // |
| #include "ClWorkloadFactory.hpp" |
| #include "ClBackendId.hpp" |
| |
| #include <Layer.hpp> |
| |
| #include <armnn/Exceptions.hpp> |
| #include <armnn/Utils.hpp> |
| |
| #include <backendsCommon/CpuTensorHandle.hpp> |
| #include <backendsCommon/MakeWorkloadHelper.hpp> |
| #include <backendsCommon/MemCopyWorkload.hpp> |
| |
| #include <cl/ClTensorHandle.hpp> |
| #include <cl/workloads/ClWorkloads.hpp> |
| #include <cl/workloads/ClWorkloadUtils.hpp> |
| |
| #include <arm_compute/core/CL/CLKernelLibrary.h> |
| #include <arm_compute/runtime/CL/CLBufferAllocator.h> |
| #include <arm_compute/runtime/CL/CLScheduler.h> |
| |
| #include <boost/polymorphic_cast.hpp> |
| #include <boost/format.hpp> |
| #include <boost/log/trivial.hpp> |
| |
| namespace armnn |
| { |
| |
| namespace |
| { |
| static const BackendId s_Id{ClBackendId()}; |
| } |
| |
| bool ClWorkloadFactory::IsLayerSupported(const Layer& layer, |
| Optional<DataType> dataType, |
| std::string& outReasonIfUnsupported) |
| { |
| return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported); |
| } |
| |
| const BackendId& ClWorkloadFactory::GetBackendId() const |
| { |
| return s_Id; |
| } |
| |
| template <typename FloatWorkload, typename Uint8Workload, typename QueueDescriptorType, typename... Args> |
| std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor, |
| const WorkloadInfo& info, |
| Args&&... args) |
| { |
| try |
| { |
| return MakeWorkloadHelper<FloatWorkload, Uint8Workload>(descriptor, info, std::forward<Args>(args)...); |
| } |
| catch (const cl::Error& clError) |
| { |
| throw WrapClError(clError, CHECK_LOCATION()); |
| } |
| } |
| |
| template <typename Workload, typename QueueDescriptorType, typename... Args> |
| std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor, |
| const WorkloadInfo& info, |
| Args&&... args) |
| { |
| try |
| { |
| return std::make_unique<Workload>(descriptor, info, std::forward<Args>(args)...); |
| } |
| catch (const cl::Error& clError) |
| { |
| throw WrapClError(clError, CHECK_LOCATION()); |
| } |
| } |
| |
| ClWorkloadFactory::ClWorkloadFactory(const std::shared_ptr<ClMemoryManager>& memoryManager) |
| : m_MemoryManager(memoryManager) |
| { |
| } |
| |
| std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo) const |
| { |
| std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo); |
| tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup()); |
| |
| return tensorHandle; |
| } |
| |
| std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo, |
| DataLayout dataLayout) const |
| { |
| std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo, dataLayout); |
| tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup()); |
| |
| return tensorHandle; |
| } |
| |
| std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateSubTensorHandle(ITensorHandle& parent, |
| TensorShape const& subTensorShape, |
| unsigned int const* subTensorOrigin) const |
| { |
| arm_compute::Coordinates coords; |
| arm_compute::TensorShape shape = armcomputetensorutils::BuildArmComputeTensorShape(subTensorShape); |
| |
| coords.set_num_dimensions(subTensorShape.GetNumDimensions()); |
| for (unsigned int i = 0; i < subTensorShape.GetNumDimensions(); i++) |
| { |
| // Arm compute indexes tensor coords in reverse order. |
| unsigned int revertedIndex = subTensorShape.GetNumDimensions() - i - 1; |
| coords.set(i, boost::numeric_cast<int>(subTensorOrigin[revertedIndex])); |
| } |
| |
| return std::make_unique<ClSubTensorHandle>( |
| boost::polymorphic_downcast<IClTensorHandle*>(&parent), shape, coords); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateInput(const InputQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<CopyMemGenericWorkload, CopyMemGenericWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateOutput(const OutputQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<CopyMemGenericWorkload, CopyMemGenericWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClActivationWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClSoftmaxFloatWorkload, ClSoftmaxUint8Workload>(descriptor, info, |
| m_MemoryManager->GetIntraLayerManager()); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClSplitterWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateMerger(const MergerQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClMergerWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateFullyConnected( |
| const FullyConnectedQueueDescriptor& descriptor, const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClFullyConnectedWorkload, ClFullyConnectedWorkload>(descriptor, info, |
| m_MemoryManager->GetIntraLayerManager()); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClPermuteWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClPooling2dWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClConvolution2dWorkload>(descriptor, info, m_MemoryManager->GetIntraLayerManager()); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDepthwiseConvolution2d( |
| const DepthwiseConvolution2dQueueDescriptor& descriptor, const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClDepthwiseConvolutionWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClNormalizationFloatWorkload, NullWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClAdditionWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateMultiplication( |
| const MultiplicationQueueDescriptor& descriptor, const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClMultiplicationWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateDivision( |
| const DivisionQueueDescriptor& descriptor, const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClDivisionFloatWorkload, NullWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateSubtraction(const SubtractionQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClSubtractionWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateBatchNormalization( |
| const BatchNormalizationQueueDescriptor& descriptor, const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClBatchNormalizationFloatWorkload, NullWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| if (descriptor.m_Inputs.empty() || !descriptor.m_Inputs[0]) |
| { |
| throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemCopy workload"); |
| } |
| |
| return MakeWorkload<CopyMemGenericWorkload, CopyMemGenericWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateResizeBilinear( |
| const ResizeBilinearQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClResizeBilinearFloatWorkload, NullWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFakeQuantization( |
| const FakeQuantizationQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return nullptr; |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClL2NormalizationFloatWorkload, NullWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClConstantWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClReshapeWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return nullptr; |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClFloorFloatWorkload, NullWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClLstmFloatWorkload, NullWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvertFp16ToFp32( |
| const ConvertFp16ToFp32QueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClConvertFp16ToFp32Workload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvertFp32ToFp16( |
| const ConvertFp32ToFp16QueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClConvertFp32ToFp16Workload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMaximum(const MaximumQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMean(const MeanQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return std::make_unique<ClMeanWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePad(const PadQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClPadWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<ClBatchToSpaceNdWorkload>(descriptor, info); |
| } |
| |
| std::unique_ptr<IWorkload> ClWorkloadFactory::CreateStridedSlice(const StridedSliceQueueDescriptor& descriptor, |
| const WorkloadInfo& info) const |
| { |
| return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info); |
| } |
| |
| } // namespace armnn |