| /* Copyright 2019 The TensorFlow Authors. All Rights Reserved. |
| |
| Licensed under the Apache License, Version 2.0 (the "License"); |
| you may not use this file except in compliance with the License. |
| You may obtain a copy of the License at |
| |
| http://www.apache.org/licenses/LICENSE-2.0 |
| |
| Unless required by applicable law or agreed to in writing, software |
| distributed under the License is distributed on an "AS IS" BASIS, |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| See the License for the specific language governing permissions and |
| limitations under the License. |
| ==============================================================================*/ |
| |
| #ifndef TENSORFLOW_LITE_DELEGATES_GPU_GL_COMMAND_QUEUE_H_ |
| #define TENSORFLOW_LITE_DELEGATES_GPU_GL_COMMAND_QUEUE_H_ |
| |
| #include <memory> |
| |
| #include "tensorflow/lite/delegates/gpu/common/status.h" |
| #include "tensorflow/lite/delegates/gpu/common/types.h" |
| #include "tensorflow/lite/delegates/gpu/gl/gl_program.h" |
| #include "tensorflow/lite/delegates/gpu/gl/gpu_info.h" |
| |
| namespace tflite { |
| namespace gpu { |
| namespace gl { |
| |
| // GL programs can be executed directly via dispatch call or using a queue |
| // abstraction similar to one in OpenCL and Vulkan. |
| // CommandQueue executes given programs in order as they come. |
| class CommandQueue { |
| public: |
| virtual ~CommandQueue() = default; |
| |
| // Dispatches a program. It may or may not call glFlush. |
| virtual Status Dispatch(const GlProgram& program, |
| const uint3& workgroups) = 0; |
| |
| // Waits until all programs dispatched prior this call are completed. |
| virtual Status WaitForCompletion() = 0; |
| }; |
| |
| // By default memory barrier is inserted after every dispatch. |
| std::unique_ptr<CommandQueue> NewCommandQueue(const GpuInfo& gpu_info); |
| |
| } // namespace gl |
| } // namespace gpu |
| } // namespace tflite |
| |
| #endif // TENSORFLOW_LITE_DELEGATES_GPU_GL_COMMAND_QUEUE_H_ |