| /* Copyright 2019 The TensorFlow Authors. All Rights Reserved. |
| |
| Licensed under the Apache License, Version 2.0 (the "License"); |
| you may not use this file except in compliance with the License. |
| You may obtain a copy of the License at |
| |
| http://www.apache.org/licenses/LICENSE-2.0 |
| |
| Unless required by applicable law or agreed to in writing, software |
| distributed under the License is distributed on an "AS IS" BASIS, |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| See the License for the specific language governing permissions and |
| limitations under the License. |
| ==============================================================================*/ |
| |
| #if GOOGLE_CUDA |
| #include "absl/memory/memory.h" |
| #include "tensorflow/core/platform/test.h" |
| #include "tensorflow/stream_executor/device_memory.h" |
| #include "tensorflow/stream_executor/multi_platform_manager.h" |
| #include "tensorflow/stream_executor/stream.h" |
| #include "tensorflow/stream_executor/stream_executor.h" |
| |
| namespace stream_executor { |
| |
| TEST(MemcpyTest, PinnedHostMemory) { |
| Platform* platform = |
| MultiPlatformManager::PlatformWithName("CUDA").ValueOrDie(); |
| StreamExecutor* executor = platform->ExecutorForDevice(0).ValueOrDie(); |
| Stream stream(executor); |
| stream.Init(); |
| ASSERT_TRUE(stream.ok()); |
| |
| void* d_ptr = executor->HostMemoryAllocate(sizeof(int)); |
| DeviceMemoryBase d_mem(d_ptr, sizeof(int)); |
| int h_ptr; |
| stream.ThenMemcpy(&h_ptr, d_mem, d_mem.size()); |
| EXPECT_TRUE(stream.BlockHostUntilDone().ok()); |
| } |
| |
| } // namespace stream_executor |
| |
| #endif // GOOGLE_CUDA |