| /* |
| * |
| * Copyright 2018 gRPC authors. |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| * |
| */ |
| |
| #include <grpcpp/channel.h> |
| #include <grpcpp/client_context.h> |
| #include <grpcpp/create_channel.h> |
| #include <grpcpp/generic/generic_stub.h> |
| #include <grpcpp/impl/codegen/proto_utils.h> |
| #include <grpcpp/server.h> |
| #include <grpcpp/server_builder.h> |
| #include <grpcpp/server_context.h> |
| #include <grpcpp/support/client_callback.h> |
| #include <gtest/gtest.h> |
| |
| #include <algorithm> |
| #include <condition_variable> |
| #include <functional> |
| #include <mutex> |
| #include <sstream> |
| #include <thread> |
| |
| #include "absl/memory/memory.h" |
| #include "src/core/lib/gpr/env.h" |
| #include "src/core/lib/iomgr/iomgr.h" |
| #include "src/proto/grpc/testing/echo.grpc.pb.h" |
| #include "test/core/util/port.h" |
| #include "test/core/util/test_config.h" |
| #include "test/cpp/end2end/interceptors_util.h" |
| #include "test/cpp/end2end/test_service_impl.h" |
| #include "test/cpp/util/byte_buffer_proto_helper.h" |
| #include "test/cpp/util/string_ref_helper.h" |
| #include "test/cpp/util/test_credentials_provider.h" |
| |
| // MAYBE_SKIP_TEST is a macro to determine if this particular test configuration |
| // should be skipped based on a decision made at SetUp time. In particular, any |
| // callback tests can only be run if the iomgr can run in the background or if |
| // the transport is in-process. |
| #define MAYBE_SKIP_TEST \ |
| do { \ |
| if (do_not_test_) { \ |
| return; \ |
| } \ |
| } while (0) |
| |
| namespace grpc { |
| namespace testing { |
| namespace { |
| |
| enum class Protocol { INPROC, TCP }; |
| |
| class TestScenario { |
| public: |
| TestScenario(bool serve_callback, Protocol protocol, bool intercept, |
| const std::string& creds_type) |
| : callback_server(serve_callback), |
| protocol(protocol), |
| use_interceptors(intercept), |
| credentials_type(creds_type) {} |
| void Log() const; |
| bool callback_server; |
| Protocol protocol; |
| bool use_interceptors; |
| const std::string credentials_type; |
| }; |
| |
| static std::ostream& operator<<(std::ostream& out, |
| const TestScenario& scenario) { |
| return out << "TestScenario{callback_server=" |
| << (scenario.callback_server ? "true" : "false") << ",protocol=" |
| << (scenario.protocol == Protocol::INPROC ? "INPROC" : "TCP") |
| << ",intercept=" << (scenario.use_interceptors ? "true" : "false") |
| << ",creds=" << scenario.credentials_type << "}"; |
| } |
| |
| void TestScenario::Log() const { |
| std::ostringstream out; |
| out << *this; |
| gpr_log(GPR_DEBUG, "%s", out.str().c_str()); |
| } |
| |
| class ClientCallbackEnd2endTest |
| : public ::testing::TestWithParam<TestScenario> { |
| protected: |
| ClientCallbackEnd2endTest() { GetParam().Log(); } |
| |
| void SetUp() override { |
| ServerBuilder builder; |
| |
| auto server_creds = GetCredentialsProvider()->GetServerCredentials( |
| GetParam().credentials_type); |
| // TODO(vjpai): Support testing of AuthMetadataProcessor |
| |
| if (GetParam().protocol == Protocol::TCP) { |
| picked_port_ = grpc_pick_unused_port_or_die(); |
| server_address_ << "localhost:" << picked_port_; |
| builder.AddListeningPort(server_address_.str(), server_creds); |
| } |
| if (!GetParam().callback_server) { |
| builder.RegisterService(&service_); |
| } else { |
| builder.RegisterService(&callback_service_); |
| } |
| |
| if (GetParam().use_interceptors) { |
| std::vector< |
| std::unique_ptr<experimental::ServerInterceptorFactoryInterface>> |
| creators; |
| // Add 20 dummy server interceptors |
| creators.reserve(20); |
| for (auto i = 0; i < 20; i++) { |
| creators.push_back(absl::make_unique<DummyInterceptorFactory>()); |
| } |
| builder.experimental().SetInterceptorCreators(std::move(creators)); |
| } |
| |
| server_ = builder.BuildAndStart(); |
| is_server_started_ = true; |
| if (GetParam().protocol == Protocol::TCP && |
| !grpc_iomgr_run_in_background()) { |
| do_not_test_ = true; |
| } |
| } |
| |
| void ResetStub() { |
| ChannelArguments args; |
| auto channel_creds = GetCredentialsProvider()->GetChannelCredentials( |
| GetParam().credentials_type, &args); |
| switch (GetParam().protocol) { |
| case Protocol::TCP: |
| if (!GetParam().use_interceptors) { |
| channel_ = ::grpc::CreateCustomChannel(server_address_.str(), |
| channel_creds, args); |
| } else { |
| channel_ = CreateCustomChannelWithInterceptors( |
| server_address_.str(), channel_creds, args, |
| CreateDummyClientInterceptors()); |
| } |
| break; |
| case Protocol::INPROC: |
| if (!GetParam().use_interceptors) { |
| channel_ = server_->InProcessChannel(args); |
| } else { |
| channel_ = server_->experimental().InProcessChannelWithInterceptors( |
| args, CreateDummyClientInterceptors()); |
| } |
| break; |
| default: |
| assert(false); |
| } |
| stub_ = grpc::testing::EchoTestService::NewStub(channel_); |
| generic_stub_ = absl::make_unique<GenericStub>(channel_); |
| DummyInterceptor::Reset(); |
| } |
| |
| void TearDown() override { |
| if (is_server_started_) { |
| // Although we would normally do an explicit shutdown, the server |
| // should also work correctly with just a destructor call. The regular |
| // end2end test uses explicit shutdown, so let this one just do reset. |
| server_.reset(); |
| } |
| if (picked_port_ > 0) { |
| grpc_recycle_unused_port(picked_port_); |
| } |
| } |
| |
| void SendRpcs(int num_rpcs, bool with_binary_metadata) { |
| std::string test_string(""); |
| for (int i = 0; i < num_rpcs; i++) { |
| EchoRequest request; |
| EchoResponse response; |
| ClientContext cli_ctx; |
| |
| test_string += "Hello world. "; |
| request.set_message(test_string); |
| std::string val; |
| if (with_binary_metadata) { |
| request.mutable_param()->set_echo_metadata(true); |
| char bytes[8] = {'\0', '\1', '\2', '\3', |
| '\4', '\5', '\6', static_cast<char>(i)}; |
| val = std::string(bytes, 8); |
| cli_ctx.AddMetadata("custom-bin", val); |
| } |
| |
| cli_ctx.set_compression_algorithm(GRPC_COMPRESS_GZIP); |
| |
| std::mutex mu; |
| std::condition_variable cv; |
| bool done = false; |
| stub_->experimental_async()->Echo( |
| &cli_ctx, &request, &response, |
| [&cli_ctx, &request, &response, &done, &mu, &cv, val, |
| with_binary_metadata](Status s) { |
| GPR_ASSERT(s.ok()); |
| |
| EXPECT_EQ(request.message(), response.message()); |
| if (with_binary_metadata) { |
| EXPECT_EQ( |
| 1u, cli_ctx.GetServerTrailingMetadata().count("custom-bin")); |
| EXPECT_EQ(val, ToString(cli_ctx.GetServerTrailingMetadata() |
| .find("custom-bin") |
| ->second)); |
| } |
| std::lock_guard<std::mutex> l(mu); |
| done = true; |
| cv.notify_one(); |
| }); |
| std::unique_lock<std::mutex> l(mu); |
| while (!done) { |
| cv.wait(l); |
| } |
| } |
| } |
| |
| void SendRpcsGeneric(int num_rpcs, bool maybe_except) { |
| const std::string kMethodName("/grpc.testing.EchoTestService/Echo"); |
| std::string test_string(""); |
| for (int i = 0; i < num_rpcs; i++) { |
| EchoRequest request; |
| std::unique_ptr<ByteBuffer> send_buf; |
| ByteBuffer recv_buf; |
| ClientContext cli_ctx; |
| |
| test_string += "Hello world. "; |
| request.set_message(test_string); |
| send_buf = SerializeToByteBuffer(&request); |
| |
| std::mutex mu; |
| std::condition_variable cv; |
| bool done = false; |
| generic_stub_->experimental().UnaryCall( |
| &cli_ctx, kMethodName, send_buf.get(), &recv_buf, |
| [&request, &recv_buf, &done, &mu, &cv, maybe_except](Status s) { |
| GPR_ASSERT(s.ok()); |
| |
| EchoResponse response; |
| EXPECT_TRUE(ParseFromByteBuffer(&recv_buf, &response)); |
| EXPECT_EQ(request.message(), response.message()); |
| std::lock_guard<std::mutex> l(mu); |
| done = true; |
| cv.notify_one(); |
| #if GRPC_ALLOW_EXCEPTIONS |
| if (maybe_except) { |
| throw - 1; |
| } |
| #else |
| GPR_ASSERT(!maybe_except); |
| #endif |
| }); |
| std::unique_lock<std::mutex> l(mu); |
| while (!done) { |
| cv.wait(l); |
| } |
| } |
| } |
| |
| void SendGenericEchoAsBidi(int num_rpcs, int reuses, bool do_writes_done) { |
| const std::string kMethodName("/grpc.testing.EchoTestService/Echo"); |
| std::string test_string(""); |
| for (int i = 0; i < num_rpcs; i++) { |
| test_string += "Hello world. "; |
| class Client : public grpc::experimental::ClientBidiReactor<ByteBuffer, |
| ByteBuffer> { |
| public: |
| Client(ClientCallbackEnd2endTest* test, const std::string& method_name, |
| const std::string& test_str, int reuses, bool do_writes_done) |
| : reuses_remaining_(reuses), do_writes_done_(do_writes_done) { |
| activate_ = [this, test, method_name, test_str] { |
| if (reuses_remaining_ > 0) { |
| cli_ctx_ = absl::make_unique<ClientContext>(); |
| reuses_remaining_--; |
| test->generic_stub_->experimental().PrepareBidiStreamingCall( |
| cli_ctx_.get(), method_name, this); |
| request_.set_message(test_str); |
| send_buf_ = SerializeToByteBuffer(&request_); |
| StartWrite(send_buf_.get()); |
| StartRead(&recv_buf_); |
| StartCall(); |
| } else { |
| std::unique_lock<std::mutex> l(mu_); |
| done_ = true; |
| cv_.notify_one(); |
| } |
| }; |
| activate_(); |
| } |
| void OnWriteDone(bool /*ok*/) override { |
| if (do_writes_done_) { |
| StartWritesDone(); |
| } |
| } |
| void OnReadDone(bool /*ok*/) override { |
| EchoResponse response; |
| EXPECT_TRUE(ParseFromByteBuffer(&recv_buf_, &response)); |
| EXPECT_EQ(request_.message(), response.message()); |
| }; |
| void OnDone(const Status& s) override { |
| EXPECT_TRUE(s.ok()); |
| activate_(); |
| } |
| void Await() { |
| std::unique_lock<std::mutex> l(mu_); |
| while (!done_) { |
| cv_.wait(l); |
| } |
| } |
| |
| EchoRequest request_; |
| std::unique_ptr<ByteBuffer> send_buf_; |
| ByteBuffer recv_buf_; |
| std::unique_ptr<ClientContext> cli_ctx_; |
| int reuses_remaining_; |
| std::function<void()> activate_; |
| std::mutex mu_; |
| std::condition_variable cv_; |
| bool done_ = false; |
| const bool do_writes_done_; |
| }; |
| |
| Client rpc(this, kMethodName, test_string, reuses, do_writes_done); |
| |
| rpc.Await(); |
| } |
| } |
| bool do_not_test_{false}; |
| bool is_server_started_{false}; |
| int picked_port_{0}; |
| std::shared_ptr<Channel> channel_; |
| std::unique_ptr<grpc::testing::EchoTestService::Stub> stub_; |
| std::unique_ptr<grpc::GenericStub> generic_stub_; |
| TestServiceImpl service_; |
| CallbackTestServiceImpl callback_service_; |
| std::unique_ptr<Server> server_; |
| std::ostringstream server_address_; |
| }; |
| |
| TEST_P(ClientCallbackEnd2endTest, SimpleRpc) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SendRpcs(1, false); |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SimpleRpcExpectedError) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| |
| EchoRequest request; |
| EchoResponse response; |
| ClientContext cli_ctx; |
| ErrorStatus error_status; |
| |
| request.set_message("Hello failure"); |
| error_status.set_code(1); // CANCELLED |
| error_status.set_error_message("cancel error message"); |
| *request.mutable_param()->mutable_expected_error() = error_status; |
| |
| std::mutex mu; |
| std::condition_variable cv; |
| bool done = false; |
| |
| stub_->experimental_async()->Echo( |
| &cli_ctx, &request, &response, |
| [&response, &done, &mu, &cv, &error_status](Status s) { |
| EXPECT_EQ("", response.message()); |
| EXPECT_EQ(error_status.code(), s.error_code()); |
| EXPECT_EQ(error_status.error_message(), s.error_message()); |
| std::lock_guard<std::mutex> l(mu); |
| done = true; |
| cv.notify_one(); |
| }); |
| |
| std::unique_lock<std::mutex> l(mu); |
| while (!done) { |
| cv.wait(l); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SimpleRpcUnderLockNested) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| |
| // The request/response state associated with an RPC and the synchronization |
| // variables needed to notify its completion. |
| struct RpcState { |
| std::mutex mu; |
| std::condition_variable cv; |
| bool done = false; |
| EchoRequest request; |
| EchoResponse response; |
| ClientContext cli_ctx; |
| |
| RpcState() = default; |
| ~RpcState() { |
| // Grab the lock to prevent destruction while another is still holding |
| // lock |
| std::lock_guard<std::mutex> lock(mu); |
| } |
| }; |
| std::vector<RpcState> rpc_state(3); |
| for (size_t i = 0; i < rpc_state.size(); i++) { |
| std::string message = "Hello locked world"; |
| message += std::to_string(i); |
| rpc_state[i].request.set_message(message); |
| } |
| |
| // Grab a lock and then start an RPC whose callback grabs the same lock and |
| // then calls this function to start the next RPC under lock (up to a limit of |
| // the size of the rpc_state vector). |
| std::function<void(int)> nested_call = [this, &nested_call, |
| &rpc_state](int index) { |
| std::lock_guard<std::mutex> l(rpc_state[index].mu); |
| stub_->experimental_async()->Echo( |
| &rpc_state[index].cli_ctx, &rpc_state[index].request, |
| &rpc_state[index].response, |
| [index, &nested_call, &rpc_state](Status s) { |
| std::lock_guard<std::mutex> l1(rpc_state[index].mu); |
| EXPECT_TRUE(s.ok()); |
| rpc_state[index].done = true; |
| rpc_state[index].cv.notify_all(); |
| // Call the next level of nesting if possible |
| if (index + 1 < rpc_state.size()) { |
| nested_call(index + 1); |
| } |
| }); |
| }; |
| |
| nested_call(0); |
| |
| // Wait for completion notifications from all RPCs. Order doesn't matter. |
| for (RpcState& state : rpc_state) { |
| std::unique_lock<std::mutex> l(state.mu); |
| while (!state.done) { |
| state.cv.wait(l); |
| } |
| EXPECT_EQ(state.request.message(), state.response.message()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SimpleRpcUnderLock) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| std::mutex mu; |
| std::condition_variable cv; |
| bool done = false; |
| EchoRequest request; |
| request.set_message("Hello locked world."); |
| EchoResponse response; |
| ClientContext cli_ctx; |
| { |
| std::lock_guard<std::mutex> l(mu); |
| stub_->experimental_async()->Echo( |
| &cli_ctx, &request, &response, |
| [&mu, &cv, &done, &request, &response](Status s) { |
| std::lock_guard<std::mutex> l(mu); |
| EXPECT_TRUE(s.ok()); |
| EXPECT_EQ(request.message(), response.message()); |
| done = true; |
| cv.notify_one(); |
| }); |
| } |
| std::unique_lock<std::mutex> l(mu); |
| while (!done) { |
| cv.wait(l); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SequentialRpcs) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SendRpcs(10, false); |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SendClientInitialMetadata) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SimpleRequest request; |
| SimpleResponse response; |
| ClientContext cli_ctx; |
| |
| cli_ctx.AddMetadata(kCheckClientInitialMetadataKey, |
| kCheckClientInitialMetadataVal); |
| |
| std::mutex mu; |
| std::condition_variable cv; |
| bool done = false; |
| stub_->experimental_async()->CheckClientInitialMetadata( |
| &cli_ctx, &request, &response, [&done, &mu, &cv](Status s) { |
| GPR_ASSERT(s.ok()); |
| |
| std::lock_guard<std::mutex> l(mu); |
| done = true; |
| cv.notify_one(); |
| }); |
| std::unique_lock<std::mutex> l(mu); |
| while (!done) { |
| cv.wait(l); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SimpleRpcWithBinaryMetadata) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SendRpcs(1, true); |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SequentialRpcsWithVariedBinaryMetadataValue) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SendRpcs(10, true); |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcs) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SendRpcsGeneric(10, false); |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsAsBidi) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SendGenericEchoAsBidi(10, 1, /*do_writes_done=*/true); |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsAsBidiWithReactorReuse) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SendGenericEchoAsBidi(10, 10, /*do_writes_done=*/true); |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, GenericRpcNoWritesDone) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SendGenericEchoAsBidi(1, 1, /*do_writes_done=*/false); |
| } |
| |
| #if GRPC_ALLOW_EXCEPTIONS |
| TEST_P(ClientCallbackEnd2endTest, ExceptingRpc) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| SendRpcsGeneric(10, true); |
| } |
| #endif |
| |
| TEST_P(ClientCallbackEnd2endTest, MultipleRpcsWithVariedBinaryMetadataValue) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| std::vector<std::thread> threads; |
| threads.reserve(10); |
| for (int i = 0; i < 10; ++i) { |
| threads.emplace_back([this] { SendRpcs(10, true); }); |
| } |
| for (int i = 0; i < 10; ++i) { |
| threads[i].join(); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, MultipleRpcs) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| std::vector<std::thread> threads; |
| threads.reserve(10); |
| for (int i = 0; i < 10; ++i) { |
| threads.emplace_back([this] { SendRpcs(10, false); }); |
| } |
| for (int i = 0; i < 10; ++i) { |
| threads[i].join(); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, CancelRpcBeforeStart) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| EchoRequest request; |
| EchoResponse response; |
| ClientContext context; |
| request.set_message("hello"); |
| context.TryCancel(); |
| |
| std::mutex mu; |
| std::condition_variable cv; |
| bool done = false; |
| stub_->experimental_async()->Echo( |
| &context, &request, &response, [&response, &done, &mu, &cv](Status s) { |
| EXPECT_EQ("", response.message()); |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| std::lock_guard<std::mutex> l(mu); |
| done = true; |
| cv.notify_one(); |
| }); |
| std::unique_lock<std::mutex> l(mu); |
| while (!done) { |
| cv.wait(l); |
| } |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, RequestEchoServerCancel) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| EchoRequest request; |
| EchoResponse response; |
| ClientContext context; |
| request.set_message("hello"); |
| context.AddMetadata(kServerTryCancelRequest, |
| std::to_string(CANCEL_BEFORE_PROCESSING)); |
| |
| std::mutex mu; |
| std::condition_variable cv; |
| bool done = false; |
| stub_->experimental_async()->Echo( |
| &context, &request, &response, [&done, &mu, &cv](Status s) { |
| EXPECT_FALSE(s.ok()); |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| std::lock_guard<std::mutex> l(mu); |
| done = true; |
| cv.notify_one(); |
| }); |
| std::unique_lock<std::mutex> l(mu); |
| while (!done) { |
| cv.wait(l); |
| } |
| } |
| |
| struct ClientCancelInfo { |
| bool cancel{false}; |
| int ops_before_cancel; |
| |
| ClientCancelInfo() : cancel{false} {} |
| explicit ClientCancelInfo(int ops) : cancel{true}, ops_before_cancel{ops} {} |
| }; |
| |
| class WriteClient : public grpc::experimental::ClientWriteReactor<EchoRequest> { |
| public: |
| WriteClient(grpc::testing::EchoTestService::Stub* stub, |
| ServerTryCancelRequestPhase server_try_cancel, |
| int num_msgs_to_send, ClientCancelInfo client_cancel = {}) |
| : server_try_cancel_(server_try_cancel), |
| num_msgs_to_send_(num_msgs_to_send), |
| client_cancel_{client_cancel} { |
| std::string msg{"Hello server."}; |
| for (int i = 0; i < num_msgs_to_send; i++) { |
| desired_ += msg; |
| } |
| if (server_try_cancel != DO_NOT_CANCEL) { |
| // Send server_try_cancel value in the client metadata |
| context_.AddMetadata(kServerTryCancelRequest, |
| std::to_string(server_try_cancel)); |
| } |
| context_.set_initial_metadata_corked(true); |
| stub->experimental_async()->RequestStream(&context_, &response_, this); |
| StartCall(); |
| request_.set_message(msg); |
| MaybeWrite(); |
| } |
| void OnWriteDone(bool ok) override { |
| if (ok) { |
| num_msgs_sent_++; |
| MaybeWrite(); |
| } |
| } |
| void OnDone(const Status& s) override { |
| gpr_log(GPR_INFO, "Sent %d messages", num_msgs_sent_); |
| int num_to_send = |
| (client_cancel_.cancel) |
| ? std::min(num_msgs_to_send_, client_cancel_.ops_before_cancel) |
| : num_msgs_to_send_; |
| switch (server_try_cancel_) { |
| case CANCEL_BEFORE_PROCESSING: |
| case CANCEL_DURING_PROCESSING: |
| // If the RPC is canceled by server before / during messages from the |
| // client, it means that the client most likely did not get a chance to |
| // send all the messages it wanted to send. i.e num_msgs_sent <= |
| // num_msgs_to_send |
| EXPECT_LE(num_msgs_sent_, num_to_send); |
| break; |
| case DO_NOT_CANCEL: |
| case CANCEL_AFTER_PROCESSING: |
| // If the RPC was not canceled or canceled after all messages were read |
| // by the server, the client did get a chance to send all its messages |
| EXPECT_EQ(num_msgs_sent_, num_to_send); |
| break; |
| default: |
| assert(false); |
| break; |
| } |
| if ((server_try_cancel_ == DO_NOT_CANCEL) && !client_cancel_.cancel) { |
| EXPECT_TRUE(s.ok()); |
| EXPECT_EQ(response_.message(), desired_); |
| } else { |
| EXPECT_FALSE(s.ok()); |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| } |
| std::unique_lock<std::mutex> l(mu_); |
| done_ = true; |
| cv_.notify_one(); |
| } |
| void Await() { |
| std::unique_lock<std::mutex> l(mu_); |
| while (!done_) { |
| cv_.wait(l); |
| } |
| } |
| |
| private: |
| void MaybeWrite() { |
| if (client_cancel_.cancel && |
| num_msgs_sent_ == client_cancel_.ops_before_cancel) { |
| context_.TryCancel(); |
| } else if (num_msgs_to_send_ > num_msgs_sent_ + 1) { |
| StartWrite(&request_); |
| } else if (num_msgs_to_send_ == num_msgs_sent_ + 1) { |
| StartWriteLast(&request_, WriteOptions()); |
| } |
| } |
| EchoRequest request_; |
| EchoResponse response_; |
| ClientContext context_; |
| const ServerTryCancelRequestPhase server_try_cancel_; |
| int num_msgs_sent_{0}; |
| const int num_msgs_to_send_; |
| std::string desired_; |
| const ClientCancelInfo client_cancel_; |
| std::mutex mu_; |
| std::condition_variable cv_; |
| bool done_ = false; |
| }; |
| |
| TEST_P(ClientCallbackEnd2endTest, RequestStream) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| WriteClient test{stub_.get(), DO_NOT_CANCEL, 3}; |
| test.Await(); |
| // Make sure that the server interceptors were not notified to cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, ClientCancelsRequestStream) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| WriteClient test{stub_.get(), DO_NOT_CANCEL, 3, ClientCancelInfo{2}}; |
| test.Await(); |
| // Make sure that the server interceptors got the cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| // Server to cancel before doing reading the request |
| TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelBeforeReads) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| WriteClient test{stub_.get(), CANCEL_BEFORE_PROCESSING, 1}; |
| test.Await(); |
| // Make sure that the server interceptors were notified |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| // Server to cancel while reading a request from the stream in parallel |
| TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelDuringRead) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| WriteClient test{stub_.get(), CANCEL_DURING_PROCESSING, 10}; |
| test.Await(); |
| // Make sure that the server interceptors were notified |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| // Server to cancel after reading all the requests but before returning to the |
| // client |
| TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelAfterReads) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| WriteClient test{stub_.get(), CANCEL_AFTER_PROCESSING, 4}; |
| test.Await(); |
| // Make sure that the server interceptors were notified |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, UnaryReactor) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| class UnaryClient : public grpc::experimental::ClientUnaryReactor { |
| public: |
| UnaryClient(grpc::testing::EchoTestService::Stub* stub) { |
| cli_ctx_.AddMetadata("key1", "val1"); |
| cli_ctx_.AddMetadata("key2", "val2"); |
| request_.mutable_param()->set_echo_metadata_initially(true); |
| request_.set_message("Hello metadata"); |
| stub->experimental_async()->Echo(&cli_ctx_, &request_, &response_, this); |
| StartCall(); |
| } |
| void OnReadInitialMetadataDone(bool ok) override { |
| EXPECT_TRUE(ok); |
| EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key1")); |
| EXPECT_EQ( |
| "val1", |
| ToString(cli_ctx_.GetServerInitialMetadata().find("key1")->second)); |
| EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key2")); |
| EXPECT_EQ( |
| "val2", |
| ToString(cli_ctx_.GetServerInitialMetadata().find("key2")->second)); |
| initial_metadata_done_ = true; |
| } |
| void OnDone(const Status& s) override { |
| EXPECT_TRUE(initial_metadata_done_); |
| EXPECT_EQ(0u, cli_ctx_.GetServerTrailingMetadata().size()); |
| EXPECT_TRUE(s.ok()); |
| EXPECT_EQ(request_.message(), response_.message()); |
| std::unique_lock<std::mutex> l(mu_); |
| done_ = true; |
| cv_.notify_one(); |
| } |
| void Await() { |
| std::unique_lock<std::mutex> l(mu_); |
| while (!done_) { |
| cv_.wait(l); |
| } |
| } |
| |
| private: |
| EchoRequest request_; |
| EchoResponse response_; |
| ClientContext cli_ctx_; |
| std::mutex mu_; |
| std::condition_variable cv_; |
| bool done_{false}; |
| bool initial_metadata_done_{false}; |
| }; |
| |
| UnaryClient test{stub_.get()}; |
| test.Await(); |
| // Make sure that the server interceptors were not notified of a cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, GenericUnaryReactor) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| const std::string kMethodName("/grpc.testing.EchoTestService/Echo"); |
| class UnaryClient : public grpc::experimental::ClientUnaryReactor { |
| public: |
| UnaryClient(grpc::GenericStub* stub, const std::string& method_name) { |
| cli_ctx_.AddMetadata("key1", "val1"); |
| cli_ctx_.AddMetadata("key2", "val2"); |
| request_.mutable_param()->set_echo_metadata_initially(true); |
| request_.set_message("Hello metadata"); |
| send_buf_ = SerializeToByteBuffer(&request_); |
| |
| stub->experimental().PrepareUnaryCall(&cli_ctx_, method_name, |
| send_buf_.get(), &recv_buf_, this); |
| StartCall(); |
| } |
| void OnReadInitialMetadataDone(bool ok) override { |
| EXPECT_TRUE(ok); |
| EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key1")); |
| EXPECT_EQ( |
| "val1", |
| ToString(cli_ctx_.GetServerInitialMetadata().find("key1")->second)); |
| EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key2")); |
| EXPECT_EQ( |
| "val2", |
| ToString(cli_ctx_.GetServerInitialMetadata().find("key2")->second)); |
| initial_metadata_done_ = true; |
| } |
| void OnDone(const Status& s) override { |
| EXPECT_TRUE(initial_metadata_done_); |
| EXPECT_EQ(0u, cli_ctx_.GetServerTrailingMetadata().size()); |
| EXPECT_TRUE(s.ok()); |
| EchoResponse response; |
| EXPECT_TRUE(ParseFromByteBuffer(&recv_buf_, &response)); |
| EXPECT_EQ(request_.message(), response.message()); |
| std::unique_lock<std::mutex> l(mu_); |
| done_ = true; |
| cv_.notify_one(); |
| } |
| void Await() { |
| std::unique_lock<std::mutex> l(mu_); |
| while (!done_) { |
| cv_.wait(l); |
| } |
| } |
| |
| private: |
| EchoRequest request_; |
| std::unique_ptr<ByteBuffer> send_buf_; |
| ByteBuffer recv_buf_; |
| ClientContext cli_ctx_; |
| std::mutex mu_; |
| std::condition_variable cv_; |
| bool done_{false}; |
| bool initial_metadata_done_{false}; |
| }; |
| |
| UnaryClient test{generic_stub_.get(), kMethodName}; |
| test.Await(); |
| // Make sure that the server interceptors were not notified of a cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| class ReadClient : public grpc::experimental::ClientReadReactor<EchoResponse> { |
| public: |
| ReadClient(grpc::testing::EchoTestService::Stub* stub, |
| ServerTryCancelRequestPhase server_try_cancel, |
| ClientCancelInfo client_cancel = {}) |
| : server_try_cancel_(server_try_cancel), client_cancel_{client_cancel} { |
| if (server_try_cancel_ != DO_NOT_CANCEL) { |
| // Send server_try_cancel value in the client metadata |
| context_.AddMetadata(kServerTryCancelRequest, |
| std::to_string(server_try_cancel)); |
| } |
| request_.set_message("Hello client "); |
| stub->experimental_async()->ResponseStream(&context_, &request_, this); |
| if (client_cancel_.cancel && |
| reads_complete_ == client_cancel_.ops_before_cancel) { |
| context_.TryCancel(); |
| } |
| // Even if we cancel, read until failure because there might be responses |
| // pending |
| StartRead(&response_); |
| StartCall(); |
| } |
| void OnReadDone(bool ok) override { |
| if (!ok) { |
| if (server_try_cancel_ == DO_NOT_CANCEL && !client_cancel_.cancel) { |
| EXPECT_EQ(reads_complete_, kServerDefaultResponseStreamsToSend); |
| } |
| } else { |
| EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend); |
| EXPECT_EQ(response_.message(), |
| request_.message() + std::to_string(reads_complete_)); |
| reads_complete_++; |
| if (client_cancel_.cancel && |
| reads_complete_ == client_cancel_.ops_before_cancel) { |
| context_.TryCancel(); |
| } |
| // Even if we cancel, read until failure because there might be responses |
| // pending |
| StartRead(&response_); |
| } |
| } |
| void OnDone(const Status& s) override { |
| gpr_log(GPR_INFO, "Read %d messages", reads_complete_); |
| switch (server_try_cancel_) { |
| case DO_NOT_CANCEL: |
| if (!client_cancel_.cancel || client_cancel_.ops_before_cancel > |
| kServerDefaultResponseStreamsToSend) { |
| EXPECT_TRUE(s.ok()); |
| EXPECT_EQ(reads_complete_, kServerDefaultResponseStreamsToSend); |
| } else { |
| EXPECT_GE(reads_complete_, client_cancel_.ops_before_cancel); |
| EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend); |
| // Status might be ok or cancelled depending on whether server |
| // sent status before client cancel went through |
| if (!s.ok()) { |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| } |
| } |
| break; |
| case CANCEL_BEFORE_PROCESSING: |
| EXPECT_FALSE(s.ok()); |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| EXPECT_EQ(reads_complete_, 0); |
| break; |
| case CANCEL_DURING_PROCESSING: |
| case CANCEL_AFTER_PROCESSING: |
| // If server canceled while writing messages, client must have read |
| // less than or equal to the expected number of messages. Even if the |
| // server canceled after writing all messages, the RPC may be canceled |
| // before the Client got a chance to read all the messages. |
| EXPECT_FALSE(s.ok()); |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend); |
| break; |
| default: |
| assert(false); |
| } |
| std::unique_lock<std::mutex> l(mu_); |
| done_ = true; |
| cv_.notify_one(); |
| } |
| void Await() { |
| std::unique_lock<std::mutex> l(mu_); |
| while (!done_) { |
| cv_.wait(l); |
| } |
| } |
| |
| private: |
| EchoRequest request_; |
| EchoResponse response_; |
| ClientContext context_; |
| const ServerTryCancelRequestPhase server_try_cancel_; |
| int reads_complete_{0}; |
| const ClientCancelInfo client_cancel_; |
| std::mutex mu_; |
| std::condition_variable cv_; |
| bool done_ = false; |
| }; |
| |
| TEST_P(ClientCallbackEnd2endTest, ResponseStream) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| ReadClient test{stub_.get(), DO_NOT_CANCEL}; |
| test.Await(); |
| // Make sure that the server interceptors were not notified of a cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, ClientCancelsResponseStream) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| ReadClient test{stub_.get(), DO_NOT_CANCEL, ClientCancelInfo{2}}; |
| test.Await(); |
| // Because cancel in this case races with server finish, we can't be sure that |
| // server interceptors even see cancellation |
| } |
| |
| // Server to cancel before sending any response messages |
| TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelBefore) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| ReadClient test{stub_.get(), CANCEL_BEFORE_PROCESSING}; |
| test.Await(); |
| // Make sure that the server interceptors were notified |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| // Server to cancel while writing a response to the stream in parallel |
| TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelDuring) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| ReadClient test{stub_.get(), CANCEL_DURING_PROCESSING}; |
| test.Await(); |
| // Make sure that the server interceptors were notified |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| // Server to cancel after writing all the respones to the stream but before |
| // returning to the client |
| TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelAfter) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| ReadClient test{stub_.get(), CANCEL_AFTER_PROCESSING}; |
| test.Await(); |
| // Make sure that the server interceptors were notified |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| class BidiClient |
| : public grpc::experimental::ClientBidiReactor<EchoRequest, EchoResponse> { |
| public: |
| BidiClient(grpc::testing::EchoTestService::Stub* stub, |
| ServerTryCancelRequestPhase server_try_cancel, |
| int num_msgs_to_send, bool cork_metadata, bool first_write_async, |
| ClientCancelInfo client_cancel = {}) |
| : server_try_cancel_(server_try_cancel), |
| msgs_to_send_{num_msgs_to_send}, |
| client_cancel_{client_cancel} { |
| if (server_try_cancel_ != DO_NOT_CANCEL) { |
| // Send server_try_cancel value in the client metadata |
| context_.AddMetadata(kServerTryCancelRequest, |
| std::to_string(server_try_cancel)); |
| } |
| request_.set_message("Hello fren "); |
| context_.set_initial_metadata_corked(cork_metadata); |
| stub->experimental_async()->BidiStream(&context_, this); |
| MaybeAsyncWrite(first_write_async); |
| StartRead(&response_); |
| StartCall(); |
| } |
| void OnReadDone(bool ok) override { |
| if (!ok) { |
| if (server_try_cancel_ == DO_NOT_CANCEL) { |
| if (!client_cancel_.cancel) { |
| EXPECT_EQ(reads_complete_, msgs_to_send_); |
| } else { |
| EXPECT_LE(reads_complete_, writes_complete_); |
| } |
| } |
| } else { |
| EXPECT_LE(reads_complete_, msgs_to_send_); |
| EXPECT_EQ(response_.message(), request_.message()); |
| reads_complete_++; |
| StartRead(&response_); |
| } |
| } |
| void OnWriteDone(bool ok) override { |
| if (async_write_thread_.joinable()) { |
| async_write_thread_.join(); |
| RemoveHold(); |
| } |
| if (server_try_cancel_ == DO_NOT_CANCEL) { |
| EXPECT_TRUE(ok); |
| } else if (!ok) { |
| return; |
| } |
| writes_complete_++; |
| MaybeWrite(); |
| } |
| void OnDone(const Status& s) override { |
| gpr_log(GPR_INFO, "Sent %d messages", writes_complete_); |
| gpr_log(GPR_INFO, "Read %d messages", reads_complete_); |
| switch (server_try_cancel_) { |
| case DO_NOT_CANCEL: |
| if (!client_cancel_.cancel || |
| client_cancel_.ops_before_cancel > msgs_to_send_) { |
| EXPECT_TRUE(s.ok()); |
| EXPECT_EQ(writes_complete_, msgs_to_send_); |
| EXPECT_EQ(reads_complete_, writes_complete_); |
| } else { |
| EXPECT_FALSE(s.ok()); |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| EXPECT_EQ(writes_complete_, client_cancel_.ops_before_cancel); |
| EXPECT_LE(reads_complete_, writes_complete_); |
| } |
| break; |
| case CANCEL_BEFORE_PROCESSING: |
| EXPECT_FALSE(s.ok()); |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| // The RPC is canceled before the server did any work or returned any |
| // reads, but it's possible that some writes took place first from the |
| // client |
| EXPECT_LE(writes_complete_, msgs_to_send_); |
| EXPECT_EQ(reads_complete_, 0); |
| break; |
| case CANCEL_DURING_PROCESSING: |
| EXPECT_FALSE(s.ok()); |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| EXPECT_LE(writes_complete_, msgs_to_send_); |
| EXPECT_LE(reads_complete_, writes_complete_); |
| break; |
| case CANCEL_AFTER_PROCESSING: |
| EXPECT_FALSE(s.ok()); |
| EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); |
| EXPECT_EQ(writes_complete_, msgs_to_send_); |
| // The Server canceled after reading the last message and after writing |
| // the message to the client. However, the RPC cancellation might have |
| // taken effect before the client actually read the response. |
| EXPECT_LE(reads_complete_, writes_complete_); |
| break; |
| default: |
| assert(false); |
| } |
| std::unique_lock<std::mutex> l(mu_); |
| done_ = true; |
| cv_.notify_one(); |
| } |
| void Await() { |
| std::unique_lock<std::mutex> l(mu_); |
| while (!done_) { |
| cv_.wait(l); |
| } |
| } |
| |
| private: |
| void MaybeAsyncWrite(bool first_write_async) { |
| if (first_write_async) { |
| // Make sure that we have a write to issue. |
| // TODO(vjpai): Make this work with 0 writes case as well. |
| assert(msgs_to_send_ >= 1); |
| |
| AddHold(); |
| async_write_thread_ = std::thread([this] { |
| std::unique_lock<std::mutex> lock(async_write_thread_mu_); |
| async_write_thread_cv_.wait( |
| lock, [this] { return async_write_thread_start_; }); |
| MaybeWrite(); |
| }); |
| std::lock_guard<std::mutex> lock(async_write_thread_mu_); |
| async_write_thread_start_ = true; |
| async_write_thread_cv_.notify_one(); |
| return; |
| } |
| MaybeWrite(); |
| } |
| void MaybeWrite() { |
| if (client_cancel_.cancel && |
| writes_complete_ == client_cancel_.ops_before_cancel) { |
| context_.TryCancel(); |
| } else if (writes_complete_ == msgs_to_send_) { |
| StartWritesDone(); |
| } else { |
| StartWrite(&request_); |
| } |
| } |
| EchoRequest request_; |
| EchoResponse response_; |
| ClientContext context_; |
| const ServerTryCancelRequestPhase server_try_cancel_; |
| int reads_complete_{0}; |
| int writes_complete_{0}; |
| const int msgs_to_send_; |
| const ClientCancelInfo client_cancel_; |
| std::mutex mu_; |
| std::condition_variable cv_; |
| bool done_ = false; |
| std::thread async_write_thread_; |
| bool async_write_thread_start_ = false; |
| std::mutex async_write_thread_mu_; |
| std::condition_variable async_write_thread_cv_; |
| }; |
| |
| TEST_P(ClientCallbackEnd2endTest, BidiStream) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| BidiClient test(stub_.get(), DO_NOT_CANCEL, |
| kServerDefaultResponseStreamsToSend, |
| /*cork_metadata=*/false, /*first_write_async=*/false); |
| test.Await(); |
| // Make sure that the server interceptors were not notified of a cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, BidiStreamFirstWriteAsync) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| BidiClient test(stub_.get(), DO_NOT_CANCEL, |
| kServerDefaultResponseStreamsToSend, |
| /*cork_metadata=*/false, /*first_write_async=*/true); |
| test.Await(); |
| // Make sure that the server interceptors were not notified of a cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, BidiStreamCorked) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| BidiClient test(stub_.get(), DO_NOT_CANCEL, |
| kServerDefaultResponseStreamsToSend, |
| /*cork_metadata=*/true, /*first_write_async=*/false); |
| test.Await(); |
| // Make sure that the server interceptors were not notified of a cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, BidiStreamCorkedFirstWriteAsync) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| BidiClient test(stub_.get(), DO_NOT_CANCEL, |
| kServerDefaultResponseStreamsToSend, |
| /*cork_metadata=*/true, /*first_write_async=*/true); |
| test.Await(); |
| // Make sure that the server interceptors were not notified of a cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, ClientCancelsBidiStream) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| BidiClient test(stub_.get(), DO_NOT_CANCEL, |
| kServerDefaultResponseStreamsToSend, |
| /*cork_metadata=*/false, /*first_write_async=*/false, |
| ClientCancelInfo(2)); |
| test.Await(); |
| // Make sure that the server interceptors were notified of a cancel |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| // Server to cancel before reading/writing any requests/responses on the stream |
| TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelBefore) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| BidiClient test(stub_.get(), CANCEL_BEFORE_PROCESSING, /*num_msgs_to_send=*/2, |
| /*cork_metadata=*/false, /*first_write_async=*/false); |
| test.Await(); |
| // Make sure that the server interceptors were notified |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| // Server to cancel while reading/writing requests/responses on the stream in |
| // parallel |
| TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelDuring) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| BidiClient test(stub_.get(), CANCEL_DURING_PROCESSING, |
| /*num_msgs_to_send=*/10, /*cork_metadata=*/false, |
| /*first_write_async=*/false); |
| test.Await(); |
| // Make sure that the server interceptors were notified |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| // Server to cancel after reading/writing all requests/responses on the stream |
| // but before returning to the client |
| TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelAfter) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| BidiClient test(stub_.get(), CANCEL_AFTER_PROCESSING, /*num_msgs_to_send=*/5, |
| /*cork_metadata=*/false, /*first_write_async=*/false); |
| test.Await(); |
| // Make sure that the server interceptors were notified |
| if (GetParam().use_interceptors) { |
| EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel()); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, SimultaneousReadAndWritesDone) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| class Client : public grpc::experimental::ClientBidiReactor<EchoRequest, |
| EchoResponse> { |
| public: |
| Client(grpc::testing::EchoTestService::Stub* stub) { |
| request_.set_message("Hello bidi "); |
| stub->experimental_async()->BidiStream(&context_, this); |
| StartWrite(&request_); |
| StartCall(); |
| } |
| void OnReadDone(bool ok) override { |
| EXPECT_TRUE(ok); |
| EXPECT_EQ(response_.message(), request_.message()); |
| } |
| void OnWriteDone(bool ok) override { |
| EXPECT_TRUE(ok); |
| // Now send out the simultaneous Read and WritesDone |
| StartWritesDone(); |
| StartRead(&response_); |
| } |
| void OnDone(const Status& s) override { |
| EXPECT_TRUE(s.ok()); |
| EXPECT_EQ(response_.message(), request_.message()); |
| std::unique_lock<std::mutex> l(mu_); |
| done_ = true; |
| cv_.notify_one(); |
| } |
| void Await() { |
| std::unique_lock<std::mutex> l(mu_); |
| while (!done_) { |
| cv_.wait(l); |
| } |
| } |
| |
| private: |
| EchoRequest request_; |
| EchoResponse response_; |
| ClientContext context_; |
| std::mutex mu_; |
| std::condition_variable cv_; |
| bool done_ = false; |
| } test{stub_.get()}; |
| |
| test.Await(); |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, UnimplementedRpc) { |
| MAYBE_SKIP_TEST; |
| ChannelArguments args; |
| const auto& channel_creds = GetCredentialsProvider()->GetChannelCredentials( |
| GetParam().credentials_type, &args); |
| std::shared_ptr<Channel> channel = |
| (GetParam().protocol == Protocol::TCP) |
| ? ::grpc::CreateCustomChannel(server_address_.str(), channel_creds, |
| args) |
| : server_->InProcessChannel(args); |
| std::unique_ptr<grpc::testing::UnimplementedEchoService::Stub> stub; |
| stub = grpc::testing::UnimplementedEchoService::NewStub(channel); |
| EchoRequest request; |
| EchoResponse response; |
| ClientContext cli_ctx; |
| request.set_message("Hello world."); |
| std::mutex mu; |
| std::condition_variable cv; |
| bool done = false; |
| stub->experimental_async()->Unimplemented( |
| &cli_ctx, &request, &response, [&done, &mu, &cv](Status s) { |
| EXPECT_EQ(StatusCode::UNIMPLEMENTED, s.error_code()); |
| EXPECT_EQ("", s.error_message()); |
| |
| std::lock_guard<std::mutex> l(mu); |
| done = true; |
| cv.notify_one(); |
| }); |
| std::unique_lock<std::mutex> l(mu); |
| while (!done) { |
| cv.wait(l); |
| } |
| } |
| |
| TEST_P(ClientCallbackEnd2endTest, |
| ResponseStreamExtraReactionFlowReadsUntilDone) { |
| MAYBE_SKIP_TEST; |
| ResetStub(); |
| class ReadAllIncomingDataClient |
| : public grpc::experimental::ClientReadReactor<EchoResponse> { |
| public: |
| ReadAllIncomingDataClient(grpc::testing::EchoTestService::Stub* stub) { |
| request_.set_message("Hello client "); |
| stub->experimental_async()->ResponseStream(&context_, &request_, this); |
| } |
| bool WaitForReadDone() { |
| std::unique_lock<std::mutex> l(mu_); |
| while (!read_done_) { |
| read_cv_.wait(l); |
| } |
| read_done_ = false; |
| return read_ok_; |
| } |
| void Await() { |
| std::unique_lock<std::mutex> l(mu_); |
| while (!done_) { |
| done_cv_.wait(l); |
| } |
| } |
| // RemoveHold under the same lock used for OnDone to make sure that we don't |
| // call OnDone directly or indirectly from the RemoveHold function. |
| void RemoveHoldUnderLock() { |
| std::unique_lock<std::mutex> l(mu_); |
| RemoveHold(); |
| } |
| const Status& status() { |
| std::unique_lock<std::mutex> l(mu_); |
| return status_; |
| } |
| |
| private: |
| void OnReadDone(bool ok) override { |
| std::unique_lock<std::mutex> l(mu_); |
| read_ok_ = ok; |
| read_done_ = true; |
| read_cv_.notify_one(); |
| } |
| void OnDone(const Status& s) override { |
| std::unique_lock<std::mutex> l(mu_); |
| done_ = true; |
| status_ = s; |
| done_cv_.notify_one(); |
| } |
| |
| EchoRequest request_; |
| EchoResponse response_; |
| ClientContext context_; |
| bool read_ok_ = false; |
| bool read_done_ = false; |
| std::mutex mu_; |
| std::condition_variable read_cv_; |
| std::condition_variable done_cv_; |
| bool done_ = false; |
| Status status_; |
| } client{stub_.get()}; |
| |
| int reads_complete = 0; |
| client.AddHold(); |
| client.StartCall(); |
| |
| EchoResponse response; |
| bool read_ok = true; |
| while (read_ok) { |
| client.StartRead(&response); |
| read_ok = client.WaitForReadDone(); |
| if (read_ok) { |
| ++reads_complete; |
| } |
| } |
| client.RemoveHoldUnderLock(); |
| client.Await(); |
| |
| EXPECT_EQ(kServerDefaultResponseStreamsToSend, reads_complete); |
| EXPECT_EQ(client.status().error_code(), grpc::StatusCode::OK); |
| } |
| |
| std::vector<TestScenario> CreateTestScenarios(bool test_insecure) { |
| #if TARGET_OS_IPHONE |
| // Workaround Apple CFStream bug |
| gpr_setenv("grpc_cfstream", "0"); |
| #endif |
| |
| std::vector<TestScenario> scenarios; |
| std::vector<std::string> credentials_types{ |
| GetCredentialsProvider()->GetSecureCredentialsTypeList()}; |
| auto insec_ok = [] { |
| // Only allow insecure credentials type when it is registered with the |
| // provider. User may create providers that do not have insecure. |
| return GetCredentialsProvider()->GetChannelCredentials( |
| kInsecureCredentialsType, nullptr) != nullptr; |
| }; |
| if (test_insecure && insec_ok()) { |
| credentials_types.push_back(kInsecureCredentialsType); |
| } |
| GPR_ASSERT(!credentials_types.empty()); |
| |
| bool barr[]{false, true}; |
| Protocol parr[]{Protocol::INPROC, Protocol::TCP}; |
| for (Protocol p : parr) { |
| for (const auto& cred : credentials_types) { |
| // TODO(vjpai): Test inproc with secure credentials when feasible |
| if (p == Protocol::INPROC && |
| (cred != kInsecureCredentialsType || !insec_ok())) { |
| continue; |
| } |
| for (bool callback_server : barr) { |
| for (bool use_interceptors : barr) { |
| scenarios.emplace_back(callback_server, p, use_interceptors, cred); |
| } |
| } |
| } |
| } |
| return scenarios; |
| } |
| |
| INSTANTIATE_TEST_SUITE_P(ClientCallbackEnd2endTest, ClientCallbackEnd2endTest, |
| ::testing::ValuesIn(CreateTestScenarios(true))); |
| |
| } // namespace |
| } // namespace testing |
| } // namespace grpc |
| |
| int main(int argc, char** argv) { |
| ::testing::InitGoogleTest(&argc, argv); |
| grpc::testing::TestEnvironment env(argc, argv); |
| grpc_init(); |
| int ret = RUN_ALL_TESTS(); |
| grpc_shutdown(); |
| return ret; |
| } |