The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#) https://grpc.io/
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

530 lines
21 KiB

//
//
// Copyright 2015 gRPC authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//
//
#include <stdio.h>
#include <string.h>
#include <string>
#include <gmock/gmock.h>
#include "absl/strings/str_cat.h"
#include "absl/strings/str_format.h"
#include <grpc/byte_buffer.h>
#include <grpc/grpc.h>
#include <grpc/grpc_security.h>
#include <grpc/support/alloc.h>
#include <grpc/support/log.h>
#include <grpc/support/time.h>
#include "src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.h"
#include "src/core/lib/channel/channel_args.h"
#include "src/core/lib/config/config_vars.h"
#include "src/core/lib/config/core_configuration.h"
#include "src/core/lib/debug/stats.h"
#include "src/core/lib/debug/stats_data.h"
#include "src/core/lib/event_engine/ares_resolver.h"
#include "src/core/lib/event_engine/default_event_engine.h"
[EventEngine] PosixEventEngine DNS Resolver (#32701) This PR implements a c-ares based DNS resolver for EventEngine with the reference from the original [grpc_ares_wrapper.h](../blob/master/src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.h). The PosixEventEngine DNSResolver is implemented on top of that. Tests which use the client channel resolver API ([resolver.h](../blob/master/src/core/lib/resolver/resolver.h#L54)) are ported, namely the [resolver_component_test.cc](../blob/master/test/cpp/naming/resolver_component_test.cc) and the [cancel_ares_query_test.cc](../blob/master/test/cpp/naming/cancel_ares_query_test.cc). The WindowsEventEngine DNSResolver will use the same EventEngine's grpc_ares_wrapper and will be worked on next. The [resolve_address_test.cc](https://github.com/grpc/grpc/blob/master/test/core/iomgr/resolve_address_test.cc) which uses the iomgr [DNSResolver](../blob/master/src/core/lib/iomgr/resolve_address.h#L44) API has been ported to EventEngine's dns_test.cc. That leaves only 2 tests which use iomgr's API, notably the [dns_resolver_cooldown_test.cc](../blob/master/test/core/client_channel/resolvers/dns_resolver_cooldown_test.cc) and the [goaway_server_test.cc](../blob/master/test/core/end2end/goaway_server_test.cc) which probably need to be restructured to use EventEngine DNSResolver (for one thing they override the original grpc_ares_wrapper's free functions). I will try to tackle these in the next step. <!-- If you know who should review your pull request, please assign it to that person, otherwise the pull request would get assigned randomly. If your pull request is for a specific language, please add the appropriate lang label. -->
1 year ago
#include "src/core/lib/experiments/experiments.h"
#include "src/core/lib/gpr/string.h"
#include "src/core/lib/gprpp/crash.h"
#include "src/core/lib/gprpp/orphanable.h"
#include "src/core/lib/gprpp/thd.h"
#include "src/core/lib/gprpp/work_serializer.h"
#include "src/core/lib/iomgr/iomgr.h"
#include "src/core/lib/iomgr/pollset.h"
#include "src/core/lib/iomgr/pollset_set.h"
#include "src/core/lib/resolver/resolver.h"
#include "src/core/lib/resolver/resolver_registry.h"
#include "test/core/end2end/cq_verifier.h"
#include "test/core/util/cmdline.h"
#include "test/core/util/fake_udp_and_tcp_server.h"
#include "test/core/util/port.h"
#include "test/core/util/socket_use_after_close_detector.h"
#include "test/core/util/test_config.h"
[EventEngine] PosixEventEngine DNS Resolver (#32701) This PR implements a c-ares based DNS resolver for EventEngine with the reference from the original [grpc_ares_wrapper.h](../blob/master/src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.h). The PosixEventEngine DNSResolver is implemented on top of that. Tests which use the client channel resolver API ([resolver.h](../blob/master/src/core/lib/resolver/resolver.h#L54)) are ported, namely the [resolver_component_test.cc](../blob/master/test/cpp/naming/resolver_component_test.cc) and the [cancel_ares_query_test.cc](../blob/master/test/cpp/naming/cancel_ares_query_test.cc). The WindowsEventEngine DNSResolver will use the same EventEngine's grpc_ares_wrapper and will be worked on next. The [resolve_address_test.cc](https://github.com/grpc/grpc/blob/master/test/core/iomgr/resolve_address_test.cc) which uses the iomgr [DNSResolver](../blob/master/src/core/lib/iomgr/resolve_address.h#L44) API has been ported to EventEngine's dns_test.cc. That leaves only 2 tests which use iomgr's API, notably the [dns_resolver_cooldown_test.cc](../blob/master/test/core/client_channel/resolvers/dns_resolver_cooldown_test.cc) and the [goaway_server_test.cc](../blob/master/test/core/end2end/goaway_server_test.cc) which probably need to be restructured to use EventEngine DNSResolver (for one thing they override the original grpc_ares_wrapper's free functions). I will try to tackle these in the next step. <!-- If you know who should review your pull request, please assign it to that person, otherwise the pull request would get assigned randomly. If your pull request is for a specific language, please add the appropriate lang label. -->
1 year ago
#include "test/cpp/util/test_config.h"
#ifdef GPR_WINDOWS
#include "src/core/lib/iomgr/sockaddr_windows.h"
#include "src/core/lib/iomgr/socket_windows.h"
#define BAD_SOCKET_RETURN_VAL INVALID_SOCKET
#else
#include "src/core/lib/iomgr/sockaddr_posix.h"
#define BAD_SOCKET_RETURN_VAL (-1)
#endif
namespace {
using ::grpc_event_engine::experimental::GetDefaultEventEngine;
void* Tag(intptr_t t) { return reinterpret_cast<void*>(t); }
gpr_timespec FiveSecondsFromNow(void) {
return grpc_timeout_seconds_to_deadline(5);
}
void DrainCq(grpc_completion_queue* cq) {
grpc_event ev;
do {
ev = grpc_completion_queue_next(cq, FiveSecondsFromNow(), nullptr);
} while (ev.type != GRPC_QUEUE_SHUTDOWN);
}
void EndTest(grpc_channel* client, grpc_completion_queue* cq) {
grpc_channel_destroy(client);
grpc_completion_queue_shutdown(cq);
DrainCq(cq);
grpc_completion_queue_destroy(cq);
}
struct ArgsStruct {
gpr_atm done_atm;
gpr_mu* mu;
grpc_pollset* pollset;
grpc_pollset_set* pollset_set;
std::shared_ptr<grpc_core::WorkSerializer> lock;
grpc_channel_args* channel_args;
};
void ArgsInit(ArgsStruct* args) {
args->pollset = static_cast<grpc_pollset*>(gpr_zalloc(grpc_pollset_size()));
grpc_pollset_init(args->pollset, &args->mu);
args->pollset_set = grpc_pollset_set_create();
grpc_pollset_set_add_pollset(args->pollset_set, args->pollset);
args->lock = std::make_shared<grpc_core::WorkSerializer>();
gpr_atm_rel_store(&args->done_atm, 0);
args->channel_args = nullptr;
}
void DoNothing(void* /*arg*/, grpc_error_handle /*error*/) {}
void ArgsFinish(ArgsStruct* args) {
grpc_pollset_set_del_pollset(args->pollset_set, args->pollset);
grpc_pollset_set_destroy(args->pollset_set);
grpc_closure DoNothing_cb;
GRPC_CLOSURE_INIT(&DoNothing_cb, DoNothing, nullptr,
grpc_schedule_on_exec_ctx);
grpc_pollset_shutdown(args->pollset, &DoNothing_cb);
// exec_ctx needs to be flushed before calling grpc_pollset_destroy()
grpc_channel_args_destroy(args->channel_args);
grpc_core::ExecCtx::Get()->Flush();
grpc_pollset_destroy(args->pollset);
gpr_free(args->pollset);
}
void PollPollsetUntilRequestDone(ArgsStruct* args) {
while (true) {
bool done = gpr_atm_acq_load(&args->done_atm) != 0;
if (done) {
break;
}
grpc_pollset_worker* worker = nullptr;
grpc_core::ExecCtx exec_ctx;
gpr_mu_lock(args->mu);
GRPC_LOG_IF_ERROR("pollset_work",
grpc_pollset_work(args->pollset, &worker,
grpc_core::Timestamp::InfFuture()));
gpr_mu_unlock(args->mu);
}
}
class AssertFailureResultHandler : public grpc_core::Resolver::ResultHandler {
public:
explicit AssertFailureResultHandler(ArgsStruct* args) : args_(args) {}
~AssertFailureResultHandler() override {
gpr_atm_rel_store(&args_->done_atm, 1);
gpr_mu_lock(args_->mu);
GRPC_LOG_IF_ERROR("pollset_kick",
grpc_pollset_kick(args_->pollset, nullptr));
gpr_mu_unlock(args_->mu);
}
void ReportResult(grpc_core::Resolver::Result /*result*/) override {
grpc_core::Crash("unreachable");
}
private:
ArgsStruct* args_;
};
void TestCancelActiveDNSQuery(ArgsStruct* args) {
grpc_core::testing::FakeUdpAndTcpServer fake_dns_server(
grpc_core::testing::FakeUdpAndTcpServer::AcceptMode::
kWaitForClientToSendFirstBytes,
grpc_core::testing::FakeUdpAndTcpServer::CloseSocketUponCloseFromPeer);
std::string client_target = absl::StrFormat(
"dns://[::1]:%d/dont-care-since-wont-be-resolved.test.com:1234",
fake_dns_server.port());
// create resolver and resolve
grpc_core::OrphanablePtr<grpc_core::Resolver> resolver =
grpc_core::CoreConfiguration::Get().resolver_registry().CreateResolver(
client_target.c_str(),
grpc_core::ChannelArgs().SetObject(GetDefaultEventEngine()),
args->pollset_set, args->lock,
std::unique_ptr<grpc_core::Resolver::ResultHandler>(
new AssertFailureResultHandler(args)));
resolver->StartLocked();
// Without resetting and causing resolver shutdown, the
// PollPollsetUntilRequestDone call should never finish.
resolver.reset();
grpc_core::ExecCtx::Get()->Flush();
PollPollsetUntilRequestDone(args);
ArgsFinish(args);
}
class CancelDuringAresQuery : public ::testing::Test {
protected:
static void SetUpTestSuite() {
grpc_core::ConfigVars::Overrides overrides;
overrides.dns_resolver = "ares";
grpc_core::ConfigVars::SetOverrides(overrides);
grpc_init();
}
static void TearDownTestSuite() { grpc_shutdown(); }
};
TEST_F(CancelDuringAresQuery, TestCancelActiveDNSQuery) {
grpc_core::ExecCtx exec_ctx;
ArgsStruct args;
ArgsInit(&args);
TestCancelActiveDNSQuery(&args);
}
#ifdef GPR_WINDOWS
void MaybePollArbitraryPollsetTwice() {
grpc_pollset* pollset = (grpc_pollset*)gpr_zalloc(grpc_pollset_size());
gpr_mu* mu;
grpc_pollset_init(pollset, &mu);
grpc_pollset_worker* worker = nullptr;
// Make a zero timeout poll
gpr_mu_lock(mu);
GRPC_LOG_IF_ERROR(
"pollset_work",
grpc_pollset_work(pollset, &worker, grpc_core::Timestamp::Now()));
gpr_mu_unlock(mu);
grpc_core::ExecCtx::Get()->Flush();
// Make a second zero-timeout poll (in case the first one
// short-circuited by picking up a previous "kick")
gpr_mu_lock(mu);
GRPC_LOG_IF_ERROR(
"pollset_work",
grpc_pollset_work(pollset, &worker, grpc_core::Timestamp::Now()));
gpr_mu_unlock(mu);
grpc_core::ExecCtx::Get()->Flush();
grpc_pollset_destroy(pollset);
gpr_free(pollset);
}
#else
void MaybePollArbitraryPollsetTwice() {}
#endif
TEST_F(CancelDuringAresQuery, TestFdsAreDeletedFromPollsetSet) {
grpc_core::ExecCtx exec_ctx;
ArgsStruct args;
ArgsInit(&args);
// Add fake_other_pollset_set into the mix to test
// that we're explicitly deleting fd's from their pollset.
// If we aren't doing so, then the remaining presence of
// "fake_other_pollset_set" after the request is done and the resolver
// pollset set is destroyed should keep the resolver's fd alive and
// fail the test.
grpc_pollset_set* fake_other_pollset_set = grpc_pollset_set_create();
grpc_pollset_set_add_pollset_set(fake_other_pollset_set, args.pollset_set);
// Note that running the cancellation c-ares test is somewhat irrelevant for
// this test. This test only cares about what happens to fd's that c-ares
// opens.
TestCancelActiveDNSQuery(&args);
// This test relies on the assumption that cancelling a c-ares query
// will flush out all callbacks on the current exec ctx, which is true
// on posix platforms but not on Windows, because fd shutdown on Windows
// requires a trip through the polling loop to schedule the callback.
// So we need to do extra polling work on Windows to free things up.
MaybePollArbitraryPollsetTwice();
EXPECT_EQ(grpc_iomgr_count_objects_for_testing(), 0u);
grpc_pollset_set_destroy(fake_other_pollset_set);
}
std::string kFakeName = "dont-care-since-wont-be-resolved.test.com:1234";
void TestCancelDuringActiveQuery(
grpc_status_code expected_status_code,
absl::string_view expected_error_message_substring,
gpr_timespec rpc_deadline, int dns_query_timeout_ms,
int fake_dns_server_port) {
// Create a call that will try to use the fake DNS server
std::string client_target =
absl::StrFormat("dns://[::1]:%d/%s", fake_dns_server_port, kFakeName);
grpc_channel_args* client_args = nullptr;
if (dns_query_timeout_ms >= 0) {
grpc_arg arg;
arg.type = GRPC_ARG_INTEGER;
arg.key = const_cast<char*>(GRPC_ARG_DNS_ARES_QUERY_TIMEOUT_MS);
arg.value.integer = dns_query_timeout_ms;
client_args = grpc_channel_args_copy_and_add(nullptr, &arg, 1);
}
grpc_channel_credentials* creds = grpc_insecure_credentials_create();
grpc_channel* client =
grpc_channel_create(client_target.c_str(), creds, client_args);
grpc_channel_credentials_release(creds);
grpc_completion_queue* cq = grpc_completion_queue_create_for_next(nullptr);
grpc_core::CqVerifier cqv(cq);
grpc_call* call = grpc_channel_create_call(
client, nullptr, GRPC_PROPAGATE_DEFAULTS, cq,
grpc_slice_from_static_string("/foo"), nullptr, rpc_deadline, nullptr);
GPR_ASSERT(call);
grpc_metadata_array initial_metadata_recv;
grpc_metadata_array trailing_metadata_recv;
grpc_metadata_array request_metadata_recv;
grpc_metadata_array_init(&initial_metadata_recv);
grpc_metadata_array_init(&trailing_metadata_recv);
grpc_metadata_array_init(&request_metadata_recv);
grpc_call_details call_details;
grpc_call_details_init(&call_details);
grpc_status_code status;
const char* error_string;
grpc_slice details;
// Set ops for client the request
grpc_op ops_base[6];
memset(ops_base, 0, sizeof(ops_base));
grpc_op* op = ops_base;
op->op = GRPC_OP_SEND_INITIAL_METADATA;
op->data.send_initial_metadata.count = 0;
op->flags = 0;
op->reserved = nullptr;
op++;
op->op = GRPC_OP_SEND_CLOSE_FROM_CLIENT;
op->flags = 0;
op->reserved = nullptr;
op++;
op->op = GRPC_OP_RECV_INITIAL_METADATA;
op->data.recv_initial_metadata.recv_initial_metadata = &initial_metadata_recv;
op->flags = 0;
op->reserved = nullptr;
op++;
op->op = GRPC_OP_RECV_STATUS_ON_CLIENT;
op->data.recv_status_on_client.trailing_metadata = &trailing_metadata_recv;
op->data.recv_status_on_client.status = &status;
op->data.recv_status_on_client.status_details = &details;
op->data.recv_status_on_client.error_string = &error_string;
op->flags = 0;
op->reserved = nullptr;
op++;
// Run the call and sanity check it failed as expected
grpc_call_error error = grpc_call_start_batch(
call, ops_base, static_cast<size_t>(op - ops_base), Tag(1), nullptr);
EXPECT_EQ(GRPC_CALL_OK, error);
cqv.Expect(Tag(1), true);
cqv.Verify();
EXPECT_EQ(status, expected_status_code);
EXPECT_THAT(std::string(error_string),
testing::HasSubstr(expected_error_message_substring));
// Teardown
grpc_channel_args_destroy(client_args);
grpc_slice_unref(details);
gpr_free(const_cast<char*>(error_string));
grpc_metadata_array_destroy(&initial_metadata_recv);
grpc_metadata_array_destroy(&trailing_metadata_recv);
grpc_metadata_array_destroy(&request_metadata_recv);
grpc_call_details_destroy(&call_details);
grpc_call_unref(call);
EndTest(client, cq);
}
TEST_F(CancelDuringAresQuery,
TestHitDeadlineAndDestroyChannelDuringAresResolutionIsGraceful) {
grpc_core::testing::SocketUseAfterCloseDetector
socket_use_after_close_detector;
grpc_core::testing::FakeUdpAndTcpServer fake_dns_server(
grpc_core::testing::FakeUdpAndTcpServer::AcceptMode::
kWaitForClientToSendFirstBytes,
grpc_core::testing::FakeUdpAndTcpServer::CloseSocketUponCloseFromPeer);
grpc_status_code expected_status_code = GRPC_STATUS_DEADLINE_EXCEEDED;
// The RPC deadline should go off well before the DNS resolution
// timeout fires.
gpr_timespec rpc_deadline = grpc_timeout_milliseconds_to_deadline(100);
int dns_query_timeout_ms = -1; // don't set query timeout
TestCancelDuringActiveQuery(
expected_status_code, "" /* expected error message substring */,
rpc_deadline, dns_query_timeout_ms, fake_dns_server.port());
}
TEST_F(
CancelDuringAresQuery,
TestHitDeadlineAndDestroyChannelDuringAresResolutionWithQueryTimeoutIsGraceful) {
grpc_core::testing::SocketUseAfterCloseDetector
socket_use_after_close_detector;
grpc_core::testing::FakeUdpAndTcpServer fake_dns_server(
grpc_core::testing::FakeUdpAndTcpServer::AcceptMode::
kWaitForClientToSendFirstBytes,
grpc_core::testing::FakeUdpAndTcpServer::CloseSocketUponCloseFromPeer);
grpc_status_code expected_status_code = GRPC_STATUS_UNAVAILABLE;
std::string expected_error_message_substring;
if (grpc_core::IsEventEngineDnsEnabled()) {
expected_error_message_substring =
absl::StrCat("errors resolving ", kFakeName);
} else {
expected_error_message_substring =
absl::StrCat("DNS resolution failed for ", kFakeName);
}
// The DNS resolution timeout should fire well before the
// RPC's deadline expires.
gpr_timespec rpc_deadline = grpc_timeout_seconds_to_deadline(10);
int dns_query_timeout_ms = 1;
TestCancelDuringActiveQuery(expected_status_code,
expected_error_message_substring, rpc_deadline,
dns_query_timeout_ms, fake_dns_server.port());
}
TEST_F(
CancelDuringAresQuery,
TestHitDeadlineAndDestroyChannelDuringAresResolutionWithZeroQueryTimeoutIsGraceful) {
grpc_core::testing::SocketUseAfterCloseDetector
socket_use_after_close_detector;
grpc_core::testing::FakeUdpAndTcpServer fake_dns_server(
grpc_core::testing::FakeUdpAndTcpServer::AcceptMode::
kWaitForClientToSendFirstBytes,
grpc_core::testing::FakeUdpAndTcpServer::CloseSocketUponCloseFromPeer);
grpc_status_code expected_status_code = GRPC_STATUS_DEADLINE_EXCEEDED;
// The RPC deadline should go off well before the DNS resolution
// timeout fires.
gpr_timespec rpc_deadline = grpc_timeout_milliseconds_to_deadline(100);
int dns_query_timeout_ms = 0; // disable query timeouts
TestCancelDuringActiveQuery(
expected_status_code, "" /* expected error message substring */,
rpc_deadline, dns_query_timeout_ms, fake_dns_server.port());
}
TEST_F(CancelDuringAresQuery, TestQueryFailsBecauseTcpServerClosesSocket) {
grpc_core::testing::SocketUseAfterCloseDetector
socket_use_after_close_detector;
// Use a fake TCP server that immediately closes the socket and causes
// c-ares to pick up a socket read error, while the previous socket
// connect/writes succeeded. Meanwhile, force c-ares to only use TCP.
// The goal is to hit a socket use-after-close bug described in
// https://github.com/grpc/grpc/pull/33871.
grpc_core::testing::FakeUdpAndTcpServer fake_dns_server(
grpc_core::testing::FakeUdpAndTcpServer::AcceptMode::
kWaitForClientToSendFirstBytes,
grpc_core::testing::FakeUdpAndTcpServer::
CloseSocketUponReceivingBytesFromPeer);
if (grpc_core::IsEventEngineDnsEnabled()) {
g_event_engine_grpc_ares_test_only_force_tcp = true;
} else {
g_grpc_ares_test_only_force_tcp = true;
}
grpc_status_code expected_status_code = GRPC_STATUS_UNAVAILABLE;
std::string expected_error_message_substring;
if (grpc_core::IsEventEngineDnsEnabled()) {
expected_error_message_substring =
absl::StrCat("errors resolving ", kFakeName);
} else {
expected_error_message_substring =
absl::StrCat("DNS resolution failed for ", kFakeName);
}
// Don't really care about the deadline - we should quickly hit a DNS
// resolution failure.
gpr_timespec rpc_deadline = grpc_timeout_seconds_to_deadline(100);
int dns_query_timeout_ms = -1; // don't set query timeout
TestCancelDuringActiveQuery(expected_status_code,
expected_error_message_substring, rpc_deadline,
dns_query_timeout_ms, fake_dns_server.port());
if (grpc_core::IsEventEngineDnsEnabled()) {
g_event_engine_grpc_ares_test_only_force_tcp = false;
} else {
g_grpc_ares_test_only_force_tcp = false;
}
}
// This test is meant to repro a bug noticed in internal issue b/297538255.
// The general issue is the loop in
// https://github.com/grpc/grpc/blob/f6a994229e72bc771963706de7a0cd8aa9150bb6/src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc#L371.
// The problem with that loop is that c-ares *can* in certain situations stop
// caring about the fd being processed without reading all of the data out of
// the read buffer. In that case, we keep looping because
// IsFdStillReadableLocked() keeps returning true, but we never make progress.
// Meanwhile, we are holding a lock which prevents cancellation or timeouts from
// kicking in, and thus we spin-loop forever.
//
// At the time of writing, this test case illustrates one way to hit that bug.
// It works as follows:
// 1) We force c-ares to use TCP for its DNS queries
// 2) We stand up a fake DNS server that, for each incoming connection, sends
// three all-zero bytes and then closes the socket.
// 3) When the c-ares library receives the three-zero-byte response from the
// DNS server, it parses the first two-bytes as a length field:
// https://github.com/c-ares/c-ares/blob/6360e96b5cf8e5980c887ce58ef727e53d77243a/src/lib/ares_process.c#L410.
// 4) Because the first two bytes were zero, c-ares attempts to malloc a
// zero-length buffer:
// https://github.com/c-ares/c-ares/blob/6360e96b5cf8e5980c887ce58ef727e53d77243a/src/lib/ares_process.c#L428.
// 5) Because c-ares' default_malloc(0) returns NULL
// (https://github.com/c-ares/c-ares/blob/7f3262312f246556d8c1bdd8ccc1844847f42787/src/lib/ares_library_init.c#L38),
// c-ares invokes handle_error and stops reading on the socket:
// https://github.com/c-ares/c-ares/blob/6360e96b5cf8e5980c887ce58ef727e53d77243a/src/lib/ares_process.c#L430.
// 6) Because we overwrite the socket "close" method, c-ares attempt to close
// the socket in handle_error does nothing except for removing the socket
// from ARES_GETSOCK_READABLE:
// https://github.com/grpc/grpc/blob/f6a994229e72bc771963706de7a0cd8aa9150bb6/src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_ev_driver_posix.cc#L156.
// 7) Because there is still one byte left in the TCP read buffer,
// IsFdStillReadableLocked will keep returning true:
// https://github.com/grpc/grpc/blob/f6a994229e72bc771963706de7a0cd8aa9150bb6/src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_ev_driver_posix.cc#L82.
// But c-ares will never try to read from that socket again, so we have an
// infinite busy loop.
TEST_F(CancelDuringAresQuery, TestQueryFailsWithDataRemainingInReadBuffer) {
if (grpc_core::IsEventEngineDnsEnabled()) {
g_event_engine_grpc_ares_test_only_force_tcp = true;
} else {
g_grpc_ares_test_only_force_tcp = true;
}
grpc_core::testing::SocketUseAfterCloseDetector
socket_use_after_close_detector;
grpc_core::testing::FakeUdpAndTcpServer fake_dns_server(
grpc_core::testing::FakeUdpAndTcpServer::AcceptMode::
kWaitForClientToSendFirstBytes,
grpc_core::testing::FakeUdpAndTcpServer::SendThreeAllZeroBytes);
grpc_status_code expected_status_code = GRPC_STATUS_UNAVAILABLE;
// Don't really care about the deadline - we'll hit a DNS
// resolution failure quickly in any case.
gpr_timespec rpc_deadline = grpc_timeout_seconds_to_deadline(100);
int dns_query_timeout_ms = -1; // don't set query timeout
TestCancelDuringActiveQuery(
expected_status_code, "" /* expected error message substring */,
rpc_deadline, dns_query_timeout_ms, fake_dns_server.port());
if (grpc_core::IsEventEngineDnsEnabled()) {
g_event_engine_grpc_ares_test_only_force_tcp = false;
} else {
g_grpc_ares_test_only_force_tcp = false;
}
}
} // namespace
int main(int argc, char** argv) {
::testing::InitGoogleTest(&argc, argv);
[EventEngine] PosixEventEngine DNS Resolver (#32701) This PR implements a c-ares based DNS resolver for EventEngine with the reference from the original [grpc_ares_wrapper.h](../blob/master/src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.h). The PosixEventEngine DNSResolver is implemented on top of that. Tests which use the client channel resolver API ([resolver.h](../blob/master/src/core/lib/resolver/resolver.h#L54)) are ported, namely the [resolver_component_test.cc](../blob/master/test/cpp/naming/resolver_component_test.cc) and the [cancel_ares_query_test.cc](../blob/master/test/cpp/naming/cancel_ares_query_test.cc). The WindowsEventEngine DNSResolver will use the same EventEngine's grpc_ares_wrapper and will be worked on next. The [resolve_address_test.cc](https://github.com/grpc/grpc/blob/master/test/core/iomgr/resolve_address_test.cc) which uses the iomgr [DNSResolver](../blob/master/src/core/lib/iomgr/resolve_address.h#L44) API has been ported to EventEngine's dns_test.cc. That leaves only 2 tests which use iomgr's API, notably the [dns_resolver_cooldown_test.cc](../blob/master/test/core/client_channel/resolvers/dns_resolver_cooldown_test.cc) and the [goaway_server_test.cc](../blob/master/test/core/end2end/goaway_server_test.cc) which probably need to be restructured to use EventEngine DNSResolver (for one thing they override the original grpc_ares_wrapper's free functions). I will try to tackle these in the next step. <!-- If you know who should review your pull request, please assign it to that person, otherwise the pull request would get assigned randomly. If your pull request is for a specific language, please add the appropriate lang label. -->
1 year ago
grpc::testing::InitTest(&argc, &argv, true);
grpc::testing::TestEnvironment env(&argc, argv);
auto result = RUN_ALL_TESTS();
return result;
}