// Protocol Buffers - Google's data interchange format // Copyright 2008 Google Inc. All rights reserved. // // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file or at // https://developers.google.com/open-source/licenses/bsd // This file defines a protocol for running the conformance test suite // in-process. In other words, the suite itself will run in the same process as // the code under test. // // For pros and cons of this approach, please see conformance.proto. #ifndef CONFORMANCE_CONFORMANCE_TEST_H #define CONFORMANCE_CONFORMANCE_TEST_H #include #include #include #include #include #include #include "google/protobuf/descriptor.pb.h" #include "absl/container/btree_map.h" #include "absl/container/flat_hash_set.h" #include "conformance/conformance.pb.h" #include "failure_list_trie_node.h" #include "google/protobuf/descriptor.h" namespace conformance { class ConformanceRequest; class ConformanceResponse; } // namespace conformance namespace protobuf_test_messages { namespace proto3 { class TestAllTypesProto3; } // namespace proto3 } // namespace protobuf_test_messages namespace google { namespace protobuf { class ConformanceTestSuite; class ConformanceTestRunner { public: virtual ~ConformanceTestRunner() = default; // Call to run a single conformance test. // // "len" is the byte length of a serialized conformance.ConformanceRequest. // "input" is a serialized conformance.ConformanceRequest. // "output" should be set to a serialized conformance.ConformanceResponse. // // If there is any error in running the test itself, set "runtime_error" in // the response. virtual void RunTest(const std::string& test_name, uint32_t len, const std::string& input, std::string* output) = 0; }; // Test runner that spawns the process being tested and communicates with it // over a pipe. class ForkPipeRunner : public ConformanceTestRunner { public: // Note: Run() doesn't take ownership of the pointers inside suites. static int Run(int argc, char* argv[], const std::vector& suites); ForkPipeRunner(const std::string& executable, const std::vector& executable_args, bool performance) : child_pid_(-1), executable_(executable), executable_args_(executable_args), performance_(performance) {} explicit ForkPipeRunner(const std::string& executable) : child_pid_(-1), executable_(executable) {} ~ForkPipeRunner() override = default; void RunTest(const std::string& test_name, uint32_t len, const std::string& request, std::string* response) override; private: void SpawnTestProgram(); void CheckedWrite(int fd, const void* buf, size_t len); bool TryRead(int fd, void* buf, size_t len); void CheckedRead(int fd, void* buf, size_t len); int write_fd_; int read_fd_; pid_t child_pid_; std::string executable_; const std::vector executable_args_; bool performance_; std::string current_test_name_; }; // Class representing the test suite itself. To run it, implement your own // class derived from ConformanceTestRunner, class derived from // ConformanceTestSuite and then write code like: // // class MyConformanceTestSuite : public ConformanceTestSuite { // public: // void RunSuiteImpl() { // // INSERT ACTUAL TESTS. // } // }; // // class MyConformanceTestRunner : public ConformanceTestRunner { // public: // static int Run(int argc, char *argv[], // ConformanceTestSuite* suite); // // private: // virtual void RunTest(...) { // // INSERT YOUR FRAMEWORK-SPECIFIC CODE HERE. // } // }; // // int main() { // MyConformanceTestSuite suite; // MyConformanceTestRunner::Run(argc, argv, &suite); // } // class ConformanceTestSuite { public: ConformanceTestSuite() = default; virtual ~ConformanceTestSuite() = default; void SetPerformance(bool performance) { performance_ = performance; } void SetVerbose(bool verbose) { verbose_ = verbose; } // Whether to require the testee to pass RECOMMENDED tests. By default failing // a RECOMMENDED test case will not fail the entire suite but will only // generated a warning. If this flag is set to true, RECOMMENDED tests will // be treated the same way as REQUIRED tests and failing a RECOMMENDED test // case will cause the entire test suite to fail as well. An implementation // can enable this if it wants to be strictly conforming to protobuf spec. // See the comments about ConformanceLevel below to learn more about the // difference between REQUIRED and RECOMMENDED test cases. void SetEnforceRecommended(bool value) { enforce_recommended_ = value; } // Sets the maximum edition (inclusive) that should be tests for conformance. void SetMaximumEdition(Edition edition) { maximum_edition_ = edition; } // Gets the flag name to the failure list file. // By default, this would return --failure_list std::string GetFailureListFlagName() { return failure_list_flag_name_; } void SetFailureListFlagName(const std::string& failure_list_flag_name) { failure_list_flag_name_ = failure_list_flag_name; } // Sets the path of the output directory. void SetOutputDir(const std::string& output_dir) { output_dir_ = output_dir; } // Sets if we are running the test in debug mode. void SetDebug(bool debug) { debug_ = debug; } // Sets if we are running ONLY the tests provided in the 'names_to_test_' set. void SetIsolated(bool isolated) { isolated_ = isolated; } // Sets the file path of the testee. void SetTestee(const std::string& testee) { testee_ = testee; } // Sets the names of tests to ONLY be run isolated from all the others. void SetNamesToTest(absl::flat_hash_set names_to_test) { names_to_test_ = std::move(names_to_test); } absl::flat_hash_set GetExpectedTestsNotRun() { return names_to_test_; } // Run all the conformance tests against the given test runner. // Test output will be stored in "output". // // Returns true if the set of failing tests was exactly the same as the // failure list. // The filename here is *only* used to create/format useful error messages for // how to update the failure list. We do NOT read this file at all. bool RunSuite(ConformanceTestRunner* runner, std::string* output, const std::string& filename, conformance::FailureSet* failure_list); protected: // Test cases are classified into a few categories: // REQUIRED: the test case must be passed for an implementation to be // interoperable with other implementations. For example, a // parser implementation must accept both packed and unpacked // form of repeated primitive fields. // RECOMMENDED: the test case is not required for the implementation to // be interoperable with other implementations, but is // recommended for best performance and compatibility. For // example, a proto3 serializer should serialize repeated // primitive fields in packed form, but an implementation // failing to do so will still be able to communicate with // other implementations. enum ConformanceLevel { REQUIRED = 0, RECOMMENDED = 1, }; class ConformanceRequestSetting { public: ConformanceRequestSetting(ConformanceLevel level, conformance::WireFormat input_format, conformance::WireFormat output_format, conformance::TestCategory test_category, const Message& prototype_message, const std::string& test_name, const std::string& input); virtual ~ConformanceRequestSetting() = default; std::unique_ptr NewTestMessage() const; std::string GetSyntaxIdentifier() const; std::string GetTestName() const; const conformance::ConformanceRequest& GetRequest() const { return request_; } ConformanceLevel GetLevel() const { return level_; } std::string ConformanceLevelToString(ConformanceLevel level) const; void SetPrintUnknownFields(bool print_unknown_fields) { request_.set_print_unknown_fields(true); } void SetPrototypeMessageForCompare(const Message& message) { prototype_message_for_compare_.reset(message.New()); } protected: virtual std::string InputFormatString(conformance::WireFormat format) const; virtual std::string OutputFormatString( conformance::WireFormat format) const; conformance::ConformanceRequest request_; private: ConformanceLevel level_; ::conformance::WireFormat input_format_; ::conformance::WireFormat output_format_; const Message& prototype_message_; std::unique_ptr prototype_message_for_compare_; std::string test_name_; }; std::string WireFormatToString(conformance::WireFormat wire_format); // Parse payload in the response to the given message. Returns true on // success. virtual bool ParseResponse(const conformance::ConformanceResponse& response, const ConformanceRequestSetting& setting, Message* test_message) = 0; void VerifyResponse(const ConformanceRequestSetting& setting, const std::string& equivalent_wire_format, const conformance::ConformanceResponse& response, bool need_report_success, bool require_same_wire_format); void TruncateDebugPayload(std::string* payload); conformance::ConformanceRequest TruncateRequest( const conformance::ConformanceRequest& request); conformance::ConformanceResponse TruncateResponse( const conformance::ConformanceResponse& response); void ReportSuccess(const conformance::TestStatus& test); void ReportFailure(conformance::TestStatus& test, ConformanceLevel level, const conformance::ConformanceRequest& request, const conformance::ConformanceResponse& response); void ReportSkip(const conformance::TestStatus& test, const conformance::ConformanceRequest& request, const conformance::ConformanceResponse& response); void RunValidInputTest(const ConformanceRequestSetting& setting, const std::string& equivalent_text_format); void RunValidBinaryInputTest(const ConformanceRequestSetting& setting, const std::string& equivalent_wire_format, bool require_same_wire_format = false); // Returns true if our runner_ ran the test and false if it did not. bool RunTest(const std::string& test_name, const conformance::ConformanceRequest& request, conformance::ConformanceResponse* response); // Will return false if an entry from the failure list was either a // duplicate of an already added one to the trie or it contained invalid // wildcards; otherwise, returns true. bool AddExpectedFailedTest(const conformance::TestStatus& failure); virtual void RunSuiteImpl() = 0; ConformanceTestRunner* runner_; FailureListTrieNode failure_list_root_; std::string testee_; int successes_; int expected_failures_; bool verbose_ = false; bool performance_ = false; bool enforce_recommended_ = false; Edition maximum_edition_ = Edition::EDITION_PROTO3; std::string output_; std::string output_dir_; std::string failure_list_flag_name_ = "--failure_list"; std::string failure_list_filename_; absl::flat_hash_set names_to_test_; bool debug_ = false; // If names were given for names_to_test_, only those tests // will be run and this bool will be set to true. bool isolated_ = false; // The set of test names (expanded from wildcard(s) and non-expanded) that are // expected to fail in this run, but haven't failed yet. absl::btree_map expected_to_fail_; // The set of tests that failed because their failure message did not match // the actual failure message. These are failure messages that may need to be // removed from our failure lists. absl::btree_map expected_failure_messages_; // The set of test names that have been run. Used to ensure that there are no // duplicate names in the suite. absl::flat_hash_set test_names_ran_; // The set of tests that failed, but weren't expected to: They weren't // present in our failure lists. absl::btree_map unexpected_failing_tests_; // The set of tests that succeeded, but weren't expected to: They were present // in our failure lists, but managed to succeed. absl::btree_map unexpected_succeeding_tests_; // The set of tests that failed because their failure message did not match // the actual failure message. These are failure messages that may need to be // added to our failure lists. absl::btree_map unexpected_failure_messages_; // The set of test names (wildcarded or not) from the failure list that did // not match any actual test name. absl::btree_map unmatched_; // The set of tests that the testee opted out of; absl::btree_map skipped_; // Allows us to remove from unmatched_. absl::btree_map saved_failure_messages_; // If a failure list entry served as a match for more than 'max_matches_', // those will be added here for removal. absl::btree_map exceeded_max_matches_; // Keeps track of how many tests matched to each failure list entry. absl::btree_map number_of_matches_; }; } // namespace protobuf } // namespace google #endif // CONFORMANCE_CONFORMANCE_TEST_H