// Copyright 2005, Google Inc.
// All rights reserved.
//
// Redistribution and use in source and binary forms, with or without
// modification, are permitted provided that the following conditions are
// met:
//
//     * Redistributions of source code must retain the above copyright
// notice, this list of conditions and the following disclaimer.
//     * Redistributions in binary form must reproduce the above
// copyright notice, this list of conditions and the following disclaimer
// in the documentation and/or other materials provided with the
// distribution.
//     * Neither the name of Google Inc. nor the names of its
// contributors may be used to endorse or promote products derived from
// this software without specific prior written permission.
//
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

//
// The Google C++ Testing and Mocking Framework (Google Test)

#include "gtest/gtest.h"
#include "gtest/internal/custom/gtest.h"
#include "gtest/gtest-spi.h"

#include <ctype.h>
#include <math.h>
#include <stdarg.h>
#include <stdio.h>
#include <stdlib.h>
#include <time.h>
#include <wchar.h>
#include <wctype.h>

#include <algorithm>
#include <cstdint>
#include <iomanip>
#include <limits>
#include <list>
#include <map>
#include <ostream> // NOLINT
#include <sstream>
#include <vector>

#if GTEST_OS_LINUX

#define GTEST_HAS_GETTIMEOFDAY_ 1

#include <fcntl.h> // NOLINT
#include <limits.h> // NOLINT
#include <sched.h> // NOLINT
// Declares vsnprintf().  This header is not available on Windows.
#include <strings.h> // NOLINT
#include <sys/mman.h> // NOLINT
#include <sys/time.h> // NOLINT
#include <unistd.h> // NOLINT
#include <string>

#elif GTEST_OS_ZOS
#define GTEST_HAS_GETTIMEOFDAY_ 1
#include <sys/time.h> // NOLINT

// On z/OS we additionally need strings.h for strcasecmp.
#include <strings.h> // NOLINT

#elif GTEST_OS_WINDOWS_MOBILE // We are on Windows CE.

#include <windows.h> // NOLINT
#undef min

#elif GTEST_OS_WINDOWS // We are on Windows proper.

#include <windows.h> // NOLINT
#undef min

#ifdef _MSC_VER
#include <crtdbg.h> // NOLINT
#endif

#include <io.h> // NOLINT
#include <sys/timeb.h> // NOLINT
#include <sys/types.h> // NOLINT
#include <sys/stat.h> // NOLINT

#if GTEST_OS_WINDOWS_MINGW
// MinGW has gettimeofday() but not _ftime64().
#define GTEST_HAS_GETTIMEOFDAY_ 1
#include <sys/time.h> // NOLINT
#endif // GTEST_OS_WINDOWS_MINGW

#else

// Assume other platforms have gettimeofday().
#define GTEST_HAS_GETTIMEOFDAY_ 1

// cpplint thinks that the header is already included, so we want to
// silence it.
#include <sys/time.h> // NOLINT
#include <unistd.h> // NOLINT

#endif // GTEST_OS_LINUX

#if GTEST_HAS_EXCEPTIONS
#include <stdexcept>
#endif

#if GTEST_CAN_STREAM_RESULTS_
#include <arpa/inet.h> // NOLINT
#include <netdb.h> // NOLINT
#include <sys/socket.h> // NOLINT
#include <sys/types.h> // NOLINT
#endif

#include "src/gtest-internal-inl.h"

#if GTEST_OS_WINDOWS
#define vsnprintf _vsnprintf
#endif // GTEST_OS_WINDOWS

#if GTEST_OS_MAC
#ifndef GTEST_OS_IOS
#include <crt_externs.h>
#endif
#endif

#if GTEST_HAS_ABSL
#include "absl/debugging/failure_signal_handler.h"
#include "absl/debugging/stacktrace.h"
#include "absl/debugging/symbolize.h"
#include "absl/strings/str_cat.h"
#endif // GTEST_HAS_ABSL

namespace testing {

using internal::CountIf;
using internal::ForEach;
using internal::GetElementOr;
using internal::Shuffle;

// Constants.

// A test whose test suite name or test name matches this filter is
// disabled and not run.
static const char kDisableTestFilter[] = "DISABLED_*:*/DISABLED_*";

// A test suite whose name matches this filter is considered a death
// test suite and will be run before test suites whose name doesn't
// match this filter.
static const char kDeathTestSuiteFilter[] = "*DeathTest:*DeathTest/*";

// A test filter that matches everything.
static const char kUniversalFilter[] = "*";

// The default output format.
static const char kDefaultOutputFormat[] = "xml";
// The default output file.
static const char kDefaultOutputFile[] = "test_detail";

// The environment variable name for the test shard index.
static const char kTestShardIndex[] = "GTEST_SHARD_INDEX";
// The environment variable name for the total number of test shards.
static const char kTestTotalShards[] = "GTEST_TOTAL_SHARDS";
// The environment variable name for the test shard status file.
static const char kTestShardStatusFile[] = "GTEST_SHARD_STATUS_FILE";

namespace internal {

// The text used in failure messages to indicate the start of the
// stack trace.
const char kStackTraceMarker[] = "\nStack trace:\n";

// g_help_flag is true if and only if the --help flag or an equivalent form
// is specified on the command line.
bool g_help_flag = false;

// Utilty function to Open File for Writing
static FILE *OpenFileForWriting(const std::string &output_file)
{
    FILE *fileout = nullptr;
    FilePath output_file_path(output_file);
    FilePath output_dir(output_file_path.RemoveFileName());

    if (output_dir.CreateDirectoriesRecursively()) {
        fileout = posix::FOpen(output_file.c_str(), "w");
    }
    if (fileout == nullptr) {
        GTEST_LOG_(FATAL) << "Unable to open file \"" << output_file << "\"";
    }
    return fileout;
}

} // namespace internal

// Bazel passes in the argument to '--test_filter' via the TESTBRIDGE_TEST_ONLY
// environment variable.
static const char *GetDefaultFilter()
{
    const char *const testbridge_test_only =
        internal::posix::GetEnv("TESTBRIDGE_TEST_ONLY");
    if (testbridge_test_only != nullptr) {
        return testbridge_test_only;
    }
    return kUniversalFilter;
}

// Bazel passes in the argument to '--test_runner_fail_fast' via the
// TESTBRIDGE_TEST_RUNNER_FAIL_FAST environment variable.
static bool GetDefaultFailFast()
{
    const char *const testbridge_test_runner_fail_fast =
        internal::posix::GetEnv("TESTBRIDGE_TEST_RUNNER_FAIL_FAST");
    if (testbridge_test_runner_fail_fast != nullptr) {
        return strcmp(testbridge_test_runner_fail_fast, "1") == 0;
    }
    return false;
}

GTEST_DEFINE_bool_(
    fail_fast, internal::BoolFromGTestEnv("fail_fast", GetDefaultFailFast()),
    "True if and only if a test failure should stop further test execution.");

GTEST_DEFINE_bool_(
    also_run_disabled_tests,
    internal::BoolFromGTestEnv("also_run_disabled_tests", false),
    "Run disabled tests too, in addition to the tests normally being run.");

GTEST_DEFINE_bool_(
    break_on_failure, internal::BoolFromGTestEnv("break_on_failure", false),
    "True if and only if a failed assertion should be a debugger "
    "break-point.");

GTEST_DEFINE_bool_(catch_exceptions,
                   internal::BoolFromGTestEnv("catch_exceptions", true),
                   "True if and only if " GTEST_NAME_
                   " should catch exceptions and treat them as test failures.");

GTEST_DEFINE_string_(
    color,
    internal::StringFromGTestEnv("color", "auto"),
    "Whether to use colors in the output.  Valid values: yes, no, "
    "and auto.  'auto' means to use colors if the output is "
    "being sent to a terminal and the TERM environment variable "
    "is set to a terminal type that supports colors.");

GTEST_DEFINE_string_(
    filter,
    internal::StringFromGTestEnv("filter", GetDefaultFilter()),
    "A colon-separated list of glob (not regex) patterns "
    "for filtering the tests to run, optionally followed by a "
    "'-' and a : separated list of negative patterns (tests to "
    "exclude).  A test is run if it matches one of the positive "
    "patterns and does not match any of the negative patterns.");

GTEST_DEFINE_bool_(
    install_failure_signal_handler,
    internal::BoolFromGTestEnv("install_failure_signal_handler", false),
    "If true and supported on the current platform, " GTEST_NAME_ " should "
    "install a signal handler that dumps debugging information when fatal "
    "signals are raised.");

GTEST_DEFINE_bool_(list_tests, false,
                   "List all tests without running them.");

// The net priority order after flag processing is thus:
//   --gtest_output command line flag
//   GTEST_OUTPUT environment variable
//   XML_OUTPUT_FILE environment variable
//   ''
GTEST_DEFINE_string_(
    output,
    internal::StringFromGTestEnv("output",
                                 internal::OutputFlagAlsoCheckEnvVar().c_str()),
    "A format (defaults to \"xml\" but can be specified to be \"json\"), "
    "optionally followed by a colon and an output file name or directory. "
    "A directory is indicated by a trailing pathname separator. "
    "Examples: \"xml:filename.xml\", \"xml::directoryname/\". "
    "If a directory is specified, output files will be created "
    "within that directory, with file-names based on the test "
    "executable's name and, if necessary, made unique by adding "
    "digits.");

GTEST_DEFINE_bool_(
    brief, internal::BoolFromGTestEnv("brief", false),
    "True if only test failures should be displayed in text output.");

GTEST_DEFINE_bool_(print_time, internal::BoolFromGTestEnv("print_time", true),
                   "True if and only if " GTEST_NAME_
                   " should display elapsed time in text output.");

GTEST_DEFINE_bool_(print_utf8, internal::BoolFromGTestEnv("print_utf8", true),
                   "True if and only if " GTEST_NAME_
                   " prints UTF8 characters as text.");

GTEST_DEFINE_int32_(
    random_seed,
    internal::Int32FromGTestEnv("random_seed", 0),
    "Random number seed to use when shuffling test orders.  Must be in range "
    "[1, 99999], or 0 to use a seed based on the current time.");

GTEST_DEFINE_int32_(
    repeat,
    internal::Int32FromGTestEnv("repeat", 1),
    "How many times to repeat each test.  Specify a negative number "
    "for repeating forever.  Useful for shaking out flaky tests.");

GTEST_DEFINE_bool_(show_internal_stack_frames, false,
                   "True if and only if " GTEST_NAME_
                   " should include internal stack frames when "
                   "printing test failure stack traces.");

GTEST_DEFINE_bool_(shuffle, internal::BoolFromGTestEnv("shuffle", false),
                   "True if and only if " GTEST_NAME_
                   " should randomize tests' order on every run.");

GTEST_DEFINE_int32_(
    stack_trace_depth,
    internal::Int32FromGTestEnv("stack_trace_depth", kMaxStackTraceDepth),
    "The maximum number of stack frames to print when an "
    "assertion fails.  The valid range is 0 through 100, inclusive.");

GTEST_DEFINE_string_(
    stream_result_to,
    internal::StringFromGTestEnv("stream_result_to", ""),
    "This flag specifies the host name and the port number on which to stream "
    "test results. Example: \"localhost:555\". The flag is effective only on "
    "Linux.");

GTEST_DEFINE_bool_(
    throw_on_failure,
    internal::BoolFromGTestEnv("throw_on_failure", false),
    "When this flag is specified, a failed assertion will throw an exception "
    "if exceptions are enabled or exit the program with a non-zero code "
    "otherwise. For use with an external test framework.");

#if GTEST_USE_OWN_FLAGFILE_FLAG_
GTEST_DEFINE_string_(
    flagfile,
    internal::StringFromGTestEnv("flagfile", ""),
    "This flag specifies the flagfile to read command-line flags from.");
#endif // GTEST_USE_OWN_FLAGFILE_FLAG_

namespace internal {

// Generates a random number from [0, range), using a Linear
// Congruential Generator (LCG).  Crashes if 'range' is 0 or greater
// than kMaxRange.
uint32_t Random::Generate(uint32_t range)
{
    // These constants are the same as are used in glibc's rand(3).
    // Use wider types than necessary to prevent unsigned overflow diagnostics.
    state_ = static_cast<uint32_t>(1103515245ULL * state_ + 12345U) % kMaxRange;

    GTEST_CHECK_(range > 0)
        << "Cannot generate a number in the range [0, 0).";
    GTEST_CHECK_(range <= kMaxRange)
        << "Generation of a number in [0, " << range << ") was requested, "
        << "but this can only generate numbers in [0, " << kMaxRange << ").";

    // Converting via modulus introduces a bit of downward bias, but
    // it's simple, and a linear congruential generator isn't too good
    // to begin with.
    return state_ % range;
}

// GTestIsInitialized() returns true if and only if the user has initialized
// Google Test.  Useful for catching the user mistake of not initializing
// Google Test before calling RUN_ALL_TESTS().
static bool GTestIsInitialized()
{
    return GetArgvs().size() > 0;
}

// Iterates over a vector of TestSuites, keeping a running sum of the
// results of calling a given int-returning method on each.
// Returns the sum.
static int SumOverTestSuiteList(const std::vector<TestSuite *> &case_list,
                                int (TestSuite::*method)() const)
{
    int sum = 0;
    for (size_t i = 0; i < case_list.size(); i++) {
        sum += (case_list[i]->*method)();
    }
    return sum;
}

// Returns true if and only if the test suite passed.
static bool TestSuitePassed(const TestSuite *test_suite)
{
    return test_suite->should_run() && test_suite->Passed();
}

// Returns true if and only if the test suite failed.
static bool TestSuiteFailed(const TestSuite *test_suite)
{
    return test_suite->should_run() && test_suite->Failed();
}

// Returns true if and only if test_suite contains at least one test that
// should run.
static bool ShouldRunTestSuite(const TestSuite *test_suite)
{
    return test_suite->should_run();
}

// AssertHelper constructor.
AssertHelper::AssertHelper(TestPartResult::Type type,
                           const char *file,
                           int line,
                           const char *message)
    : data_(new AssertHelperData(type, file, line, message))
{
}

AssertHelper::~AssertHelper()
{
    delete data_;
}

// Message assignment, for assertion streaming support.
void AssertHelper::operator=(const Message &message) const
{
    UnitTest::GetInstance()->AddTestPartResult(data_->type, data_->file, data_->line,
                                               AppendUserMessage(data_->message, message),
                                               UnitTest::GetInstance()->impl()->CurrentOsStackTraceExceptTop(1)
                                               // Skips the stack frame for this function itself.
    ); // NOLINT
}

namespace {

// When TEST_P is found without a matching INSTANTIATE_TEST_SUITE_P
// to creates test cases for it, a syntetic test case is
// inserted to report ether an error or a log message.
//
// This configuration bit will likely be removed at some point.
constexpr bool kErrorOnUninstantiatedParameterizedTest = true;
constexpr bool kErrorOnUninstantiatedTypeParameterizedTest = true;

// A test that fails at a given file/line location with a given message.
class FailureTest : public Test
{
public:
    explicit FailureTest(const CodeLocation &loc, std::string error_message,
                         bool as_error)
        : loc_(loc)
        , error_message_(std::move(error_message))
        , as_error_(as_error)
    {
    }

    void TestBody() override
    {
        if (as_error_) {
            AssertHelper(TestPartResult::kNonFatalFailure, loc_.file.c_str(),
                         loc_.line, "") = Message() << error_message_;
        } else {
            std::cout << error_message_ << std::endl;
        }
    }

private:
    const CodeLocation loc_;
    const std::string error_message_;
    const bool as_error_;
};

} // namespace

std::set<std::string> *GetIgnoredParameterizedTestSuites()
{
    return UnitTest::GetInstance()->impl()->ignored_parameterized_test_suites();
}

// Add a given test_suit to the list of them allow to go un-instantiated.
MarkAsIgnored::MarkAsIgnored(const char *test_suite)
{
    GetIgnoredParameterizedTestSuites()->insert(test_suite);
}

// If this parameterized test suite has no instantiations (and that
// has not been marked as okay), emit a test case reporting that.
void InsertSyntheticTestCase(const std::string &name, CodeLocation location,
                             bool has_test_p)
{
    const auto &ignored = *GetIgnoredParameterizedTestSuites();
    if (ignored.find(name) != ignored.end())
        return;

    const char kMissingInstantiation[] = //
        " is defined via TEST_P, but never instantiated. None of the test cases "
        "will run. Either no INSTANTIATE_TEST_SUITE_P is provided or the only "
        "ones provided expand to nothing."
        "\n\n"
        "Ideally, TEST_P definitions should only ever be included as part of "
        "binaries that intend to use them. (As opposed to, for example, being "
        "placed in a library that may be linked in to get other utilities.)";

    const char kMissingTestCase[] = //
        " is instantiated via INSTANTIATE_TEST_SUITE_P, but no tests are "
        "defined via TEST_P . No test cases will run."
        "\n\n"
        "Ideally, INSTANTIATE_TEST_SUITE_P should only ever be invoked from "
        "code that always depend on code that provides TEST_P. Failing to do "
        "so is often an indication of dead code, e.g. the last TEST_P was "
        "removed but the rest got left behind.";

    std::string message =
        "Parameterized test suite " + name + (has_test_p ? kMissingInstantiation : kMissingTestCase) + "\n\n"
                                                                                                       "To suppress this error for this test suite, insert the following line "
                                                                                                       "(in a non-header) in the namespace it is defined in:"
                                                                                                       "\n\n"
                                                                                                       "GTEST_ALLOW_UNINSTANTIATED_PARAMETERIZED_TEST("
        + name + ");";

    std::string full_name = "UninstantiatedParameterizedTestSuite<" + name + ">";
    RegisterTest( //
        "GoogleTestVerification", full_name.c_str(),
        nullptr, // No type parameter.
        nullptr, // No value parameter.
        location.file.c_str(), location.line, [message, location] {
            return new FailureTest(location, message,
                                   kErrorOnUninstantiatedParameterizedTest);
        });
}

void RegisterTypeParameterizedTestSuite(const char *test_suite_name,
                                        CodeLocation code_location)
{
    GetUnitTestImpl()->type_parameterized_test_registry().RegisterTestSuite(
        test_suite_name, code_location);
}

void RegisterTypeParameterizedTestSuiteInstantiation(const char *case_name)
{
    GetUnitTestImpl()
        ->type_parameterized_test_registry()
        .RegisterInstantiation(case_name);
}

void TypeParameterizedTestSuiteRegistry::RegisterTestSuite(
    const char *test_suite_name, CodeLocation code_location)
{
    suites_.emplace(std::string(test_suite_name),
                    TypeParameterizedTestSuiteInfo(code_location));
}

void TypeParameterizedTestSuiteRegistry::RegisterInstantiation(
    const char *test_suite_name)
{
    auto it = suites_.find(std::string(test_suite_name));
    if (it != suites_.end()) {
        it->second.instantiated = true;
    } else {
        GTEST_LOG_(ERROR) << "Unknown type parameterized test suit '"
                          << test_suite_name << "'";
    }
}

void TypeParameterizedTestSuiteRegistry::CheckForInstantiations()
{
    const auto &ignored = *GetIgnoredParameterizedTestSuites();
    for (const auto &testcase : suites_) {
        if (testcase.second.instantiated)
            continue;
        if (ignored.find(testcase.first) != ignored.end())
            continue;

        std::string message =
            "Type parameterized test suite " + testcase.first + " is defined via REGISTER_TYPED_TEST_SUITE_P, but never instantiated "
                                                                "via INSTANTIATE_TYPED_TEST_SUITE_P. None of the test cases will run."
                                                                "\n\n"
                                                                "Ideally, TYPED_TEST_P definitions should only ever be included as "
                                                                "part of binaries that intend to use them. (As opposed to, for "
                                                                "example, being placed in a library that may be linked in to get other "
                                                                "utilities.)"
                                                                "\n\n"
                                                                "To suppress this error for this test suite, insert the following line "
                                                                "(in a non-header) in the namespace it is definedin in:"
                                                                "\n\n"
                                                                "GTEST_ALLOW_UNINSTANTIATED_PARAMETERIZED_TEST("
            + testcase.first + ");";

        std::string full_name =
            "UninstantiatedTypeParameterizedTestSuite<" + testcase.first + ">";
        RegisterTest( //
            "GoogleTestVerification", full_name.c_str(),
            nullptr, // No type parameter.
            nullptr, // No value parameter.
            testcase.second.code_location.file.c_str(),
            testcase.second.code_location.line, [message, testcase] {
                return new FailureTest(testcase.second.code_location, message,
                                       kErrorOnUninstantiatedTypeParameterizedTest);
            });
    }
}

// A copy of all command line arguments.  Set by InitGoogleTest().
static ::std::vector<std::string> g_argvs;

::std::vector<std::string> GetArgvs()
{
#if defined(GTEST_CUSTOM_GET_ARGVS_)
    // GTEST_CUSTOM_GET_ARGVS_() may return a container of std::string or
    // ::string. This code converts it to the appropriate type.
    const auto &custom = GTEST_CUSTOM_GET_ARGVS_();
    return ::std::vector<std::string>(custom.begin(), custom.end());
#else // defined(GTEST_CUSTOM_GET_ARGVS_)
    return g_argvs;
#endif // defined(GTEST_CUSTOM_GET_ARGVS_)
}

// Returns the current application's name, removing directory path if that
// is present.
FilePath GetCurrentExecutableName()
{
    FilePath result;

#if GTEST_OS_WINDOWS || GTEST_OS_OS2
    result.Set(FilePath(GetArgvs()[0]).RemoveExtension("exe"));
#else
    result.Set(FilePath(GetArgvs()[0]));
#endif // GTEST_OS_WINDOWS

    return result.RemoveDirectoryName();
}

// Functions for processing the gtest_output flag.

// Returns the output format, or "" for normal printed output.
std::string UnitTestOptions::GetOutputFormat()
{
    const char *const gtest_output_flag = GTEST_FLAG(output).c_str();
    const char *const colon = strchr(gtest_output_flag, ':');
    return (colon == nullptr)
               ? std::string(gtest_output_flag)
               : std::string(gtest_output_flag,
                             static_cast<size_t>(colon - gtest_output_flag));
}

// Returns the name of the requested output file, or the default if none
// was explicitly specified.
std::string UnitTestOptions::GetAbsolutePathToOutputFile()
{
    const char *const gtest_output_flag = GTEST_FLAG(output).c_str();

    std::string format = GetOutputFormat();
    if (format.empty())
        format = std::string(kDefaultOutputFormat);

    const char *const colon = strchr(gtest_output_flag, ':');
    if (colon == nullptr)
        return internal::FilePath::MakeFileName(
                   internal::FilePath(
                       UnitTest::GetInstance()->original_working_dir()),
                   internal::FilePath(kDefaultOutputFile), 0,
                   format.c_str())
            .string();

    internal::FilePath output_name(colon + 1);
    if (!output_name.IsAbsolutePath())
        output_name = internal::FilePath::ConcatPaths(
            internal::FilePath(UnitTest::GetInstance()->original_working_dir()),
            internal::FilePath(colon + 1));

    if (!output_name.IsDirectory())
        return output_name.string();

    internal::FilePath result(internal::FilePath::GenerateUniqueFileName(
        output_name, internal::GetCurrentExecutableName(),
        GetOutputFormat().c_str()));
    return result.string();
}

// Returns true if and only if the wildcard pattern matches the string.
// The first ':' or '\0' character in pattern marks the end of it.
//
// This recursive algorithm isn't very efficient, but is clear and
// works well enough for matching test names, which are short.
bool UnitTestOptions::PatternMatchesString(const char *pattern,
                                           const char *str)
{
    switch (*pattern) {
    case '\0':
    case ':': // Either ':' or '\0' marks the end of the pattern.
        return *str == '\0';
    case '?': // Matches any single character.
        return *str != '\0' && PatternMatchesString(pattern + 1, str + 1);
    case '*': // Matches any string (possibly empty) of characters.
        return (*str != '\0' && PatternMatchesString(pattern, str + 1)) || PatternMatchesString(pattern + 1, str);
    default: // Non-special character.  Matches itself.
        return *pattern == *str && PatternMatchesString(pattern + 1, str + 1);
    }
}

bool UnitTestOptions::MatchesFilter(
    const std::string &name, const char *filter)
{
    const char *cur_pattern = filter;
    for (;;) {
        if (PatternMatchesString(cur_pattern, name.c_str())) {
            return true;
        }

        // Finds the next pattern in the filter.
        cur_pattern = strchr(cur_pattern, ':');

        // Returns if no more pattern can be found.
        if (cur_pattern == nullptr) {
            return false;
        }

        // Skips the pattern separater (the ':' character).
        cur_pattern++;
    }
}

// Returns true if and only if the user-specified filter matches the test
// suite name and the test name.
bool UnitTestOptions::FilterMatchesTest(const std::string &test_suite_name,
                                        const std::string &test_name)
{
    const std::string &full_name = test_suite_name + "." + test_name.c_str();

    // Split --gtest_filter at '-', if there is one, to separate into
    // positive filter and negative filter portions
    const char *const p = GTEST_FLAG(filter).c_str();
    const char *const dash = strchr(p, '-');
    std::string positive;
    std::string negative;
    if (dash == nullptr) {
        positive = GTEST_FLAG(filter).c_str(); // Whole string is a positive filter
        negative = "";
    } else {
        positive = std::string(p, dash); // Everything up to the dash
        negative = std::string(dash + 1); // Everything after the dash
        if (positive.empty()) {
            // Treat '-test1' as the same as '*-test1'
            positive = kUniversalFilter;
        }
    }

    // A filter is a colon-separated list of patterns.  It matches a
    // test if any pattern in it matches the test.
    return (MatchesFilter(full_name, positive.c_str()) && !MatchesFilter(full_name, negative.c_str()));
}

#if GTEST_HAS_SEH
// Returns EXCEPTION_EXECUTE_HANDLER if Google Test should handle the
// given SEH exception, or EXCEPTION_CONTINUE_SEARCH otherwise.
// This function is useful as an __except condition.
int UnitTestOptions::GTestShouldProcessSEH(DWORD exception_code)
{
    // Google Test should handle a SEH exception if:
    //   1. the user wants it to, AND
    //   2. this is not a breakpoint exception, AND
    //   3. this is not a C++ exception (VC++ implements them via SEH,
    //      apparently).
    //
    // SEH exception code for C++ exceptions.
    // (see http://support.microsoft.com/kb/185294 for more information).
    const DWORD kCxxExceptionCode = 0xe06d7363;

    bool should_handle = true;

    if (!GTEST_FLAG(catch_exceptions))
        should_handle = false;
    else if (exception_code == EXCEPTION_BREAKPOINT)
        should_handle = false;
    else if (exception_code == kCxxExceptionCode)
        should_handle = false;

    return should_handle ? EXCEPTION_EXECUTE_HANDLER : EXCEPTION_CONTINUE_SEARCH;
}
#endif // GTEST_HAS_SEH

} // namespace internal

// The c'tor sets this object as the test part result reporter used by
// Google Test.  The 'result' parameter specifies where to report the
// results. Intercepts only failures from the current thread.
ScopedFakeTestPartResultReporter::ScopedFakeTestPartResultReporter(
    TestPartResultArray *result)
    : intercept_mode_(INTERCEPT_ONLY_CURRENT_THREAD)
    , result_(result)
{
    Init();
}

// The c'tor sets this object as the test part result reporter used by
// Google Test.  The 'result' parameter specifies where to report the
// results.
ScopedFakeTestPartResultReporter::ScopedFakeTestPartResultReporter(
    InterceptMode intercept_mode, TestPartResultArray *result)
    : intercept_mode_(intercept_mode)
    , result_(result)
{
    Init();
}

void ScopedFakeTestPartResultReporter::Init()
{
    internal::UnitTestImpl *const impl = internal::GetUnitTestImpl();
    if (intercept_mode_ == INTERCEPT_ALL_THREADS) {
        old_reporter_ = impl->GetGlobalTestPartResultReporter();
        impl->SetGlobalTestPartResultReporter(this);
    } else {
        old_reporter_ = impl->GetTestPartResultReporterForCurrentThread();
        impl->SetTestPartResultReporterForCurrentThread(this);
    }
}

// The d'tor restores the test part result reporter used by Google Test
// before.
ScopedFakeTestPartResultReporter::~ScopedFakeTestPartResultReporter()
{
    internal::UnitTestImpl *const impl = internal::GetUnitTestImpl();
    if (intercept_mode_ == INTERCEPT_ALL_THREADS) {
        impl->SetGlobalTestPartResultReporter(old_reporter_);
    } else {
        impl->SetTestPartResultReporterForCurrentThread(old_reporter_);
    }
}

// Increments the test part result count and remembers the result.
// This method is from the TestPartResultReporterInterface interface.
void ScopedFakeTestPartResultReporter::ReportTestPartResult(
    const TestPartResult &result)
{
    result_->Append(result);
}

namespace internal {

// Returns the type ID of ::testing::Test.  We should always call this
// instead of GetTypeId< ::testing::Test>() to get the type ID of
// testing::Test.  This is to work around a suspected linker bug when
// using Google Test as a framework on Mac OS X.  The bug causes
// GetTypeId< ::testing::Test>() to return different values depending
// on whether the call is from the Google Test framework itself or
// from user test code.  GetTestTypeId() is guaranteed to always
// return the same value, as it always calls GetTypeId<>() from the
// gtest.cc, which is within the Google Test framework.
TypeId GetTestTypeId()
{
    return GetTypeId<Test>();
}

// The value of GetTestTypeId() as seen from within the Google Test
// library.  This is solely for testing GetTestTypeId().
extern const TypeId kTestTypeIdInGoogleTest = GetTestTypeId();

// This predicate-formatter checks that 'results' contains a test part
// failure of the given type and that the failure message contains the
// given substring.
static AssertionResult HasOneFailure(const char * /* results_expr */,
                                     const char * /* type_expr */,
                                     const char * /* substr_expr */,
                                     const TestPartResultArray &results,
                                     TestPartResult::Type type,
                                     const std::string &substr)
{
    const std::string expected(type == TestPartResult::kFatalFailure ? "1 fatal failure" : "1 non-fatal failure");
    Message msg;
    if (results.size() != 1) {
        msg << "Expected: " << expected << "\n"
            << "  Actual: " << results.size() << " failures";
        for (int i = 0; i < results.size(); i++) {
            msg << "\n"
                << results.GetTestPartResult(i);
        }
        return AssertionFailure() << msg;
    }

    const TestPartResult &r = results.GetTestPartResult(0);
    if (r.type() != type) {
        return AssertionFailure() << "Expected: " << expected << "\n"
                                  << "  Actual:\n"
                                  << r;
    }

    if (strstr(r.message(), substr.c_str()) == nullptr) {
        return AssertionFailure() << "Expected: " << expected << " containing \""
                                  << substr << "\"\n"
                                  << "  Actual:\n"
                                  << r;
    }

    return AssertionSuccess();
}

// The constructor of SingleFailureChecker remembers where to look up
// test part results, what type of failure we expect, and what
// substring the failure message should contain.
SingleFailureChecker::SingleFailureChecker(const TestPartResultArray *results,
                                           TestPartResult::Type type,
                                           const std::string &substr)
    : results_(results)
    , type_(type)
    , substr_(substr)
{
}

// The destructor of SingleFailureChecker verifies that the given
// TestPartResultArray contains exactly one failure that has the given
// type and contains the given substring.  If that's not the case, a
// non-fatal failure will be generated.
SingleFailureChecker::~SingleFailureChecker()
{
    EXPECT_PRED_FORMAT3(HasOneFailure, *results_, type_, substr_);
}

DefaultGlobalTestPartResultReporter::DefaultGlobalTestPartResultReporter(
    UnitTestImpl *unit_test)
    : unit_test_(unit_test)
{
}

void DefaultGlobalTestPartResultReporter::ReportTestPartResult(
    const TestPartResult &result)
{
    unit_test_->current_test_result()->AddTestPartResult(result);
    unit_test_->listeners()->repeater()->OnTestPartResult(result);
}

DefaultPerThreadTestPartResultReporter::DefaultPerThreadTestPartResultReporter(
    UnitTestImpl *unit_test)
    : unit_test_(unit_test)
{
}

void DefaultPerThreadTestPartResultReporter::ReportTestPartResult(
    const TestPartResult &result)
{
    unit_test_->GetGlobalTestPartResultReporter()->ReportTestPartResult(result);
}

// Returns the global test part result reporter.
TestPartResultReporterInterface *
UnitTestImpl::GetGlobalTestPartResultReporter()
{
    internal::MutexLock lock(&global_test_part_result_reporter_mutex_);
    return global_test_part_result_repoter_;
}

// Sets the global test part result reporter.
void UnitTestImpl::SetGlobalTestPartResultReporter(
    TestPartResultReporterInterface *reporter)
{
    internal::MutexLock lock(&global_test_part_result_reporter_mutex_);
    global_test_part_result_repoter_ = reporter;
}

// Returns the test part result reporter for the current thread.
TestPartResultReporterInterface *
UnitTestImpl::GetTestPartResultReporterForCurrentThread()
{
    return per_thread_test_part_result_reporter_.get();
}

// Sets the test part result reporter for the current thread.
void UnitTestImpl::SetTestPartResultReporterForCurrentThread(
    TestPartResultReporterInterface *reporter)
{
    per_thread_test_part_result_reporter_.set(reporter);
}

// Gets the number of successful test suites.
int UnitTestImpl::successful_test_suite_count() const
{
    return CountIf(test_suites_, TestSuitePassed);
}

// Gets the number of failed test suites.
int UnitTestImpl::failed_test_suite_count() const
{
    return CountIf(test_suites_, TestSuiteFailed);
}

// Gets the number of all test suites.
int UnitTestImpl::total_test_suite_count() const
{
    return static_cast<int>(test_suites_.size());
}

// Gets the number of all test suites that contain at least one test
// that should run.
int UnitTestImpl::test_suite_to_run_count() const
{
    return CountIf(test_suites_, ShouldRunTestSuite);
}

// Gets the number of successful tests.
int UnitTestImpl::successful_test_count() const
{
    return SumOverTestSuiteList(test_suites_, &TestSuite::successful_test_count);
}

// Gets the number of skipped tests.
int UnitTestImpl::skipped_test_count() const
{
    return SumOverTestSuiteList(test_suites_, &TestSuite::skipped_test_count);
}

// Gets the number of failed tests.
int UnitTestImpl::failed_test_count() const
{
    return SumOverTestSuiteList(test_suites_, &TestSuite::failed_test_count);
}

// Gets the number of disabled tests that will be reported in the XML report.
int UnitTestImpl::reportable_disabled_test_count() const
{
    return SumOverTestSuiteList(test_suites_,
                                &TestSuite::reportable_disabled_test_count);
}

// Gets the number of disabled tests.
int UnitTestImpl::disabled_test_count() const
{
    return SumOverTestSuiteList(test_suites_, &TestSuite::disabled_test_count);
}

// Gets the number of tests to be printed in the XML report.
int UnitTestImpl::reportable_test_count() const
{
    return SumOverTestSuiteList(test_suites_, &TestSuite::reportable_test_count);
}

// Gets the number of all tests.
int UnitTestImpl::total_test_count() const
{
    return SumOverTestSuiteList(test_suites_, &TestSuite::total_test_count);
}

// Gets the number of tests that should run.
int UnitTestImpl::test_to_run_count() const
{
    return SumOverTestSuiteList(test_suites_, &TestSuite::test_to_run_count);
}

// Returns the current OS stack trace as an std::string.
//
// The maximum number of stack frames to be included is specified by
// the gtest_stack_trace_depth flag.  The skip_count parameter
// specifies the number of top frames to be skipped, which doesn't
// count against the number of frames to be included.
//
// For example, if Foo() calls Bar(), which in turn calls
// CurrentOsStackTraceExceptTop(1), Foo() will be included in the
// trace but Bar() and CurrentOsStackTraceExceptTop() won't.
std::string UnitTestImpl::CurrentOsStackTraceExceptTop(int skip_count)
{
    return os_stack_trace_getter()->CurrentStackTrace(
        static_cast<int>(GTEST_FLAG(stack_trace_depth)),
        skip_count + 1
        // Skips the user-specified number of frames plus this function
        // itself.
    ); // NOLINT
}

// Returns the current time in milliseconds.
TimeInMillis GetTimeInMillis()
{
#if GTEST_OS_WINDOWS_MOBILE || defined(__BORLANDC__)
    // Difference between 1970-01-01 and 1601-01-01 in milliseconds.
    // http://analogous.blogspot.com/2005/04/epoch.html
    const TimeInMillis kJavaEpochToWinFileTimeDelta =
        static_cast<TimeInMillis>(116444736UL) * 100000UL;
    const DWORD kTenthMicrosInMilliSecond = 10000;

    SYSTEMTIME now_systime;
    FILETIME now_filetime;
    ULARGE_INTEGER now_int64;
    GetSystemTime(&now_systime);
    if (SystemTimeToFileTime(&now_systime, &now_filetime)) {
        now_int64.LowPart = now_filetime.dwLowDateTime;
        now_int64.HighPart = now_filetime.dwHighDateTime;
        now_int64.QuadPart = (now_int64.QuadPart / kTenthMicrosInMilliSecond) - kJavaEpochToWinFileTimeDelta;
        return now_int64.QuadPart;
    }
    return 0;
#elif GTEST_OS_WINDOWS && !GTEST_HAS_GETTIMEOFDAY_
    __timeb64 now;

    // MSVC 8 deprecates _ftime64(), so we want to suppress warning 4996
    // (deprecated function) there.
    GTEST_DISABLE_MSC_DEPRECATED_PUSH_()
    _ftime64(&now);
    GTEST_DISABLE_MSC_DEPRECATED_POP_()

    return static_cast<TimeInMillis>(now.time) * 1000 + now.millitm;
#elif GTEST_HAS_GETTIMEOFDAY_
    struct timeval now;
    gettimeofday(&now, nullptr);
    return static_cast<TimeInMillis>(now.tv_sec) * 1000 + now.tv_usec / 1000;
#else
#error "Don't know how to get the current time on your system."
#endif
}

// Utilities

// class String.

#if GTEST_OS_WINDOWS_MOBILE
// Creates a UTF-16 wide string from the given ANSI string, allocating
// memory using new. The caller is responsible for deleting the return
// value using delete[]. Returns the wide string, or NULL if the
// input is NULL.
LPCWSTR String::AnsiToUtf16(const char *ansi)
{
    if (!ansi)
        return nullptr;
    const int length = strlen(ansi);
    const int unicode_length =
        MultiByteToWideChar(CP_ACP, 0, ansi, length, nullptr, 0);
    WCHAR *unicode = new WCHAR[unicode_length + 1];
    MultiByteToWideChar(CP_ACP, 0, ansi, length,
                        unicode, unicode_length);
    unicode[unicode_length] = 0;
    return unicode;
}

// Creates an ANSI string from the given wide string, allocating
// memory using new. The caller is responsible for deleting the return
// value using delete[]. Returns the ANSI string, or NULL if the
// input is NULL.
const char *String::Utf16ToAnsi(LPCWSTR utf16_str)
{
    if (!utf16_str)
        return nullptr;
    const int ansi_length = WideCharToMultiByte(CP_ACP, 0, utf16_str, -1, nullptr,
                                                0, nullptr, nullptr);
    char *ansi = new char[ansi_length + 1];
    WideCharToMultiByte(CP_ACP, 0, utf16_str, -1, ansi, ansi_length, nullptr,
                        nullptr);
    ansi[ansi_length] = 0;
    return ansi;
}

#endif // GTEST_OS_WINDOWS_MOBILE

// Compares two C strings.  Returns true if and only if they have the same
// content.
//
// Unlike strcmp(), this function can handle NULL argument(s).  A NULL
// C string is considered different to any non-NULL C string,
// including the empty string.
bool String::CStringEquals(const char *lhs, const char *rhs)
{
    if (lhs == nullptr)
        return rhs == nullptr;

    if (rhs == nullptr)
        return false;

    return strcmp(lhs, rhs) == 0;
}

#if GTEST_HAS_STD_WSTRING

// Converts an array of wide chars to a narrow string using the UTF-8
// encoding, and streams the result to the given Message object.
static void StreamWideCharsToMessage(const wchar_t *wstr, size_t length,
                                     Message *msg)
{
    for (size_t i = 0; i != length;) { // NOLINT
        if (wstr[i] != L'\0') {
            *msg << WideStringToUtf8(wstr + i, static_cast<int>(length - i));
            while (i != length && wstr[i] != L'\0')
                i++;
        } else {
            *msg << '\0';
            i++;
        }
    }
}

#endif // GTEST_HAS_STD_WSTRING

void SplitString(const ::std::string &str, char delimiter,
                 ::std::vector<::std::string> *dest)
{
    ::std::vector<::std::string> parsed;
    ::std::string::size_type pos = 0;
    while (::testing::internal::AlwaysTrue()) {
        const ::std::string::size_type colon = str.find(delimiter, pos);
        if (colon == ::std::string::npos) {
            parsed.push_back(str.substr(pos));
            break;
        } else {
            parsed.push_back(str.substr(pos, colon - pos));
            pos = colon + 1;
        }
    }
    dest->swap(parsed);
}

} // namespace internal

// Constructs an empty Message.
// We allocate the stringstream separately because otherwise each use of
// ASSERT/EXPECT in a procedure adds over 200 bytes to the procedure's
// stack frame leading to huge stack frames in some cases; gcc does not reuse
// the stack space.
Message::Message()
    : ss_(new ::std::stringstream)
{
    // By default, we want there to be enough precision when printing
    // a double to a Message.
    *ss_ << std::setprecision(std::numeric_limits<double>::digits10 + 2);
}

// These two overloads allow streaming a wide C string to a Message
// using the UTF-8 encoding.
Message &Message::operator<<(const wchar_t *wide_c_str)
{
    return *this << internal::String::ShowWideCString(wide_c_str);
}
Message &Message::operator<<(wchar_t *wide_c_str)
{
    return *this << internal::String::ShowWideCString(wide_c_str);
}

#if GTEST_HAS_STD_WSTRING
// Converts the given wide string to a narrow string using the UTF-8
// encoding, and streams the result to this Message object.
Message &Message::operator<<(const ::std::wstring &wstr)
{
    internal::StreamWideCharsToMessage(wstr.c_str(), wstr.length(), this);
    return *this;
}
#endif // GTEST_HAS_STD_WSTRING

// Gets the text streamed to this object so far as an std::string.
// Each '\0' character in the buffer is replaced with "\\0".
std::string Message::GetString() const
{
    return internal::StringStreamToString(ss_.get());
}

// AssertionResult constructors.
// Used in EXPECT_TRUE/FALSE(assertion_result).
AssertionResult::AssertionResult(const AssertionResult &other)
    : success_(other.success_)
    , message_(other.message_.get() != nullptr
                   ? new ::std::string(*other.message_)
                   : static_cast<::std::string *>(nullptr))
{
}

// Swaps two AssertionResults.
void AssertionResult::swap(AssertionResult &other)
{
    using std::swap;
    swap(success_, other.success_);
    swap(message_, other.message_);
}

// Returns the assertion's negation. Used with EXPECT/ASSERT_FALSE.
AssertionResult AssertionResult::operator!() const
{
    AssertionResult negation(!success_);
    if (message_.get() != nullptr)
        negation << *message_;
    return negation;
}

// Makes a successful assertion result.
AssertionResult AssertionSuccess()
{
    return AssertionResult(true);
}

// Makes a failed assertion result.
AssertionResult AssertionFailure()
{
    return AssertionResult(false);
}

// Makes a failed assertion result with the given failure message.
// Deprecated; use AssertionFailure() << message.
AssertionResult AssertionFailure(const Message &message)
{
    return AssertionFailure() << message;
}

namespace internal {

namespace edit_distance {
std::vector<EditType> CalculateOptimalEdits(const std::vector<size_t> &left,
                                            const std::vector<size_t> &right)
{
    std::vector<std::vector<double>> costs(
        left.size() + 1, std::vector<double>(right.size() + 1));
    std::vector<std::vector<EditType>> best_move(
        left.size() + 1, std::vector<EditType>(right.size() + 1));

    // Populate for empty right.
    for (size_t l_i = 0; l_i < costs.size(); ++l_i) {
        costs[l_i][0] = static_cast<double>(l_i);
        best_move[l_i][0] = kRemove;
    }
    // Populate for empty left.
    for (size_t r_i = 1; r_i < costs[0].size(); ++r_i) {
        costs[0][r_i] = static_cast<double>(r_i);
        best_move[0][r_i] = kAdd;
    }

    for (size_t l_i = 0; l_i < left.size(); ++l_i) {
        for (size_t r_i = 0; r_i < right.size(); ++r_i) {
            if (left[l_i] == right[r_i]) {
                // Found a match. Consume it.
                costs[l_i + 1][r_i + 1] = costs[l_i][r_i];
                best_move[l_i + 1][r_i + 1] = kMatch;
                continue;
            }

            const double add = costs[l_i + 1][r_i];
            const double remove = costs[l_i][r_i + 1];
            const double replace = costs[l_i][r_i];
            if (add < remove && add < replace) {
                costs[l_i + 1][r_i + 1] = add + 1;
                best_move[l_i + 1][r_i + 1] = kAdd;
            } else if (remove < add && remove < replace) {
                costs[l_i + 1][r_i + 1] = remove + 1;
                best_move[l_i + 1][r_i + 1] = kRemove;
            } else {
                // We make replace a little more expensive than add/remove to lower
                // their priority.
                costs[l_i + 1][r_i + 1] = replace + 1.00001;
                best_move[l_i + 1][r_i + 1] = kReplace;
            }
        }
    }

    // Reconstruct the best path. We do it in reverse order.
    std::vector<EditType> best_path;
    for (size_t l_i = left.size(), r_i = right.size(); l_i > 0 || r_i > 0;) {
        EditType move = best_move[l_i][r_i];
        best_path.push_back(move);
        l_i -= move != kAdd;
        r_i -= move != kRemove;
    }
    std::reverse(best_path.begin(), best_path.end());
    return best_path;
}

namespace {

// Helper class to convert string into ids with deduplication.
class InternalStrings
{
public:
    size_t GetId(const std::string &str)
    {
        IdMap::iterator it = ids_.find(str);
        if (it != ids_.end())
            return it->second;
        size_t id = ids_.size();
        return ids_[str] = id;
    }

private:
    typedef std::map<std::string, size_t> IdMap;
    IdMap ids_;
};

} // namespace

std::vector<EditType> CalculateOptimalEdits(
    const std::vector<std::string> &left,
    const std::vector<std::string> &right)
{
    std::vector<size_t> left_ids, right_ids;
    {
        InternalStrings intern_table;
        for (size_t i = 0; i < left.size(); ++i) {
            left_ids.push_back(intern_table.GetId(left[i]));
        }
        for (size_t i = 0; i < right.size(); ++i) {
            right_ids.push_back(intern_table.GetId(right[i]));
        }
    }
    return CalculateOptimalEdits(left_ids, right_ids);
}

namespace {

// Helper class that holds the state for one hunk and prints it out to the
// stream.
// It reorders adds/removes when possible to group all removes before all
// adds. It also adds the hunk header before printint into the stream.
class Hunk
{
public:
    Hunk(size_t left_start, size_t right_start)
        : left_start_(left_start)
        , right_start_(right_start)
        , adds_()
        , removes_()
        , common_()
    {
    }

    void PushLine(char edit, const char *line)
    {
        switch (edit) {
        case ' ':
            ++common_;
            FlushEdits();
            hunk_.push_back(std::make_pair(' ', line));
            break;
        case '-':
            ++removes_;
            hunk_removes_.push_back(std::make_pair('-', line));
            break;
        case '+':
            ++adds_;
            hunk_adds_.push_back(std::make_pair('+', line));
            break;
        }
    }

    void PrintTo(std::ostream *os)
    {
        PrintHeader(os);
        FlushEdits();
        for (std::list<std::pair<char, const char *>>::const_iterator it =
                 hunk_.begin();
             it != hunk_.end(); ++it) {
            *os << it->first << it->second << "\n";
        }
    }

    bool has_edits() const { return adds_ || removes_; }

private:
    void FlushEdits()
    {
        hunk_.splice(hunk_.end(), hunk_removes_);
        hunk_.splice(hunk_.end(), hunk_adds_);
    }

    // Print a unified diff header for one hunk.
    // The format is
    //   "@@ -<left_start>,<left_length> +<right_start>,<right_length> @@"
    // where the left/right parts are omitted if unnecessary.
    void PrintHeader(std::ostream *ss) const
    {
        *ss << "@@ ";
        if (removes_) {
            *ss << "-" << left_start_ << "," << (removes_ + common_);
        }
        if (removes_ && adds_) {
            *ss << " ";
        }
        if (adds_) {
            *ss << "+" << right_start_ << "," << (adds_ + common_);
        }
        *ss << " @@\n";
    }

    size_t left_start_, right_start_;
    size_t adds_, removes_, common_;
    std::list<std::pair<char, const char *>> hunk_, hunk_adds_, hunk_removes_;
};

} // namespace

// Create a list of diff hunks in Unified diff format.
// Each hunk has a header generated by PrintHeader above plus a body with
// lines prefixed with ' ' for no change, '-' for deletion and '+' for
// addition.
// 'context' represents the desired unchanged prefix/suffix around the diff.
// If two hunks are close enough that their contexts overlap, then they are
// joined into one hunk.
std::string CreateUnifiedDiff(const std::vector<std::string> &left,
                              const std::vector<std::string> &right,
                              size_t context)
{
    const std::vector<EditType> edits = CalculateOptimalEdits(left, right);

    size_t l_i = 0, r_i = 0, edit_i = 0;
    std::stringstream ss;
    while (edit_i < edits.size()) {
        // Find first edit.
        while (edit_i < edits.size() && edits[edit_i] == kMatch) {
            ++l_i;
            ++r_i;
            ++edit_i;
        }

        // Find the first line to include in the hunk.
        const size_t prefix_context = std::min(l_i, context);
        Hunk hunk(l_i - prefix_context + 1, r_i - prefix_context + 1);
        for (size_t i = prefix_context; i > 0; --i) {
            hunk.PushLine(' ', left[l_i - i].c_str());
        }

        // Iterate the edits until we found enough suffix for the hunk or the input
        // is over.
        size_t n_suffix = 0;
        for (; edit_i < edits.size(); ++edit_i) {
            if (n_suffix >= context) {
                // Continue only if the next hunk is very close.
                auto it = edits.begin() + static_cast<int>(edit_i);
                while (it != edits.end() && *it == kMatch)
                    ++it;
                if (it == edits.end() || static_cast<size_t>(it - edits.begin()) - edit_i >= context) {
                    // There is no next edit or it is too far away.
                    break;
                }
            }

            EditType edit = edits[edit_i];
            // Reset count when a non match is found.
            n_suffix = edit == kMatch ? n_suffix + 1 : 0;

            if (edit == kMatch || edit == kRemove || edit == kReplace) {
                hunk.PushLine(edit == kMatch ? ' ' : '-', left[l_i].c_str());
            }
            if (edit == kAdd || edit == kReplace) {
                hunk.PushLine('+', right[r_i].c_str());
            }

            // Advance indices, depending on edit type.
            l_i += edit != kAdd;
            r_i += edit != kRemove;
        }

        if (!hunk.has_edits()) {
            // We are done. We don't want this hunk.
            break;
        }

        hunk.PrintTo(&ss);
    }
    return ss.str();
}

} // namespace edit_distance

namespace {

// The string representation of the values received in EqFailure() are already
// escaped. Split them on escaped '\n' boundaries. Leave all other escaped
// characters the same.
std::vector<std::string> SplitEscapedString(const std::string &str)
{
    std::vector<std::string> lines;
    size_t start = 0, end = str.size();
    if (end > 2 && str[0] == '"' && str[end - 1] == '"') {
        ++start;
        --end;
    }
    bool escaped = false;
    for (size_t i = start; i + 1 < end; ++i) {
        if (escaped) {
            escaped = false;
            if (str[i] == 'n') {
                lines.push_back(str.substr(start, i - start - 1));
                start = i + 1;
            }
        } else {
            escaped = str[i] == '\\';
        }
    }
    lines.push_back(str.substr(start, end - start));
    return lines;
}

} // namespace

// Constructs and returns the message for an equality assertion
// (e.g. ASSERT_EQ, EXPECT_STREQ, etc) failure.
//
// The first four parameters are the expressions used in the assertion
// and their values, as strings.  For example, for ASSERT_EQ(foo, bar)
// where foo is 5 and bar is 6, we have:
//
//   lhs_expression: "foo"
//   rhs_expression: "bar"
//   lhs_value:      "5"
//   rhs_value:      "6"
//
// The ignoring_case parameter is true if and only if the assertion is a
// *_STRCASEEQ*.  When it's true, the string "Ignoring case" will
// be inserted into the message.
AssertionResult EqFailure(const char *lhs_expression,
                          const char *rhs_expression,
                          const std::string &lhs_value,
                          const std::string &rhs_value,
                          bool ignoring_case)
{
    Message msg;
    msg << "Expected equality of these values:";
    msg << "\n  " << lhs_expression;
    if (lhs_value != lhs_expression) {
        msg << "\n    Which is: " << lhs_value;
    }
    msg << "\n  " << rhs_expression;
    if (rhs_value != rhs_expression) {
        msg << "\n    Which is: " << rhs_value;
    }

    if (ignoring_case) {
        msg << "\nIgnoring case";
    }

    if (!lhs_value.empty() && !rhs_value.empty()) {
        const std::vector<std::string> lhs_lines =
            SplitEscapedString(lhs_value);
        const std::vector<std::string> rhs_lines =
            SplitEscapedString(rhs_value);
        if (lhs_lines.size() > 1 || rhs_lines.size() > 1) {
            msg << "\nWith diff:\n"
                << edit_distance::CreateUnifiedDiff(lhs_lines, rhs_lines);
        }
    }

    return AssertionFailure() << msg;
}

// Constructs a failure message for Boolean assertions such as EXPECT_TRUE.
std::string GetBoolAssertionFailureMessage(
    const AssertionResult &assertion_result,
    const char *expression_text,
    const char *actual_predicate_value,
    const char *expected_predicate_value)
{
    const char *actual_message = assertion_result.message();
    Message msg;
    msg << "Value of: " << expression_text
        << "\n  Actual: " << actual_predicate_value;
    if (actual_message[0] != '\0')
        msg << " (" << actual_message << ")";
    msg << "\nExpected: " << expected_predicate_value;
    return msg.GetString();
}

// Helper function for implementing ASSERT_NEAR.
AssertionResult DoubleNearPredFormat(const char *expr1,
                                     const char *expr2,
                                     const char *abs_error_expr,
                                     double val1,
                                     double val2,
                                     double abs_error)
{
    const double diff = fabs(val1 - val2);
    if (diff <= abs_error)
        return AssertionSuccess();

    return AssertionFailure()
           << "The difference between " << expr1 << " and " << expr2
           << " is " << diff << ", which exceeds " << abs_error_expr << ", where\n"
           << expr1 << " evaluates to " << val1 << ",\n"
           << expr2 << " evaluates to " << val2 << ", and\n"
           << abs_error_expr << " evaluates to " << abs_error << ".";
}

// Helper template for implementing FloatLE() and DoubleLE().
template<typename RawType>
AssertionResult FloatingPointLE(const char *expr1,
                                const char *expr2,
                                RawType val1,
                                RawType val2)
{
    // Returns success if val1 is less than val2,
    if (val1 < val2) {
        return AssertionSuccess();
    }

    // or if val1 is almost equal to val2.
    const FloatingPoint<RawType> lhs(val1), rhs(val2);
    if (lhs.AlmostEquals(rhs)) {
        return AssertionSuccess();
    }

    // Note that the above two checks will both fail if either val1 or
    // val2 is NaN, as the IEEE floating-point standard requires that
    // any predicate involving a NaN must return false.

    ::std::stringstream val1_ss;
    val1_ss << std::setprecision(std::numeric_limits<RawType>::digits10 + 2)
            << val1;

    ::std::stringstream val2_ss;
    val2_ss << std::setprecision(std::numeric_limits<RawType>::digits10 + 2)
            << val2;

    return AssertionFailure()
           << "Expected: (" << expr1 << ") <= (" << expr2 << ")\n"
           << "  Actual: " << StringStreamToString(&val1_ss) << " vs "
           << StringStreamToString(&val2_ss);
}

} // namespace internal

// Asserts that val1 is less than, or almost equal to, val2.  Fails
// otherwise.  In particular, it fails if either val1 or val2 is NaN.
AssertionResult FloatLE(const char *expr1, const char *expr2,
                        float val1, float val2)
{
    return internal::FloatingPointLE<float>(expr1, expr2, val1, val2);
}

// Asserts that val1 is less than, or almost equal to, val2.  Fails
// otherwise.  In particular, it fails if either val1 or val2 is NaN.
AssertionResult DoubleLE(const char *expr1, const char *expr2,
                         double val1, double val2)
{
    return internal::FloatingPointLE<double>(expr1, expr2, val1, val2);
}

namespace internal {

// The helper function for {ASSERT|EXPECT}_EQ with int or enum
// arguments.
AssertionResult CmpHelperEQ(const char *lhs_expression,
                            const char *rhs_expression,
                            BiggestInt lhs,
                            BiggestInt rhs)
{
    if (lhs == rhs) {
        return AssertionSuccess();
    }

    return EqFailure(lhs_expression,
                     rhs_expression,
                     FormatForComparisonFailureMessage(lhs, rhs),
                     FormatForComparisonFailureMessage(rhs, lhs),
                     false);
}

// A macro for implementing the helper functions needed to implement
// ASSERT_?? and EXPECT_?? with integer or enum arguments.  It is here
// just to avoid copy-and-paste of similar code.
#define GTEST_IMPL_CMP_HELPER_(op_name, op) \
    AssertionResult CmpHelper##op_name(const char *expr1, const char *expr2, \
                                       BiggestInt val1, BiggestInt val2) \
    { \
        if (val1 op val2) { \
            return AssertionSuccess(); \
        } else { \
            return AssertionFailure() \
                   << "Expected: (" << expr1 << ") " #op " (" << expr2 \
                   << "), actual: " << FormatForComparisonFailureMessage(val1, val2) \
                   << " vs " << FormatForComparisonFailureMessage(val2, val1); \
        } \
    }

// Implements the helper function for {ASSERT|EXPECT}_NE with int or
// enum arguments.
GTEST_IMPL_CMP_HELPER_(NE, !=)
// Implements the helper function for {ASSERT|EXPECT}_LE with int or
// enum arguments.
GTEST_IMPL_CMP_HELPER_(LE, <=)
// Implements the helper function for {ASSERT|EXPECT}_LT with int or
// enum arguments.
GTEST_IMPL_CMP_HELPER_(LT, <)
// Implements the helper function for {ASSERT|EXPECT}_GE with int or
// enum arguments.
GTEST_IMPL_CMP_HELPER_(GE, >=)
// Implements the helper function for {ASSERT|EXPECT}_GT with int or
// enum arguments.
GTEST_IMPL_CMP_HELPER_(GT, >)

#undef GTEST_IMPL_CMP_HELPER_

// The helper function for {ASSERT|EXPECT}_STREQ.
AssertionResult CmpHelperSTREQ(const char *lhs_expression,
                               const char *rhs_expression,
                               const char *lhs,
                               const char *rhs)
{
    if (String::CStringEquals(lhs, rhs)) {
        return AssertionSuccess();
    }

    return EqFailure(lhs_expression,
                     rhs_expression,
                     PrintToString(lhs),
                     PrintToString(rhs),
                     false);
}

// The helper function for {ASSERT|EXPECT}_STRCASEEQ.
AssertionResult CmpHelperSTRCASEEQ(const char *lhs_expression,
                                   const char *rhs_expression,
                                   const char *lhs,
                                   const char *rhs)
{
    if (String::CaseInsensitiveCStringEquals(lhs, rhs)) {
        return AssertionSuccess();
    }

    return EqFailure(lhs_expression,
                     rhs_expression,
                     PrintToString(lhs),
                     PrintToString(rhs),
                     true);
}

// The helper function for {ASSERT|EXPECT}_STRNE.
AssertionResult CmpHelperSTRNE(const char *s1_expression,
                               const char *s2_expression,
                               const char *s1,
                               const char *s2)
{
    if (!String::CStringEquals(s1, s2)) {
        return AssertionSuccess();
    } else {
        return AssertionFailure() << "Expected: (" << s1_expression << ") != ("
                                  << s2_expression << "), actual: \""
                                  << s1 << "\" vs \"" << s2 << "\"";
    }
}

// The helper function for {ASSERT|EXPECT}_STRCASENE.
AssertionResult CmpHelperSTRCASENE(const char *s1_expression,
                                   const char *s2_expression,
                                   const char *s1,
                                   const char *s2)
{
    if (!String::CaseInsensitiveCStringEquals(s1, s2)) {
        return AssertionSuccess();
    } else {
        return AssertionFailure()
               << "Expected: (" << s1_expression << ") != ("
               << s2_expression << ") (ignoring case), actual: \""
               << s1 << "\" vs \"" << s2 << "\"";
    }
}

} // namespace internal

namespace {

// Helper functions for implementing IsSubString() and IsNotSubstring().

// This group of overloaded functions return true if and only if needle
// is a substring of haystack.  NULL is considered a substring of
// itself only.

bool IsSubstringPred(const char *needle, const char *haystack)
{
    if (needle == nullptr || haystack == nullptr)
        return needle == haystack;

    return strstr(haystack, needle) != nullptr;
}

bool IsSubstringPred(const wchar_t *needle, const wchar_t *haystack)
{
    if (needle == nullptr || haystack == nullptr)
        return needle == haystack;

    return wcsstr(haystack, needle) != nullptr;
}

// StringType here can be either ::std::string or ::std::wstring.
template<typename StringType>
bool IsSubstringPred(const StringType &needle,
                     const StringType &haystack)
{
    return haystack.find(needle) != StringType::npos;
}

// This function implements either IsSubstring() or IsNotSubstring(),
// depending on the value of the expected_to_be_substring parameter.
// StringType here can be const char*, const wchar_t*, ::std::string,
// or ::std::wstring.
template<typename StringType>
AssertionResult IsSubstringImpl(
    bool expected_to_be_substring,
    const char *needle_expr, const char *haystack_expr,
    const StringType &needle, const StringType &haystack)
{
    if (IsSubstringPred(needle, haystack) == expected_to_be_substring)
        return AssertionSuccess();

    const bool is_wide_string = sizeof(needle[0]) > 1;
    const char *const begin_string_quote = is_wide_string ? "L\"" : "\"";
    return AssertionFailure()
           << "Value of: " << needle_expr << "\n"
           << "  Actual: " << begin_string_quote << needle << "\"\n"
           << "Expected: " << (expected_to_be_substring ? "" : "not ")
           << "a substring of " << haystack_expr << "\n"
           << "Which is: " << begin_string_quote << haystack << "\"";
}

} // namespace

// IsSubstring() and IsNotSubstring() check whether needle is a
// substring of haystack (NULL is considered a substring of itself
// only), and return an appropriate error message when they fail.

AssertionResult IsSubstring(
    const char *needle_expr, const char *haystack_expr,
    const char *needle, const char *haystack)
{
    return IsSubstringImpl(true, needle_expr, haystack_expr, needle, haystack);
}

AssertionResult IsSubstring(
    const char *needle_expr, const char *haystack_expr,
    const wchar_t *needle, const wchar_t *haystack)
{
    return IsSubstringImpl(true, needle_expr, haystack_expr, needle, haystack);
}

AssertionResult IsNotSubstring(
    const char *needle_expr, const char *haystack_expr,
    const char *needle, const char *haystack)
{
    return IsSubstringImpl(false, needle_expr, haystack_expr, needle, haystack);
}

AssertionResult IsNotSubstring(
    const char *needle_expr, const char *haystack_expr,
    const wchar_t *needle, const wchar_t *haystack)
{
    return IsSubstringImpl(false, needle_expr, haystack_expr, needle, haystack);
}

AssertionResult IsSubstring(
    const char *needle_expr, const char *haystack_expr,
    const ::std::string &needle, const ::std::string &haystack)
{
    return IsSubstringImpl(true, needle_expr, haystack_expr, needle, haystack);
}

AssertionResult IsNotSubstring(
    const char *needle_expr, const char *haystack_expr,
    const ::std::string &needle, const ::std::string &haystack)
{
    return IsSubstringImpl(false, needle_expr, haystack_expr, needle, haystack);
}

#if GTEST_HAS_STD_WSTRING
AssertionResult IsSubstring(
    const char *needle_expr, const char *haystack_expr,
    const ::std::wstring &needle, const ::std::wstring &haystack)
{
    return IsSubstringImpl(true, needle_expr, haystack_expr, needle, haystack);
}

AssertionResult IsNotSubstring(
    const char *needle_expr, const char *haystack_expr,
    const ::std::wstring &needle, const ::std::wstring &haystack)
{
    return IsSubstringImpl(false, needle_expr, haystack_expr, needle, haystack);
}
#endif // GTEST_HAS_STD_WSTRING

namespace internal {

#if GTEST_OS_WINDOWS

namespace {

// Helper function for IsHRESULT{SuccessFailure} predicates
AssertionResult HRESULTFailureHelper(const char *expr,
                                     const char *expected,
                                     long hr)
{ // NOLINT
#if GTEST_OS_WINDOWS_MOBILE || GTEST_OS_WINDOWS_TV_TITLE

    // Windows CE doesn't support FormatMessage.
    const char error_text[] = "";

#else

    // Looks up the human-readable system message for the HRESULT code
    // and since we're not passing any params to FormatMessage, we don't
    // want inserts expanded.
    const DWORD kFlags = FORMAT_MESSAGE_FROM_SYSTEM | FORMAT_MESSAGE_IGNORE_INSERTS;
    const DWORD kBufSize = 4096;
    // Gets the system's human readable message string for this HRESULT.
    char error_text[kBufSize] = {'\0'};
    DWORD message_length = ::FormatMessageA(kFlags,
                                            0, // no source, we're asking system
                                            static_cast<DWORD>(hr), // the error
                                            0, // no line width restrictions
                                            error_text, // output buffer
                                            kBufSize, // buf size
                                            nullptr); // no arguments for inserts
    // Trims tailing white space (FormatMessage leaves a trailing CR-LF)
    for (; message_length && IsSpace(error_text[message_length - 1]);
         --message_length) {
        error_text[message_length - 1] = '\0';
    }

#endif // GTEST_OS_WINDOWS_MOBILE

    const std::string error_hex("0x" + String::FormatHexInt(hr));
    return ::testing::AssertionFailure()
           << "Expected: " << expr << " " << expected << ".\n"
           << "  Actual: " << error_hex << " " << error_text << "\n";
}

} // namespace

AssertionResult IsHRESULTSuccess(const char *expr, long hr)
{ // NOLINT
    if (SUCCEEDED(hr)) {
        return AssertionSuccess();
    }
    return HRESULTFailureHelper(expr, "succeeds", hr);
}

AssertionResult IsHRESULTFailure(const char *expr, long hr)
{ // NOLINT
    if (FAILED(hr)) {
        return AssertionSuccess();
    }
    return HRESULTFailureHelper(expr, "fails", hr);
}

#endif // GTEST_OS_WINDOWS

// Utility functions for encoding Unicode text (wide strings) in
// UTF-8.

// A Unicode code-point can have up to 21 bits, and is encoded in UTF-8
// like this:
//
// Code-point length   Encoding
//   0 -  7 bits       0xxxxxxx
//   8 - 11 bits       110xxxxx 10xxxxxx
//  12 - 16 bits       1110xxxx 10xxxxxx 10xxxxxx
//  17 - 21 bits       11110xxx 10xxxxxx 10xxxxxx 10xxxxxx

// The maximum code-point a one-byte UTF-8 sequence can represent.
constexpr uint32_t kMaxCodePoint1 = (static_cast<uint32_t>(1) << 7) - 1;

// The maximum code-point a two-byte UTF-8 sequence can represent.
constexpr uint32_t kMaxCodePoint2 = (static_cast<uint32_t>(1) << (5 + 6)) - 1;

// The maximum code-point a three-byte UTF-8 sequence can represent.
constexpr uint32_t kMaxCodePoint3 = (static_cast<uint32_t>(1) << (4 + 2 * 6)) - 1;

// The maximum code-point a four-byte UTF-8 sequence can represent.
constexpr uint32_t kMaxCodePoint4 = (static_cast<uint32_t>(1) << (3 + 3 * 6)) - 1;

// Chops off the n lowest bits from a bit pattern.  Returns the n
// lowest bits.  As a side effect, the original bit pattern will be
// shifted to the right by n bits.
inline uint32_t ChopLowBits(uint32_t *bits, int n)
{
    const uint32_t low_bits = *bits & ((static_cast<uint32_t>(1) << n) - 1);
    *bits >>= n;
    return low_bits;
}

// Converts a Unicode code point to a narrow string in UTF-8 encoding.
// code_point parameter is of type uint32_t because wchar_t may not be
// wide enough to contain a code point.
// If the code_point is not a valid Unicode code point
// (i.e. outside of Unicode range U+0 to U+10FFFF) it will be converted
// to "(Invalid Unicode 0xXXXXXXXX)".
std::string CodePointToUtf8(uint32_t code_point)
{
    if (code_point > kMaxCodePoint4) {
        return "(Invalid Unicode 0x" + String::FormatHexUInt32(code_point) + ")";
    }

    char str[5]; // Big enough for the largest valid code point.
    if (code_point <= kMaxCodePoint1) {
        str[1] = '\0';
        str[0] = static_cast<char>(code_point); // 0xxxxxxx
    } else if (code_point <= kMaxCodePoint2) {
        str[2] = '\0';
        str[1] = static_cast<char>(0x80 | ChopLowBits(&code_point, 6)); // 10xxxxxx
        str[0] = static_cast<char>(0xC0 | code_point); // 110xxxxx
    } else if (code_point <= kMaxCodePoint3) {
        str[3] = '\0';
        str[2] = static_cast<char>(0x80 | ChopLowBits(&code_point, 6)); // 10xxxxxx
        str[1] = static_cast<char>(0x80 | ChopLowBits(&code_point, 6)); // 10xxxxxx
        str[0] = static_cast<char>(0xE0 | code_point); // 1110xxxx
    } else { // code_point <= kMaxCodePoint4
        str[4] = '\0';
        str[3] = static_cast<char>(0x80 | ChopLowBits(&code_point, 6)); // 10xxxxxx
        str[2] = static_cast<char>(0x80 | ChopLowBits(&code_point, 6)); // 10xxxxxx
        str[1] = static_cast<char>(0x80 | ChopLowBits(&code_point, 6)); // 10xxxxxx
        str[0] = static_cast<char>(0xF0 | code_point); // 11110xxx
    }
    return str;
}

// The following two functions only make sense if the system
// uses UTF-16 for wide string encoding. All supported systems
// with 16 bit wchar_t (Windows, Cygwin) do use UTF-16.

// Determines if the arguments constitute UTF-16 surrogate pair
// and thus should be combined into a single Unicode code point
// using CreateCodePointFromUtf16SurrogatePair.
inline bool IsUtf16SurrogatePair(wchar_t first, wchar_t second)
{
    return sizeof(wchar_t) == 2 && (first & 0xFC00) == 0xD800 && (second & 0xFC00) == 0xDC00;
}

// Creates a Unicode code point from UTF16 surrogate pair.
inline uint32_t CreateCodePointFromUtf16SurrogatePair(wchar_t first,
                                                      wchar_t second)
{
    const auto first_u = static_cast<uint32_t>(first);
    const auto second_u = static_cast<uint32_t>(second);
    const uint32_t mask = (1 << 10) - 1;
    return (sizeof(wchar_t) == 2)
               ? (((first_u & mask) << 10) | (second_u & mask)) + 0x10000
               :
               // This function should not be called when the condition is
               // false, but we provide a sensible default in case it is.
               first_u;
}

// Converts a wide string to a narrow string in UTF-8 encoding.
// The wide string is assumed to have the following encoding:
//   UTF-16 if sizeof(wchar_t) == 2 (on Windows, Cygwin)
//   UTF-32 if sizeof(wchar_t) == 4 (on Linux)
// Parameter str points to a null-terminated wide string.
// Parameter num_chars may additionally limit the number
// of wchar_t characters processed. -1 is used when the entire string
// should be processed.
// If the string contains code points that are not valid Unicode code points
// (i.e. outside of Unicode range U+0 to U+10FFFF) they will be output
// as '(Invalid Unicode 0xXXXXXXXX)'. If the string is in UTF16 encoding
// and contains invalid UTF-16 surrogate pairs, values in those pairs
// will be encoded as individual Unicode characters from Basic Normal Plane.
std::string WideStringToUtf8(const wchar_t *str, int num_chars)
{
    if (num_chars == -1)
        num_chars = static_cast<int>(wcslen(str));

    ::std::stringstream stream;
    for (int i = 0; i < num_chars; ++i) {
        uint32_t unicode_code_point;

        if (str[i] == L'\0') {
            break;
        } else if (i + 1 < num_chars && IsUtf16SurrogatePair(str[i], str[i + 1])) {
            unicode_code_point = CreateCodePointFromUtf16SurrogatePair(str[i],
                                                                       str[i + 1]);
            i++;
        } else {
            unicode_code_point = static_cast<uint32_t>(str[i]);
        }

        stream << CodePointToUtf8(unicode_code_point);
    }
    return StringStreamToString(&stream);
}

// Converts a wide C string to an std::string using the UTF-8 encoding.
// NULL will be converted to "(null)".
std::string String::ShowWideCString(const wchar_t *wide_c_str)
{
    if (wide_c_str == nullptr)
        return "(null)";

    return internal::WideStringToUtf8(wide_c_str, -1);
}

// Compares two wide C strings.  Returns true if and only if they have the
// same content.
//
// Unlike wcscmp(), this function can handle NULL argument(s).  A NULL
// C string is considered different to any non-NULL C string,
// including the empty string.
bool String::WideCStringEquals(const wchar_t *lhs, const wchar_t *rhs)
{
    if (lhs == nullptr)
        return rhs == nullptr;

    if (rhs == nullptr)
        return false;

    return wcscmp(lhs, rhs) == 0;
}

// Helper function for *_STREQ on wide strings.
AssertionResult CmpHelperSTREQ(const char *lhs_expression,
                               const char *rhs_expression,
                               const wchar_t *lhs,
                               const wchar_t *rhs)
{
    if (String::WideCStringEquals(lhs, rhs)) {
        return AssertionSuccess();
    }

    return EqFailure(lhs_expression,
                     rhs_expression,
                     PrintToString(lhs),
                     PrintToString(rhs),
                     false);
}

// Helper function for *_STRNE on wide strings.
AssertionResult CmpHelperSTRNE(const char *s1_expression,
                               const char *s2_expression,
                               const wchar_t *s1,
                               const wchar_t *s2)
{
    if (!String::WideCStringEquals(s1, s2)) {
        return AssertionSuccess();
    }

    return AssertionFailure() << "Expected: (" << s1_expression << ") != ("
                              << s2_expression << "), actual: "
                              << PrintToString(s1)
                              << " vs " << PrintToString(s2);
}

// Compares two C strings, ignoring case.  Returns true if and only if they have
// the same content.
//
// Unlike strcasecmp(), this function can handle NULL argument(s).  A
// NULL C string is considered different to any non-NULL C string,
// including the empty string.
bool String::CaseInsensitiveCStringEquals(const char *lhs, const char *rhs)
{
    if (lhs == nullptr)
        return rhs == nullptr;
    if (rhs == nullptr)
        return false;
    return posix::StrCaseCmp(lhs, rhs) == 0;
}

// Compares two wide C strings, ignoring case.  Returns true if and only if they
// have the same content.
//
// Unlike wcscasecmp(), this function can handle NULL argument(s).
// A NULL C string is considered different to any non-NULL wide C string,
// including the empty string.
// NB: The implementations on different platforms slightly differ.
// On windows, this method uses _wcsicmp which compares according to LC_CTYPE
// environment variable. On GNU platform this method uses wcscasecmp
// which compares according to LC_CTYPE category of the current locale.
// On MacOS X, it uses towlower, which also uses LC_CTYPE category of the
// current locale.
bool String::CaseInsensitiveWideCStringEquals(const wchar_t *lhs,
                                              const wchar_t *rhs)
{
    if (lhs == nullptr)
        return rhs == nullptr;

    if (rhs == nullptr)
        return false;

#if GTEST_OS_WINDOWS
    return _wcsicmp(lhs, rhs) == 0;
#elif GTEST_OS_LINUX && !GTEST_OS_LINUX_ANDROID
    return wcscasecmp(lhs, rhs) == 0;
#else
    // Android, Mac OS X and Cygwin don't define wcscasecmp.
    // Other unknown OSes may not define it either.
    wint_t left, right;
    do {
        left = towlower(static_cast<wint_t>(*lhs++));
        right = towlower(static_cast<wint_t>(*rhs++));
    } while (left && left == right);
    return left == right;
#endif // OS selector
}

// Returns true if and only if str ends with the given suffix, ignoring case.
// Any string is considered to end with an empty suffix.
bool String::EndsWithCaseInsensitive(
    const std::string &str, const std::string &suffix)
{
    const size_t str_len = str.length();
    const size_t suffix_len = suffix.length();
    return (str_len >= suffix_len) && CaseInsensitiveCStringEquals(str.c_str() + str_len - suffix_len, suffix.c_str());
}

// Formats an int value as "%02d".
std::string String::FormatIntWidth2(int value)
{
    std::stringstream ss;
    ss << std::setfill('0') << std::setw(2) << value;
    return ss.str();
}

// Formats an int value as "%X".
std::string String::FormatHexUInt32(uint32_t value)
{
    std::stringstream ss;
    ss << std::hex << std::uppercase << value;
    return ss.str();
}

// Formats an int value as "%X".
std::string String::FormatHexInt(int value)
{
    return FormatHexUInt32(static_cast<uint32_t>(value));
}

// Formats a byte as "%02X".
std::string String::FormatByte(unsigned char value)
{
    std::stringstream ss;
    ss << std::setfill('0') << std::setw(2) << std::hex << std::uppercase
       << static_cast<unsigned int>(value);
    return ss.str();
}

// Converts the buffer in a stringstream to an std::string, converting NUL
// bytes to "\\0" along the way.
std::string StringStreamToString(::std::stringstream *ss)
{
    const ::std::string &str = ss->str();
    const char *const start = str.c_str();
    const char *const end = start + str.length();

    std::string result;
    result.reserve(static_cast<size_t>(2 * (end - start)));
    for (const char *ch = start; ch != end; ++ch) {
        if (*ch == '\0') {
            result += "\\0"; // Replaces NUL with "\\0";
        } else {
            result += *ch;
        }
    }

    return result;
}

// Appends the user-supplied message to the Google-Test-generated message.
std::string AppendUserMessage(const std::string &gtest_msg,
                              const Message &user_msg)
{
    // Appends the user message if it's non-empty.
    const std::string user_msg_string = user_msg.GetString();
    if (user_msg_string.empty()) {
        return gtest_msg;
    }
    if (gtest_msg.empty()) {
        return user_msg_string;
    }
    return gtest_msg + "\n" + user_msg_string;
}

} // namespace internal

// class TestResult

// Creates an empty TestResult.
TestResult::TestResult()
    : death_test_count_(0)
    , start_timestamp_(0)
    , elapsed_time_(0)
{
}

// D'tor.
TestResult::~TestResult()
{
}

// Returns the i-th test part result among all the results. i can
// range from 0 to total_part_count() - 1. If i is not in that range,
// aborts the program.
const TestPartResult &TestResult::GetTestPartResult(int i) const
{
    if (i < 0 || i >= total_part_count())
        internal::posix::Abort();
    return test_part_results_.at(static_cast<size_t>(i));
}

// Returns the i-th test property. i can range from 0 to
// test_property_count() - 1. If i is not in that range, aborts the
// program.
const TestProperty &TestResult::GetTestProperty(int i) const
{
    if (i < 0 || i >= test_property_count())
        internal::posix::Abort();
    return test_properties_.at(static_cast<size_t>(i));
}

// Clears the test part results.
void TestResult::ClearTestPartResults()
{
    test_part_results_.clear();
}

// Adds a test part result to the list.
void TestResult::AddTestPartResult(const TestPartResult &test_part_result)
{
    test_part_results_.push_back(test_part_result);
}

// Adds a test property to the list. If a property with the same key as the
// supplied property is already represented, the value of this test_property
// replaces the old value for that key.
void TestResult::RecordProperty(const std::string &xml_element,
                                const TestProperty &test_property)
{
    if (!ValidateTestProperty(xml_element, test_property)) {
        return;
    }
    internal::MutexLock lock(&test_properites_mutex_);
    const std::vector<TestProperty>::iterator property_with_matching_key =
        std::find_if(test_properties_.begin(), test_properties_.end(),
                     internal::TestPropertyKeyIs(test_property.key()));
    if (property_with_matching_key == test_properties_.end()) {
        test_properties_.push_back(test_property);
        return;
    }
    property_with_matching_key->SetValue(test_property.value());
}

// The list of reserved attributes used in the <testsuites> element of XML
// output.
static const char *const kReservedTestSuitesAttributes[] = {
    "disabled",
    "errors",
    "failures",
    "name",
    "random_seed",
    "tests",
    "time",
    "timestamp"};

// The list of reserved attributes used in the <testsuite> element of XML
// output.
static const char *const kReservedTestSuiteAttributes[] = {
    "disabled", "errors", "failures", "name",
    "tests", "time", "timestamp", "skipped"};

// The list of reserved attributes used in the <testcase> element of XML output.
static const char *const kReservedTestCaseAttributes[] = {
    "classname", "name", "status", "time", "type_param",
    "value_param", "file", "line"};

// Use a slightly different set for allowed output to ensure existing tests can
// still RecordProperty("result") or "RecordProperty(timestamp")
static const char *const kReservedOutputTestCaseAttributes[] = {
    "classname", "name", "status", "time", "type_param",
    "value_param", "file", "line", "result", "timestamp"};

template<size_t kSize>
std::vector<std::string> ArrayAsVector(const char *const (&array)[kSize])
{
    return std::vector<std::string>(array, array + kSize);
}

static std::vector<std::string> GetReservedAttributesForElement(
    const std::string &xml_element)
{
    if (xml_element == "testsuites") {
        return ArrayAsVector(kReservedTestSuitesAttributes);
    } else if (xml_element == "testsuite") {
        return ArrayAsVector(kReservedTestSuiteAttributes);
    } else if (xml_element == "testcase") {
        return ArrayAsVector(kReservedTestCaseAttributes);
    } else {
        GTEST_CHECK_(false) << "Unrecognized xml_element provided: " << xml_element;
    }
    // This code is unreachable but some compilers may not realizes that.
    return std::vector<std::string>();
}

// TODO(jdesprez): Merge the two getReserved attributes once skip is improved
static std::vector<std::string> GetReservedOutputAttributesForElement(
    const std::string &xml_element)
{
    if (xml_element == "testsuites") {
        return ArrayAsVector(kReservedTestSuitesAttributes);
    } else if (xml_element == "testsuite") {
        return ArrayAsVector(kReservedTestSuiteAttributes);
    } else if (xml_element == "testcase") {
        return ArrayAsVector(kReservedOutputTestCaseAttributes);
    } else {
        GTEST_CHECK_(false) << "Unrecognized xml_element provided: " << xml_element;
    }
    // This code is unreachable but some compilers may not realizes that.
    return std::vector<std::string>();
}

static std::string FormatWordList(const std::vector<std::string> &words)
{
    Message word_list;
    for (size_t i = 0; i < words.size(); ++i) {
        if (i > 0 && words.size() > 2) {
            word_list << ", ";
        }
        if (i == words.size() - 1) {
            word_list << "and ";
        }
        word_list << "'" << words[i] << "'";
    }
    return word_list.GetString();
}

static bool ValidateTestPropertyName(
    const std::string &property_name,
    const std::vector<std::string> &reserved_names)
{
    if (std::find(reserved_names.begin(), reserved_names.end(), property_name) != reserved_names.end()) {
        ADD_FAILURE() << "Reserved key used in RecordProperty(): " << property_name
                      << " (" << FormatWordList(reserved_names)
                      << " are reserved by " << GTEST_NAME_ << ")";
        return false;
    }
    return true;
}

// Adds a failure if the key is a reserved attribute of the element named
// xml_element.  Returns true if the property is valid.
bool TestResult::ValidateTestProperty(const std::string &xml_element,
                                      const TestProperty &test_property)
{
    return ValidateTestPropertyName(test_property.key(),
                                    GetReservedAttributesForElement(xml_element));
}

// Clears the object.
void TestResult::Clear()
{
    test_part_results_.clear();
    test_properties_.clear();
    death_test_count_ = 0;
    elapsed_time_ = 0;
}

// Returns true off the test part was skipped.
static bool TestPartSkipped(const TestPartResult &result)
{
    return result.skipped();
}

// Returns true if and only if the test was skipped.
bool TestResult::Skipped() const
{
    return !Failed() && CountIf(test_part_results_, TestPartSkipped) > 0;
}

// Returns true if and only if the test failed.
bool TestResult::Failed() const
{
    for (int i = 0; i < total_part_count(); ++i) {
        if (GetTestPartResult(i).failed())
            return true;
    }
    return false;
}

// Returns true if and only if the test part fatally failed.
static bool TestPartFatallyFailed(const TestPartResult &result)
{
    return result.fatally_failed();
}

// Returns true if and only if the test fatally failed.
bool TestResult::HasFatalFailure() const
{
    return CountIf(test_part_results_, TestPartFatallyFailed) > 0;
}

// Returns true if and only if the test part non-fatally failed.
static bool TestPartNonfatallyFailed(const TestPartResult &result)
{
    return result.nonfatally_failed();
}

// Returns true if and only if the test has a non-fatal failure.
bool TestResult::HasNonfatalFailure() const
{
    return CountIf(test_part_results_, TestPartNonfatallyFailed) > 0;
}

// Gets the number of all test parts.  This is the sum of the number
// of successful test parts and the number of failed test parts.
int TestResult::total_part_count() const
{
    return static_cast<int>(test_part_results_.size());
}

// Returns the number of the test properties.
int TestResult::test_property_count() const
{
    return static_cast<int>(test_properties_.size());
}

// class Test

// Creates a Test object.

// The c'tor saves the states of all flags.
Test::Test()
    : gtest_flag_saver_(new GTEST_FLAG_SAVER_)
{
}

// The d'tor restores the states of all flags.  The actual work is
// done by the d'tor of the gtest_flag_saver_ field, and thus not
// visible here.
Test::~Test()
{
}

// Sets up the test fixture.
//
// A sub-class may override this.
void Test::SetUp()
{
}

// Tears down the test fixture.
//
// A sub-class may override this.
void Test::TearDown()
{
}

// Allows user supplied key value pairs to be recorded for later output.
void Test::RecordProperty(const std::string &key, const std::string &value)
{
    UnitTest::GetInstance()->RecordProperty(key, value);
}

// Allows user supplied key value pairs to be recorded for later output.
void Test::RecordProperty(const std::string &key, int value)
{
    Message value_message;
    value_message << value;
    RecordProperty(key, value_message.GetString().c_str());
}

namespace internal {

void ReportFailureInUnknownLocation(TestPartResult::Type result_type,
                                    const std::string &message)
{
    // This function is a friend of UnitTest and as such has access to
    // AddTestPartResult.
    UnitTest::GetInstance()->AddTestPartResult(
        result_type,
        nullptr, // No info about the source file where the exception occurred.
        -1, // We have no info on which line caused the exception.
        message,
        ""); // No stack trace, either.
}

} // namespace internal

// Google Test requires all tests in the same test suite to use the same test
// fixture class.  This function checks if the current test has the
// same fixture class as the first test in the current test suite.  If
// yes, it returns true; otherwise it generates a Google Test failure and
// returns false.
bool Test::HasSameFixtureClass()
{
    internal::UnitTestImpl *const impl = internal::GetUnitTestImpl();
    const TestSuite *const test_suite = impl->current_test_suite();

    // Info about the first test in the current test suite.
    const TestInfo *const first_test_info = test_suite->test_info_list()[0];
    const internal::TypeId first_fixture_id = first_test_info->fixture_class_id_;
    const char *const first_test_name = first_test_info->name();

    // Info about the current test.
    const TestInfo *const this_test_info = impl->current_test_info();
    const internal::TypeId this_fixture_id = this_test_info->fixture_class_id_;
    const char *const this_test_name = this_test_info->name();

    if (this_fixture_id != first_fixture_id) {
        // Is the first test defined using TEST?
        const bool first_is_TEST = first_fixture_id == internal::GetTestTypeId();
        // Is this test defined using TEST?
        const bool this_is_TEST = this_fixture_id == internal::GetTestTypeId();

        if (first_is_TEST || this_is_TEST) {
            // Both TEST and TEST_F appear in same test suite, which is incorrect.
            // Tell the user how to fix this.

            // Gets the name of the TEST and the name of the TEST_F.  Note
            // that first_is_TEST and this_is_TEST cannot both be true, as
            // the fixture IDs are different for the two tests.
            const char *const TEST_name =
                first_is_TEST ? first_test_name : this_test_name;
            const char *const TEST_F_name =
                first_is_TEST ? this_test_name : first_test_name;

            ADD_FAILURE()
                << "All tests in the same test suite must use the same test fixture\n"
                << "class, so mixing TEST_F and TEST in the same test suite is\n"
                << "illegal.  In test suite " << this_test_info->test_suite_name()
                << ",\n"
                << "test " << TEST_F_name << " is defined using TEST_F but\n"
                << "test " << TEST_name << " is defined using TEST.  You probably\n"
                << "want to change the TEST to TEST_F or move it to another test\n"
                << "case.";
        } else {
            // Two fixture classes with the same name appear in two different
            // namespaces, which is not allowed. Tell the user how to fix this.
            ADD_FAILURE()
                << "All tests in the same test suite must use the same test fixture\n"
                << "class.  However, in test suite "
                << this_test_info->test_suite_name() << ",\n"
                << "you defined test " << first_test_name << " and test "
                << this_test_name << "\n"
                << "using two different test fixture classes.  This can happen if\n"
                << "the two classes are from different namespaces or translation\n"
                << "units and have the same name.  You should probably rename one\n"
                << "of the classes to put the tests into different test suites.";
        }
        return false;
    }

    return true;
}

#if GTEST_HAS_SEH

// Adds an "exception thrown" fatal failure to the current test.  This
// function returns its result via an output parameter pointer because VC++
// prohibits creation of objects with destructors on stack in functions
// using __try (see error C2712).
static std::string *FormatSehExceptionMessage(DWORD exception_code,
                                              const char *location)
{
    Message message;
    message << "SEH exception with code 0x" << std::setbase(16) << exception_code << std::setbase(10) << " thrown in " << location << ".";

    return new std::string(message.GetString());
}

#endif // GTEST_HAS_SEH

namespace internal {

#if GTEST_HAS_EXCEPTIONS

// Adds an "exception thrown" fatal failure to the current test.
static std::string FormatCxxExceptionMessage(const char *description,
                                             const char *location)
{
    Message message;
    if (description != nullptr) {
        message << "C++ exception with description \"" << description << "\"";
    } else {
        message << "Unknown C++ exception";
    }
    message << " thrown in " << location << ".";

    return message.GetString();
}

static std::string PrintTestPartResultToString(
    const TestPartResult &test_part_result);

GoogleTestFailureException::GoogleTestFailureException(
    const TestPartResult &failure)
    : ::std::runtime_error(PrintTestPartResultToString(failure).c_str())
{
}

#endif // GTEST_HAS_EXCEPTIONS

// We put these helper functions in the internal namespace as IBM's xlC
// compiler rejects the code if they were declared static.

// Runs the given method and handles SEH exceptions it throws, when
// SEH is supported; returns the 0-value for type Result in case of an
// SEH exception.  (Microsoft compilers cannot handle SEH and C++
// exceptions in the same function.  Therefore, we provide a separate
// wrapper function for handling SEH exceptions.)
template<class T, typename Result>
Result HandleSehExceptionsInMethodIfSupported(
    T *object, Result (T::*method)(), const char *location)
{
#if GTEST_HAS_SEH
    __try {
        return (object->*method)();
    } __except (internal::UnitTestOptions::GTestShouldProcessSEH( // NOLINT
        GetExceptionCode())) {
        // We create the exception message on the heap because VC++ prohibits
        // creation of objects with destructors on stack in functions using __try
        // (see error C2712).
        std::string *exception_message = FormatSehExceptionMessage(
            GetExceptionCode(), location);
        internal::ReportFailureInUnknownLocation(TestPartResult::kFatalFailure,
                                                 *exception_message);
        delete exception_message;
        return static_cast<Result>(0);
    }
#else
    (void)location;
    return (object->*method)();
#endif // GTEST_HAS_SEH
}

// Runs the given method and catches and reports C++ and/or SEH-style
// exceptions, if they are supported; returns the 0-value for type
// Result in case of an SEH exception.
template<class T, typename Result>
Result HandleExceptionsInMethodIfSupported(
    T *object, Result (T::*method)(), const char *location)
{
    // NOTE: The user code can affect the way in which Google Test handles
    // exceptions by setting GTEST_FLAG(catch_exceptions), but only before
    // RUN_ALL_TESTS() starts. It is technically possible to check the flag
    // after the exception is caught and either report or re-throw the
    // exception based on the flag's value:
    //
    // try {
    //   // Perform the test method.
    // } catch (...) {
    //   if (GTEST_FLAG(catch_exceptions))
    //     // Report the exception as failure.
    //   else
    //     throw;  // Re-throws the original exception.
    // }
    //
    // However, the purpose of this flag is to allow the program to drop into
    // the debugger when the exception is thrown. On most platforms, once the
    // control enters the catch block, the exception origin information is
    // lost and the debugger will stop the program at the point of the
    // re-throw in this function -- instead of at the point of the original
    // throw statement in the code under test.  For this reason, we perform
    // the check early, sacrificing the ability to affect Google Test's
    // exception handling in the method where the exception is thrown.
    if (internal::GetUnitTestImpl()->catch_exceptions()) {
#if GTEST_HAS_EXCEPTIONS
        try {
            return HandleSehExceptionsInMethodIfSupported(object, method, location);
        } catch (const AssertionException &) { // NOLINT
            // This failure was reported already.
        } catch (const internal::GoogleTestFailureException &) { // NOLINT
            // This exception type can only be thrown by a failed Google
            // Test assertion with the intention of letting another testing
            // framework catch it.  Therefore we just re-throw it.
            throw;
        } catch (const std::exception &e) { // NOLINT
            internal::ReportFailureInUnknownLocation(
                TestPartResult::kFatalFailure,
                FormatCxxExceptionMessage(e.what(), location));
        } catch (...) { // NOLINT
            internal::ReportFailureInUnknownLocation(
                TestPartResult::kFatalFailure,
                FormatCxxExceptionMessage(nullptr, location));
        }
        return static_cast<Result>(0);
#else
        return HandleSehExceptionsInMethodIfSupported(object, method, location);
#endif // GTEST_HAS_EXCEPTIONS
    } else {
        return (object->*method)();
    }
}

} // namespace internal

// Runs the test and updates the test result.
void Test::Run()
{
    if (!HasSameFixtureClass())
        return;

    internal::UnitTestImpl *const impl = internal::GetUnitTestImpl();
    impl->os_stack_trace_getter()->UponLeavingGTest();
    internal::HandleExceptionsInMethodIfSupported(this, &Test::SetUp, "SetUp()");
    // We will run the test only if SetUp() was successful and didn't call
    // GTEST_SKIP().
    if (!HasFatalFailure() && !IsSkipped()) {
        impl->os_stack_trace_getter()->UponLeavingGTest();
        internal::HandleExceptionsInMethodIfSupported(
            this, &Test::TestBody, "the test body");
    }

    // However, we want to clean up as much as possible.  Hence we will
    // always call TearDown(), even if SetUp() or the test body has
    // failed.
    impl->os_stack_trace_getter()->UponLeavingGTest();
    internal::HandleExceptionsInMethodIfSupported(
        this, &Test::TearDown, "TearDown()");
}

// Returns true if and only if the current test has a fatal failure.
bool Test::HasFatalFailure()
{
    return internal::GetUnitTestImpl()->current_test_result()->HasFatalFailure();
}

// Returns true if and only if the current test has a non-fatal failure.
bool Test::HasNonfatalFailure()
{
    return internal::GetUnitTestImpl()->current_test_result()->HasNonfatalFailure();
}

// Returns true if and only if the current test was skipped.
bool Test::IsSkipped()
{
    return internal::GetUnitTestImpl()->current_test_result()->Skipped();
}

// class TestInfo

// Constructs a TestInfo object. It assumes ownership of the test factory
// object.
TestInfo::TestInfo(const std::string &a_test_suite_name,
                   const std::string &a_name, const char *a_type_param,
                   const char *a_value_param,
                   internal::CodeLocation a_code_location,
                   internal::TypeId fixture_class_id,
                   internal::TestFactoryBase *factory)
    : test_suite_name_(a_test_suite_name)
    , name_(a_name)
    , type_param_(a_type_param ? new std::string(a_type_param) : nullptr)
    , value_param_(a_value_param ? new std::string(a_value_param) : nullptr)
    , location_(a_code_location)
    , fixture_class_id_(fixture_class_id)
    , should_run_(false)
    , is_disabled_(false)
    , matches_filter_(false)
    , factory_(factory)
    , result_()
{
}

// Destructs a TestInfo object.
TestInfo::~TestInfo()
{
    delete factory_;
}

namespace internal {

// Creates a new TestInfo object and registers it with Google Test;
// returns the created object.
//
// Arguments:
//
//   test_suite_name:   name of the test suite
//   name:             name of the test
//   type_param:       the name of the test's type parameter, or NULL if
//                     this is not a typed or a type-parameterized test.
//   value_param:      text representation of the test's value parameter,
//                     or NULL if this is not a value-parameterized test.
//   code_location:    code location where the test is defined
//   fixture_class_id: ID of the test fixture class
//   set_up_tc:        pointer to the function that sets up the test suite
//   tear_down_tc:     pointer to the function that tears down the test suite
//   factory:          pointer to the factory that creates a test object.
//                     The newly created TestInfo instance will assume
//                     ownership of the factory object.
TestInfo *MakeAndRegisterTestInfo(
    const char *test_suite_name, const char *name, const char *type_param,
    const char *value_param, CodeLocation code_location,
    TypeId fixture_class_id, SetUpTestSuiteFunc set_up_tc,
    TearDownTestSuiteFunc tear_down_tc, TestFactoryBase *factory)
{
    TestInfo *const test_info =
        new TestInfo(test_suite_name, name, type_param, value_param,
                     code_location, fixture_class_id, factory);
    GetUnitTestImpl()->AddTestInfo(set_up_tc, tear_down_tc, test_info);
    return test_info;
}

void ReportInvalidTestSuiteType(const char *test_suite_name,
                                CodeLocation code_location)
{
    Message errors;
    errors
        << "Attempted redefinition of test suite " << test_suite_name << ".\n"
        << "All tests in the same test suite must use the same test fixture\n"
        << "class.  However, in test suite " << test_suite_name << ", you tried\n"
        << "to define a test using a fixture class different from the one\n"
        << "used earlier. This can happen if the two fixture classes are\n"
        << "from different namespaces and have the same name. You should\n"
        << "probably rename one of the classes to put the tests into different\n"
        << "test suites.";

    GTEST_LOG_(ERROR) << FormatFileLocation(code_location.file.c_str(),
                                            code_location.line)
                      << " " << errors.GetString();
}
} // namespace internal

namespace {

// A predicate that checks the test name of a TestInfo against a known
// value.
//
// This is used for implementation of the TestSuite class only.  We put
// it in the anonymous namespace to prevent polluting the outer
// namespace.
//
// TestNameIs is copyable.
class TestNameIs
{
public:
    // Constructor.
    //
    // TestNameIs has NO default constructor.
    explicit TestNameIs(const char *name)
        : name_(name)
    {
    }

    // Returns true if and only if the test name of test_info matches name_.
    bool operator()(const TestInfo *test_info) const
    {
        return test_info && test_info->name() == name_;
    }

private:
    std::string name_;
};

} // namespace

namespace internal {

// This method expands all parameterized tests registered with macros TEST_P
// and INSTANTIATE_TEST_SUITE_P into regular tests and registers those.
// This will be done just once during the program runtime.
void UnitTestImpl::RegisterParameterizedTests()
{
    if (!parameterized_tests_registered_) {
        parameterized_test_registry_.RegisterTests();
        type_parameterized_test_registry_.CheckForInstantiations();
        parameterized_tests_registered_ = true;
    }
}

} // namespace internal

// Creates the test object, runs it, records its result, and then
// deletes it.
void TestInfo::Run()
{
    if (!should_run_)
        return;

    // Tells UnitTest where to store test result.
    internal::UnitTestImpl *const impl = internal::GetUnitTestImpl();
    impl->set_current_test_info(this);

    TestEventListener *repeater = UnitTest::GetInstance()->listeners().repeater();

    // Notifies the unit test event listeners that a test is about to start.
    repeater->OnTestStart(*this);

    const TimeInMillis start = internal::GetTimeInMillis();

    impl->os_stack_trace_getter()->UponLeavingGTest();

    // Creates the test object.
    Test *const test = internal::HandleExceptionsInMethodIfSupported(
        factory_, &internal::TestFactoryBase::CreateTest,
        "the test fixture's constructor");

    // Runs the test if the constructor didn't generate a fatal failure or invoke
    // GTEST_SKIP().
    // Note that the object will not be null
    if (!Test::HasFatalFailure() && !Test::IsSkipped()) {
        // This doesn't throw as all user code that can throw are wrapped into
        // exception handling code.
        test->Run();
    }

    if (test != nullptr) {
        // Deletes the test object.
        impl->os_stack_trace_getter()->UponLeavingGTest();
        internal::HandleExceptionsInMethodIfSupported(
            test, &Test::DeleteSelf_, "the test fixture's destructor");
    }

    result_.set_start_timestamp(start);
    result_.set_elapsed_time(internal::GetTimeInMillis() - start);

    // Notifies the unit test event listener that a test has just finished.
    repeater->OnTestEnd(*this);

    // Tells UnitTest to stop associating assertion results to this
    // test.
    impl->set_current_test_info(nullptr);
}

// Skip and records a skipped test result for this object.
void TestInfo::Skip()
{
    if (!should_run_)
        return;

    internal::UnitTestImpl *const impl = internal::GetUnitTestImpl();
    impl->set_current_test_info(this);

    TestEventListener *repeater = UnitTest::GetInstance()->listeners().repeater();

    // Notifies the unit test event listeners that a test is about to start.
    repeater->OnTestStart(*this);

    const TestPartResult test_part_result =
        TestPartResult(TestPartResult::kSkip, this->file(), this->line(), "");
    impl->GetTestPartResultReporterForCurrentThread()->ReportTestPartResult(
        test_part_result);

    // Notifies the unit test event listener that a test has just finished.
    repeater->OnTestEnd(*this);
    impl->set_current_test_info(nullptr);
}

// class TestSuite

// Gets the number of successful tests in this test suite.
int TestSuite::successful_test_count() const
{
    return CountIf(test_info_list_, TestPassed);
}

// Gets the number of successful tests in this test suite.
int TestSuite::skipped_test_count() const
{
    return CountIf(test_info_list_, TestSkipped);
}

// Gets the number of failed tests in this test suite.
int TestSuite::failed_test_count() const
{
    return CountIf(test_info_list_, TestFailed);
}

// Gets the number of disabled tests that will be reported in the XML report.
int TestSuite::reportable_disabled_test_count() const
{
    return CountIf(test_info_list_, TestReportableDisabled);
}

// Gets the number of disabled tests in this test suite.
int TestSuite::disabled_test_count() const
{
    return CountIf(test_info_list_, TestDisabled);
}

// Gets the number of tests to be printed in the XML report.
int TestSuite::reportable_test_count() const
{
    return CountIf(test_info_list_, TestReportable);
}

// Get the number of tests in this test suite that should run.
int TestSuite::test_to_run_count() const
{
    return CountIf(test_info_list_, ShouldRunTest);
}

// Gets the number of all tests.
int TestSuite::total_test_count() const
{
    return static_cast<int>(test_info_list_.size());
}

// Creates a TestSuite with the given name.
//
// Arguments:
//
//   name:         name of the test suite
//   a_type_param: the name of the test suite's type parameter, or NULL if
//                 this is not a typed or a type-parameterized test suite.
//   set_up_tc:    pointer to the function that sets up the test suite
//   tear_down_tc: pointer to the function that tears down the test suite
TestSuite::TestSuite(const char *a_name, const char *a_type_param,
                     internal::SetUpTestSuiteFunc set_up_tc,
                     internal::TearDownTestSuiteFunc tear_down_tc)
    : name_(a_name)
    , type_param_(a_type_param ? new std::string(a_type_param) : nullptr)
    , set_up_tc_(set_up_tc)
    , tear_down_tc_(tear_down_tc)
    , should_run_(false)
    , start_timestamp_(0)
    , elapsed_time_(0)
{
}

// Destructor of TestSuite.
TestSuite::~TestSuite()
{
    // Deletes every Test in the collection.
    ForEach(test_info_list_, internal::Delete<TestInfo>);
}

// Returns the i-th test among all the tests. i can range from 0 to
// total_test_count() - 1. If i is not in that range, returns NULL.
const TestInfo *TestSuite::GetTestInfo(int i) const
{
    const int index = GetElementOr(test_indices_, i, -1);
    return index < 0 ? nullptr : test_info_list_[static_cast<size_t>(index)];
}

// Returns the i-th test among all the tests. i can range from 0 to
// total_test_count() - 1. If i is not in that range, returns NULL.
TestInfo *TestSuite::GetMutableTestInfo(int i)
{
    const int index = GetElementOr(test_indices_, i, -1);
    return index < 0 ? nullptr : test_info_list_[static_cast<size_t>(index)];
}

// Adds a test to this test suite.  Will delete the test upon
// destruction of the TestSuite object.
void TestSuite::AddTestInfo(TestInfo *test_info)
{
    test_info_list_.push_back(test_info);
    test_indices_.push_back(static_cast<int>(test_indices_.size()));
}

// Runs every test in this TestSuite.
void TestSuite::Run()
{
    if (!should_run_)
        return;

    internal::UnitTestImpl *const impl = internal::GetUnitTestImpl();
    impl->set_current_test_suite(this);

    TestEventListener *repeater = UnitTest::GetInstance()->listeners().repeater();

    // Call both legacy and the new API
    repeater->OnTestSuiteStart(*this);
//  Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI
    repeater->OnTestCaseStart(*this);
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI

    impl->os_stack_trace_getter()->UponLeavingGTest();
    internal::HandleExceptionsInMethodIfSupported(
        this, &TestSuite::RunSetUpTestSuite, "SetUpTestSuite()");

    start_timestamp_ = internal::GetTimeInMillis();
    for (int i = 0; i < total_test_count(); i++) {
        GetMutableTestInfo(i)->Run();
        if (GTEST_FLAG(fail_fast) && GetMutableTestInfo(i)->result()->Failed()) {
            for (int j = i + 1; j < total_test_count(); j++) {
                GetMutableTestInfo(j)->Skip();
            }
            break;
        }
    }
    elapsed_time_ = internal::GetTimeInMillis() - start_timestamp_;

    impl->os_stack_trace_getter()->UponLeavingGTest();
    internal::HandleExceptionsInMethodIfSupported(
        this, &TestSuite::RunTearDownTestSuite, "TearDownTestSuite()");

    // Call both legacy and the new API
    repeater->OnTestSuiteEnd(*this);
//  Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI
    repeater->OnTestCaseEnd(*this);
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI

    impl->set_current_test_suite(nullptr);
}

// Skips all tests under this TestSuite.
void TestSuite::Skip()
{
    if (!should_run_)
        return;

    internal::UnitTestImpl *const impl = internal::GetUnitTestImpl();
    impl->set_current_test_suite(this);

    TestEventListener *repeater = UnitTest::GetInstance()->listeners().repeater();

    // Call both legacy and the new API
    repeater->OnTestSuiteStart(*this);
//  Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI
    repeater->OnTestCaseStart(*this);
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI

    for (int i = 0; i < total_test_count(); i++) {
        GetMutableTestInfo(i)->Skip();
    }

    // Call both legacy and the new API
    repeater->OnTestSuiteEnd(*this);
    // Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI
    repeater->OnTestCaseEnd(*this);
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI

    impl->set_current_test_suite(nullptr);
}

// Clears the results of all tests in this test suite.
void TestSuite::ClearResult()
{
    ad_hoc_test_result_.Clear();
    ForEach(test_info_list_, TestInfo::ClearTestResult);
}

// Shuffles the tests in this test suite.
void TestSuite::ShuffleTests(internal::Random *random)
{
    Shuffle(random, &test_indices_);
}

// Restores the test order to before the first shuffle.
void TestSuite::UnshuffleTests()
{
    for (size_t i = 0; i < test_indices_.size(); i++) {
        test_indices_[i] = static_cast<int>(i);
    }
}

// Formats a countable noun.  Depending on its quantity, either the
// singular form or the plural form is used. e.g.
//
// FormatCountableNoun(1, "formula", "formuli") returns "1 formula".
// FormatCountableNoun(5, "book", "books") returns "5 books".
static std::string FormatCountableNoun(int count,
                                       const char *singular_form,
                                       const char *plural_form)
{
    return internal::StreamableToString(count) + " " + (count == 1 ? singular_form : plural_form);
}

// Formats the count of tests.
static std::string FormatTestCount(int test_count)
{
    return FormatCountableNoun(test_count, "test", "tests");
}

// Formats the count of test suites.
static std::string FormatTestSuiteCount(int test_suite_count)
{
    return FormatCountableNoun(test_suite_count, "test suite", "test suites");
}

// Converts a TestPartResult::Type enum to human-friendly string
// representation.  Both kNonFatalFailure and kFatalFailure are translated
// to "Failure", as the user usually doesn't care about the difference
// between the two when viewing the test result.
static const char *TestPartResultTypeToString(TestPartResult::Type type)
{
    switch (type) {
    case TestPartResult::kSkip:
        return "Skipped\n";
    case TestPartResult::kSuccess:
        return "Success";

    case TestPartResult::kNonFatalFailure:
    case TestPartResult::kFatalFailure:
#ifdef _MSC_VER
        return "error: ";
#else
        return "Failure\n";
#endif
    default:
        return "Unknown result type";
    }
}

namespace internal {
namespace {
enum class GTestColor { kDefault,
                        kRed,
                        kGreen,
                        kYellow };
} // namespace

// Prints a TestPartResult to an std::string.
static std::string PrintTestPartResultToString(
    const TestPartResult &test_part_result)
{
    return (Message()
            << internal::FormatFileLocation(test_part_result.file_name(),
                                            test_part_result.line_number())
            << " " << TestPartResultTypeToString(test_part_result.type())
            << test_part_result.message())
        .GetString();
}

// Prints a TestPartResult.
static void PrintTestPartResult(const TestPartResult &test_part_result)
{
    const std::string &result =
        PrintTestPartResultToString(test_part_result);
    printf("%s\n", result.c_str());
    fflush(stdout);
    // If the test program runs in Visual Studio or a debugger, the
    // following statements add the test part result message to the Output
    // window such that the user can double-click on it to jump to the
    // corresponding source code location; otherwise they do nothing.
#if GTEST_OS_WINDOWS && !GTEST_OS_WINDOWS_MOBILE
    // We don't call OutputDebugString*() on Windows Mobile, as printing
    // to stdout is done by OutputDebugString() there already - we don't
    // want the same message printed twice.
    ::OutputDebugStringA(result.c_str());
    ::OutputDebugStringA("\n");
#endif
}

// class PrettyUnitTestResultPrinter
#if GTEST_OS_WINDOWS && !GTEST_OS_WINDOWS_MOBILE && !GTEST_OS_WINDOWS_PHONE && !GTEST_OS_WINDOWS_RT && !GTEST_OS_WINDOWS_MINGW

// Returns the character attribute for the given color.
static WORD GetColorAttribute(GTestColor color)
{
    switch (color) {
    case GTestColor::kRed:
        return FOREGROUND_RED;
    case GTestColor::kGreen:
        return FOREGROUND_GREEN;
    case GTestColor::kYellow:
        return FOREGROUND_RED | FOREGROUND_GREEN;
    default:
        return 0;
    }
}

static int GetBitOffset(WORD color_mask)
{
    if (color_mask == 0)
        return 0;

    int bitOffset = 0;
    while ((color_mask & 1) == 0) {
        color_mask >>= 1;
        ++bitOffset;
    }
    return bitOffset;
}

static WORD GetNewColor(GTestColor color, WORD old_color_attrs)
{
    // Let's reuse the BG
    static const WORD background_mask = BACKGROUND_BLUE | BACKGROUND_GREEN | BACKGROUND_RED | BACKGROUND_INTENSITY;
    static const WORD foreground_mask = FOREGROUND_BLUE | FOREGROUND_GREEN | FOREGROUND_RED | FOREGROUND_INTENSITY;
    const WORD existing_bg = old_color_attrs & background_mask;

    WORD new_color =
        GetColorAttribute(color) | existing_bg | FOREGROUND_INTENSITY;
    static const int bg_bitOffset = GetBitOffset(background_mask);
    static const int fg_bitOffset = GetBitOffset(foreground_mask);

    if (((new_color & background_mask) >> bg_bitOffset) == ((new_color & foreground_mask) >> fg_bitOffset)) {
        new_color ^= FOREGROUND_INTENSITY; // invert intensity
    }
    return new_color;
}

#else

// Returns the ANSI color code for the given color. GTestColor::kDefault is
// an invalid input.
static const char *GetAnsiColorCode(GTestColor color)
{
    switch (color) {
    case GTestColor::kRed:
        return "1";
    case GTestColor::kGreen:
        return "2";
    case GTestColor::kYellow:
        return "3";
    default:
        return nullptr;
    }
}

#endif // GTEST_OS_WINDOWS && !GTEST_OS_WINDOWS_MOBILE

// Returns true if and only if Google Test should use colors in the output.
bool ShouldUseColor(bool stdout_is_tty)
{
    const char *const gtest_color = GTEST_FLAG(color).c_str();

    if (String::CaseInsensitiveCStringEquals(gtest_color, "auto")) {
#if GTEST_OS_WINDOWS && !GTEST_OS_WINDOWS_MINGW
        // On Windows the TERM variable is usually not set, but the
        // console there does support colors.
        return stdout_is_tty;
#else
        // On non-Windows platforms, we rely on the TERM variable.
        const char *const term = posix::GetEnv("TERM");
        const bool term_supports_color =
            String::CStringEquals(term, "xterm") || String::CStringEquals(term, "xterm-color") || String::CStringEquals(term, "xterm-256color") || String::CStringEquals(term, "screen") || String::CStringEquals(term, "screen-256color") || String::CStringEquals(term, "tmux") || String::CStringEquals(term, "tmux-256color") || String::CStringEquals(term, "rxvt-unicode") || String::CStringEquals(term, "rxvt-unicode-256color") || String::CStringEquals(term, "linux") || String::CStringEquals(term, "cygwin");
        return stdout_is_tty && term_supports_color;
#endif // GTEST_OS_WINDOWS
    }

    return String::CaseInsensitiveCStringEquals(gtest_color, "yes") || String::CaseInsensitiveCStringEquals(gtest_color, "true") || String::CaseInsensitiveCStringEquals(gtest_color, "t") || String::CStringEquals(gtest_color, "1");
    // We take "yes", "true", "t", and "1" as meaning "yes".  If the
    // value is neither one of these nor "auto", we treat it as "no" to
    // be conservative.
}

// Helpers for printing colored strings to stdout. Note that on Windows, we
// cannot simply emit special characters and have the terminal change colors.
// This routine must actually emit the characters rather than return a string
// that would be colored when printed, as can be done on Linux.

void ColoredPrintf(GTestColor color, const char *fmt, ...)
{
    va_list args;
    va_start(args, fmt);

#if GTEST_OS_WINDOWS_MOBILE || GTEST_OS_ZOS || GTEST_OS_IOS || GTEST_OS_WINDOWS_PHONE || GTEST_OS_WINDOWS_RT || defined(ESP_PLATFORM)
    const bool use_color = AlwaysFalse();
#else
    static const bool in_color_mode =
        ShouldUseColor(posix::IsATTY(posix::FileNo(stdout)) != 0);
    const bool use_color = in_color_mode && (color != GTestColor::kDefault);
#endif // GTEST_OS_WINDOWS_MOBILE || GTEST_OS_ZOS

    if (!use_color) {
        vprintf(fmt, args);
        va_end(args);
        return;
    }

#if GTEST_OS_WINDOWS && !GTEST_OS_WINDOWS_MOBILE && !GTEST_OS_WINDOWS_PHONE && !GTEST_OS_WINDOWS_RT && !GTEST_OS_WINDOWS_MINGW
    const HANDLE stdout_handle = GetStdHandle(STD_OUTPUT_HANDLE);

    // Gets the current text color.
    CONSOLE_SCREEN_BUFFER_INFO buffer_info;
    GetConsoleScreenBufferInfo(stdout_handle, &buffer_info);
    const WORD old_color_attrs = buffer_info.wAttributes;
    const WORD new_color = GetNewColor(color, old_color_attrs);

    // We need to flush the stream buffers into the console before each
    // SetConsoleTextAttribute call lest it affect the text that is already
    // printed but has not yet reached the console.
    fflush(stdout);
    SetConsoleTextAttribute(stdout_handle, new_color);

    vprintf(fmt, args);

    fflush(stdout);
    // Restores the text color.
    SetConsoleTextAttribute(stdout_handle, old_color_attrs);
#else
    printf("\033[0;3%sm", GetAnsiColorCode(color));
    vprintf(fmt, args);
    printf("\033[m"); // Resets the terminal to default.
#endif // GTEST_OS_WINDOWS && !GTEST_OS_WINDOWS_MOBILE
    va_end(args);
}

// Text printed in Google Test's text output and --gtest_list_tests
// output to label the type parameter and value parameter for a test.
static const char kTypeParamLabel[] = "TypeParam";
static const char kValueParamLabel[] = "GetParam()";

static void PrintFullTestCommentIfPresent(const TestInfo &test_info)
{
    const char *const type_param = test_info.type_param();
    const char *const value_param = test_info.value_param();

    if (type_param != nullptr || value_param != nullptr) {
        printf(", where ");
        if (type_param != nullptr) {
            printf("%s = %s", kTypeParamLabel, type_param);
            if (value_param != nullptr)
                printf(" and ");
        }
        if (value_param != nullptr) {
            printf("%s = %s", kValueParamLabel, value_param);
        }
    }
}

// This class implements the TestEventListener interface.
//
// Class PrettyUnitTestResultPrinter is copyable.
class PrettyUnitTestResultPrinter : public TestEventListener
{
public:
    PrettyUnitTestResultPrinter() {}
    static void PrintTestName(const char *test_suite, const char *test)
    {
        printf("%s.%s", test_suite, test);
    }

    // The following methods override what's in the TestEventListener class.
    void OnTestProgramStart(const UnitTest & /*unit_test*/) override {}
    void OnTestIterationStart(const UnitTest &unit_test, int iteration) override;
    void OnEnvironmentsSetUpStart(const UnitTest &unit_test) override;
    void OnEnvironmentsSetUpEnd(const UnitTest & /*unit_test*/) override {}
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
    void OnTestCaseStart(const TestCase &test_case) override;
#else
    void OnTestSuiteStart(const TestSuite &test_suite) override;
#endif // OnTestCaseStart

    void OnTestStart(const TestInfo &test_info) override;

    void OnTestPartResult(const TestPartResult &result) override;
    void OnTestEnd(const TestInfo &test_info) override;
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
    void OnTestCaseEnd(const TestCase &test_case) override;
#else
    void OnTestSuiteEnd(const TestSuite &test_suite) override;
#endif // GTEST_REMOVE_LEGACY_TEST_CASEAPI_

    void OnEnvironmentsTearDownStart(const UnitTest &unit_test) override;
    void OnEnvironmentsTearDownEnd(const UnitTest & /*unit_test*/) override {}
    void OnTestIterationEnd(const UnitTest &unit_test, int iteration) override;
    void OnTestProgramEnd(const UnitTest & /*unit_test*/) override {}

private:
    static void PrintFailedTests(const UnitTest &unit_test);
    static void PrintFailedTestSuites(const UnitTest &unit_test);
    static void PrintSkippedTests(const UnitTest &unit_test);
};

// Fired before each iteration of tests starts.
void PrettyUnitTestResultPrinter::OnTestIterationStart(
    const UnitTest &unit_test, int iteration)
{
    if (GTEST_FLAG(repeat) != 1)
        printf("\nRepeating all tests (iteration %d) . . .\n\n", iteration + 1);

    const char *const filter = GTEST_FLAG(filter).c_str();

    // Prints the filter if it's not *.  This reminds the user that some
    // tests may be skipped.
    if (!String::CStringEquals(filter, kUniversalFilter)) {
        ColoredPrintf(GTestColor::kYellow, "Note: %s filter = %s\n", GTEST_NAME_,
                      filter);
    }

    if (internal::ShouldShard(kTestTotalShards, kTestShardIndex, false)) {
        const int32_t shard_index = Int32FromEnvOrDie(kTestShardIndex, -1);
        ColoredPrintf(GTestColor::kYellow, "Note: This is test shard %d of %s.\n",
                      static_cast<int>(shard_index) + 1,
                      internal::posix::GetEnv(kTestTotalShards));
    }

    if (GTEST_FLAG(shuffle)) {
        ColoredPrintf(GTestColor::kYellow,
                      "Note: Randomizing tests' orders with a seed of %d .\n",
                      unit_test.random_seed());
    }

    ColoredPrintf(GTestColor::kGreen, "[==========] ");
    printf("Running %s from %s.\n",
           FormatTestCount(unit_test.test_to_run_count()).c_str(),
           FormatTestSuiteCount(unit_test.test_suite_to_run_count()).c_str());
    fflush(stdout);
}

void PrettyUnitTestResultPrinter::OnEnvironmentsSetUpStart(
    const UnitTest & /*unit_test*/)
{
    ColoredPrintf(GTestColor::kGreen, "[----------] ");
    printf("Global test environment set-up.\n");
    fflush(stdout);
}

#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
void PrettyUnitTestResultPrinter::OnTestCaseStart(const TestCase &test_case)
{
    const std::string counts =
        FormatCountableNoun(test_case.test_to_run_count(), "test", "tests");
    ColoredPrintf(GTestColor::kGreen, "[----------] ");
    printf("%s from %s", counts.c_str(), test_case.name());
    if (test_case.type_param() == nullptr) {
        printf("\n");
    } else {
        printf(", where %s = %s\n", kTypeParamLabel, test_case.type_param());
    }
    fflush(stdout);
}
#else
void PrettyUnitTestResultPrinter::OnTestSuiteStart(
    const TestSuite &test_suite)
{
    const std::string counts =
        FormatCountableNoun(test_suite.test_to_run_count(), "test", "tests");
    ColoredPrintf(GTestColor::kGreen, "[----------] ");
    printf("%s from %s", counts.c_str(), test_suite.name());
    if (test_suite.type_param() == nullptr) {
        printf("\n");
    } else {
        printf(", where %s = %s\n", kTypeParamLabel, test_suite.type_param());
    }
    fflush(stdout);
}
#endif // GTEST_REMOVE_LEGACY_TEST_CASEAPI_

void PrettyUnitTestResultPrinter::OnTestStart(const TestInfo &test_info)
{
    ColoredPrintf(GTestColor::kGreen, "[ RUN      ] ");
    PrintTestName(test_info.test_suite_name(), test_info.name());
    printf("\n");
    fflush(stdout);
}

// Called after an assertion failure.
void PrettyUnitTestResultPrinter::OnTestPartResult(
    const TestPartResult &result)
{
    switch (result.type()) {
    // If the test part succeeded, we don't need to do anything.
    case TestPartResult::kSuccess:
        return;
    default:
        // Print failure message from the assertion
        // (e.g. expected this and got that).
        PrintTestPartResult(result);
        fflush(stdout);
    }
}

void PrettyUnitTestResultPrinter::OnTestEnd(const TestInfo &test_info)
{
    if (test_info.result()->Passed()) {
        ColoredPrintf(GTestColor::kGreen, "[       OK ] ");
    } else if (test_info.result()->Skipped()) {
        ColoredPrintf(GTestColor::kGreen, "[  SKIPPED ] ");
    } else {
        ColoredPrintf(GTestColor::kRed, "[  FAILED  ] ");
    }
    PrintTestName(test_info.test_suite_name(), test_info.name());
    if (test_info.result()->Failed())
        PrintFullTestCommentIfPresent(test_info);

    if (GTEST_FLAG(print_time)) {
        printf(" (%s ms)\n", internal::StreamableToString(
                                 test_info.result()->elapsed_time())
                                 .c_str());
    } else {
        printf("\n");
    }
    fflush(stdout);
}

#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
void PrettyUnitTestResultPrinter::OnTestCaseEnd(const TestCase &test_case)
{
    if (!GTEST_FLAG(print_time))
        return;

    const std::string counts =
        FormatCountableNoun(test_case.test_to_run_count(), "test", "tests");
    ColoredPrintf(GTestColor::kGreen, "[----------] ");
    printf("%s from %s (%s ms total)\n\n", counts.c_str(), test_case.name(),
           internal::StreamableToString(test_case.elapsed_time()).c_str());
    fflush(stdout);
}
#else
void PrettyUnitTestResultPrinter::OnTestSuiteEnd(const TestSuite &test_suite)
{
    if (!GTEST_FLAG(print_time))
        return;

    const std::string counts =
        FormatCountableNoun(test_suite.test_to_run_count(), "test", "tests");
    ColoredPrintf(GTestColor::kGreen, "[----------] ");
    printf("%s from %s (%s ms total)\n\n", counts.c_str(), test_suite.name(),
           internal::StreamableToString(test_suite.elapsed_time()).c_str());
    fflush(stdout);
}
#endif // GTEST_REMOVE_LEGACY_TEST_CASEAPI_

void PrettyUnitTestResultPrinter::OnEnvironmentsTearDownStart(
    const UnitTest & /*unit_test*/)
{
    ColoredPrintf(GTestColor::kGreen, "[----------] ");
    printf("Global test environment tear-down\n");
    fflush(stdout);
}

// Internal helper for printing the list of failed tests.
void PrettyUnitTestResultPrinter::PrintFailedTests(const UnitTest &unit_test)
{
    const int failed_test_count = unit_test.failed_test_count();
    ColoredPrintf(GTestColor::kRed, "[  FAILED  ] ");
    printf("%s, listed below:\n", FormatTestCount(failed_test_count).c_str());

    for (int i = 0; i < unit_test.total_test_suite_count(); ++i) {
        const TestSuite &test_suite = *unit_test.GetTestSuite(i);
        if (!test_suite.should_run() || (test_suite.failed_test_count() == 0)) {
            continue;
        }
        for (int j = 0; j < test_suite.total_test_count(); ++j) {
            const TestInfo &test_info = *test_suite.GetTestInfo(j);
            if (!test_info.should_run() || !test_info.result()->Failed()) {
                continue;
            }
            ColoredPrintf(GTestColor::kRed, "[  FAILED  ] ");
            printf("%s.%s", test_suite.name(), test_info.name());
            PrintFullTestCommentIfPresent(test_info);
            printf("\n");
        }
    }
    printf("\n%2d FAILED %s\n", failed_test_count,
           failed_test_count == 1 ? "TEST" : "TESTS");
}

// Internal helper for printing the list of test suite failures not covered by
// PrintFailedTests.
void PrettyUnitTestResultPrinter::PrintFailedTestSuites(
    const UnitTest &unit_test)
{
    int suite_failure_count = 0;
    for (int i = 0; i < unit_test.total_test_suite_count(); ++i) {
        const TestSuite &test_suite = *unit_test.GetTestSuite(i);
        if (!test_suite.should_run()) {
            continue;
        }
        if (test_suite.ad_hoc_test_result().Failed()) {
            ColoredPrintf(GTestColor::kRed, "[  FAILED  ] ");
            printf("%s: SetUpTestSuite or TearDownTestSuite\n", test_suite.name());
            ++suite_failure_count;
        }
    }
    if (suite_failure_count > 0) {
        printf("\n%2d FAILED TEST %s\n", suite_failure_count,
               suite_failure_count == 1 ? "SUITE" : "SUITES");
    }
}

// Internal helper for printing the list of skipped tests.
void PrettyUnitTestResultPrinter::PrintSkippedTests(const UnitTest &unit_test)
{
    const int skipped_test_count = unit_test.skipped_test_count();
    if (skipped_test_count == 0) {
        return;
    }

    for (int i = 0; i < unit_test.total_test_suite_count(); ++i) {
        const TestSuite &test_suite = *unit_test.GetTestSuite(i);
        if (!test_suite.should_run() || (test_suite.skipped_test_count() == 0)) {
            continue;
        }
        for (int j = 0; j < test_suite.total_test_count(); ++j) {
            const TestInfo &test_info = *test_suite.GetTestInfo(j);
            if (!test_info.should_run() || !test_info.result()->Skipped()) {
                continue;
            }
            ColoredPrintf(GTestColor::kGreen, "[  SKIPPED ] ");
            printf("%s.%s", test_suite.name(), test_info.name());
            printf("\n");
        }
    }
}

void PrettyUnitTestResultPrinter::OnTestIterationEnd(const UnitTest &unit_test,
                                                     int /*iteration*/)
{
    ColoredPrintf(GTestColor::kGreen, "[==========] ");
    printf("%s from %s ran.",
           FormatTestCount(unit_test.test_to_run_count()).c_str(),
           FormatTestSuiteCount(unit_test.test_suite_to_run_count()).c_str());
    if (GTEST_FLAG(print_time)) {
        printf(" (%s ms total)",
               internal::StreamableToString(unit_test.elapsed_time()).c_str());
    }
    printf("\n");
    ColoredPrintf(GTestColor::kGreen, "[  PASSED  ] ");
    printf("%s.\n", FormatTestCount(unit_test.successful_test_count()).c_str());

    const int skipped_test_count = unit_test.skipped_test_count();
    if (skipped_test_count > 0) {
        ColoredPrintf(GTestColor::kGreen, "[  SKIPPED ] ");
        printf("%s, listed below:\n", FormatTestCount(skipped_test_count).c_str());
        PrintSkippedTests(unit_test);
    }

    if (!unit_test.Passed()) {
        PrintFailedTests(unit_test);
        PrintFailedTestSuites(unit_test);
    }

    int num_disabled = unit_test.reportable_disabled_test_count();
    if (num_disabled && !GTEST_FLAG(also_run_disabled_tests)) {
        if (unit_test.Passed()) {
            printf("\n"); // Add a spacer if no FAILURE banner is displayed.
        }
        ColoredPrintf(GTestColor::kYellow, "  YOU HAVE %d DISABLED %s\n\n",
                      num_disabled, num_disabled == 1 ? "TEST" : "TESTS");
    }
    // Ensure that Google Test output is printed before, e.g., heapchecker output.
    fflush(stdout);
}

// End PrettyUnitTestResultPrinter

// This class implements the TestEventListener interface.
//
// Class BriefUnitTestResultPrinter is copyable.
class BriefUnitTestResultPrinter : public TestEventListener
{
public:
    BriefUnitTestResultPrinter() {}
    static void PrintTestName(const char *test_suite, const char *test)
    {
        printf("%s.%s", test_suite, test);
    }

    // The following methods override what's in the TestEventListener class.
    void OnTestProgramStart(const UnitTest & /*unit_test*/) override {}
    void OnTestIterationStart(const UnitTest & /*unit_test*/,
                              int /*iteration*/) override {}
    void OnEnvironmentsSetUpStart(const UnitTest & /*unit_test*/) override {}
    void OnEnvironmentsSetUpEnd(const UnitTest & /*unit_test*/) override {}
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
    void OnTestCaseStart(const TestCase & /*test_case*/) override
    {
    }
#else
    void OnTestSuiteStart(const TestSuite & /*test_suite*/) override
    {
    }
#endif // OnTestCaseStart

    void OnTestStart(const TestInfo & /*test_info*/) override
    {
    }

    void OnTestPartResult(const TestPartResult &result) override;
    void OnTestEnd(const TestInfo &test_info) override;
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
    void OnTestCaseEnd(const TestCase & /*test_case*/) override
    {
    }
#else
    void OnTestSuiteEnd(const TestSuite & /*test_suite*/) override
    {
    }
#endif // GTEST_REMOVE_LEGACY_TEST_CASEAPI_

    void OnEnvironmentsTearDownStart(const UnitTest & /*unit_test*/) override
    {
    }
    void OnEnvironmentsTearDownEnd(const UnitTest & /*unit_test*/) override {}
    void OnTestIterationEnd(const UnitTest &unit_test, int iteration) override;
    void OnTestProgramEnd(const UnitTest & /*unit_test*/) override {}
};

// Called after an assertion failure.
void BriefUnitTestResultPrinter::OnTestPartResult(
    const TestPartResult &result)
{
    switch (result.type()) {
    // If the test part succeeded, we don't need to do anything.
    case TestPartResult::kSuccess:
        return;
    default:
        // Print failure message from the assertion
        // (e.g. expected this and got that).
        PrintTestPartResult(result);
        fflush(stdout);
    }
}

void BriefUnitTestResultPrinter::OnTestEnd(const TestInfo &test_info)
{
    if (test_info.result()->Failed()) {
        ColoredPrintf(GTestColor::kRed, "[  FAILED  ] ");
        PrintTestName(test_info.test_suite_name(), test_info.name());
        PrintFullTestCommentIfPresent(test_info);

        if (GTEST_FLAG(print_time)) {
            printf(" (%s ms)\n",
                   internal::StreamableToString(test_info.result()->elapsed_time())
                       .c_str());
        } else {
            printf("\n");
        }
        fflush(stdout);
    }
}

void BriefUnitTestResultPrinter::OnTestIterationEnd(const UnitTest &unit_test,
                                                    int /*iteration*/)
{
    ColoredPrintf(GTestColor::kGreen, "[==========] ");
    printf("%s from %s ran.",
           FormatTestCount(unit_test.test_to_run_count()).c_str(),
           FormatTestSuiteCount(unit_test.test_suite_to_run_count()).c_str());
    if (GTEST_FLAG(print_time)) {
        printf(" (%s ms total)",
               internal::StreamableToString(unit_test.elapsed_time()).c_str());
    }
    printf("\n");
    ColoredPrintf(GTestColor::kGreen, "[  PASSED  ] ");
    printf("%s.\n", FormatTestCount(unit_test.successful_test_count()).c_str());

    const int skipped_test_count = unit_test.skipped_test_count();
    if (skipped_test_count > 0) {
        ColoredPrintf(GTestColor::kGreen, "[  SKIPPED ] ");
        printf("%s.\n", FormatTestCount(skipped_test_count).c_str());
    }

    int num_disabled = unit_test.reportable_disabled_test_count();
    if (num_disabled && !GTEST_FLAG(also_run_disabled_tests)) {
        if (unit_test.Passed()) {
            printf("\n"); // Add a spacer if no FAILURE banner is displayed.
        }
        ColoredPrintf(GTestColor::kYellow, "  YOU HAVE %d DISABLED %s\n\n",
                      num_disabled, num_disabled == 1 ? "TEST" : "TESTS");
    }
    // Ensure that Google Test output is printed before, e.g., heapchecker output.
    fflush(stdout);
}

// End BriefUnitTestResultPrinter

// class TestEventRepeater
//
// This class forwards events to other event listeners.
class TestEventRepeater : public TestEventListener
{
public:
    TestEventRepeater()
        : forwarding_enabled_(true)
    {
    }
    ~TestEventRepeater() override;
    void Append(TestEventListener *listener);
    TestEventListener *Release(TestEventListener *listener);

    // Controls whether events will be forwarded to listeners_. Set to false
    // in death test child processes.
    bool forwarding_enabled() const { return forwarding_enabled_; }
    void set_forwarding_enabled(bool enable) { forwarding_enabled_ = enable; }

    void OnTestProgramStart(const UnitTest &unit_test) override;
    void OnTestIterationStart(const UnitTest &unit_test, int iteration) override;
    void OnEnvironmentsSetUpStart(const UnitTest &unit_test) override;
    void OnEnvironmentsSetUpEnd(const UnitTest &unit_test) override;
//  Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
    void OnTestCaseStart(const TestSuite &parameter) override;
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI_
    void OnTestSuiteStart(const TestSuite &parameter) override;
    void OnTestStart(const TestInfo &test_info) override;
    void OnTestPartResult(const TestPartResult &result) override;
    void OnTestEnd(const TestInfo &test_info) override;
//  Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
    void OnTestCaseEnd(const TestCase &parameter) override;
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI_
    void OnTestSuiteEnd(const TestSuite &parameter) override;
    void OnEnvironmentsTearDownStart(const UnitTest &unit_test) override;
    void OnEnvironmentsTearDownEnd(const UnitTest &unit_test) override;
    void OnTestIterationEnd(const UnitTest &unit_test, int iteration) override;
    void OnTestProgramEnd(const UnitTest &unit_test) override;

private:
    // Controls whether events will be forwarded to listeners_. Set to false
    // in death test child processes.
    bool forwarding_enabled_;
    // The list of listeners that receive events.
    std::vector<TestEventListener *> listeners_;

    GTEST_DISALLOW_COPY_AND_ASSIGN_(TestEventRepeater);
};

TestEventRepeater::~TestEventRepeater()
{
    ForEach(listeners_, Delete<TestEventListener>);
}

void TestEventRepeater::Append(TestEventListener *listener)
{
    listeners_.push_back(listener);
}

TestEventListener *TestEventRepeater::Release(TestEventListener *listener)
{
    for (size_t i = 0; i < listeners_.size(); ++i) {
        if (listeners_[i] == listener) {
            listeners_.erase(listeners_.begin() + static_cast<int>(i));
            return listener;
        }
    }

    return nullptr;
}

// Since most methods are very similar, use macros to reduce boilerplate.
// This defines a member that forwards the call to all listeners.
#define GTEST_REPEATER_METHOD_(Name, Type) \
    void TestEventRepeater::Name(const Type &parameter) \
    { \
        if (forwarding_enabled_) { \
            for (size_t i = 0; i < listeners_.size(); i++) { \
                listeners_[i]->Name(parameter); \
            } \
        } \
    }
// This defines a member that forwards the call to all listeners in reverse
// order.
#define GTEST_REVERSE_REPEATER_METHOD_(Name, Type) \
    void TestEventRepeater::Name(const Type &parameter) \
    { \
        if (forwarding_enabled_) { \
            for (size_t i = listeners_.size(); i != 0; i--) { \
                listeners_[i - 1]->Name(parameter); \
            } \
        } \
    }

GTEST_REPEATER_METHOD_(OnTestProgramStart, UnitTest)
GTEST_REPEATER_METHOD_(OnEnvironmentsSetUpStart, UnitTest)
//  Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
GTEST_REPEATER_METHOD_(OnTestCaseStart, TestSuite)
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI_
GTEST_REPEATER_METHOD_(OnTestSuiteStart, TestSuite)
GTEST_REPEATER_METHOD_(OnTestStart, TestInfo)
GTEST_REPEATER_METHOD_(OnTestPartResult, TestPartResult)
GTEST_REPEATER_METHOD_(OnEnvironmentsTearDownStart, UnitTest)
GTEST_REVERSE_REPEATER_METHOD_(OnEnvironmentsSetUpEnd, UnitTest)
GTEST_REVERSE_REPEATER_METHOD_(OnEnvironmentsTearDownEnd, UnitTest)
GTEST_REVERSE_REPEATER_METHOD_(OnTestEnd, TestInfo)
//  Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
GTEST_REVERSE_REPEATER_METHOD_(OnTestCaseEnd, TestSuite)
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI_
GTEST_REVERSE_REPEATER_METHOD_(OnTestSuiteEnd, TestSuite)
GTEST_REVERSE_REPEATER_METHOD_(OnTestProgramEnd, UnitTest)

#undef GTEST_REPEATER_METHOD_
#undef GTEST_REVERSE_REPEATER_METHOD_

void TestEventRepeater::OnTestIterationStart(const UnitTest &unit_test,
                                             int iteration)
{
    if (forwarding_enabled_) {
        for (size_t i = 0; i < listeners_.size(); i++) {
            listeners_[i]->OnTestIterationStart(unit_test, iteration);
        }
    }
}

void TestEventRepeater::OnTestIterationEnd(const UnitTest &unit_test,
                                           int iteration)
{
    if (forwarding_enabled_) {
        for (size_t i = listeners_.size(); i > 0; i--) {
            listeners_[i - 1]->OnTestIterationEnd(unit_test, iteration);
        }
    }
}

// End TestEventRepeater

// This class generates an XML output file.
class XmlUnitTestResultPrinter : public EmptyTestEventListener
{
public:
    explicit XmlUnitTestResultPrinter(const char *output_file);

    void OnTestIterationEnd(const UnitTest &unit_test, int iteration) override;
    void ListTestsMatchingFilter(const std::vector<TestSuite *> &test_suites);

    // Prints an XML summary of all unit tests.
    static void PrintXmlTestsList(std::ostream *stream,
                                  const std::vector<TestSuite *> &test_suites);

private:
    // Is c a whitespace character that is normalized to a space character
    // when it appears in an XML attribute value?
    static bool IsNormalizableWhitespace(char c)
    {
        return c == 0x9 || c == 0xA || c == 0xD;
    }

    // May c appear in a well-formed XML document?
    static bool IsValidXmlCharacter(char c)
    {
        return IsNormalizableWhitespace(c) || c >= 0x20;
    }

    // Returns an XML-escaped copy of the input string str.  If
    // is_attribute is true, the text is meant to appear as an attribute
    // value, and normalizable whitespace is preserved by replacing it
    // with character references.
    static std::string EscapeXml(const std::string &str, bool is_attribute);

    // Returns the given string with all characters invalid in XML removed.
    static std::string RemoveInvalidXmlCharacters(const std::string &str);

    // Convenience wrapper around EscapeXml when str is an attribute value.
    static std::string EscapeXmlAttribute(const std::string &str)
    {
        return EscapeXml(str, true);
    }

    // Convenience wrapper around EscapeXml when str is not an attribute value.
    static std::string EscapeXmlText(const char *str)
    {
        return EscapeXml(str, false);
    }

    // Verifies that the given attribute belongs to the given element and
    // streams the attribute as XML.
    static void OutputXmlAttribute(std::ostream *stream,
                                   const std::string &element_name,
                                   const std::string &name,
                                   const std::string &value);

    // Streams an XML CDATA section, escaping invalid CDATA sequences as needed.
    static void OutputXmlCDataSection(::std::ostream *stream, const char *data);

    // Streams an XML representation of a TestInfo object.
    static void OutputXmlTestInfo(::std::ostream *stream,
                                  const char *test_suite_name,
                                  const TestInfo &test_info);

    // Prints an XML representation of a TestSuite object
    static void PrintXmlTestSuite(::std::ostream *stream,
                                  const TestSuite &test_suite);

    // Prints an XML summary of unit_test to output stream out.
    static void PrintXmlUnitTest(::std::ostream *stream,
                                 const UnitTest &unit_test);

    // Produces a string representing the test properties in a result as space
    // delimited XML attributes based on the property key="value" pairs.
    // When the std::string is not empty, it includes a space at the beginning,
    // to delimit this attribute from prior attributes.
    static std::string TestPropertiesAsXmlAttributes(const TestResult &result);

    // Streams an XML representation of the test properties of a TestResult
    // object.
    static void OutputXmlTestProperties(std::ostream *stream,
                                        const TestResult &result);

    // The output file.
    const std::string output_file_;

    GTEST_DISALLOW_COPY_AND_ASSIGN_(XmlUnitTestResultPrinter);
};

// Creates a new XmlUnitTestResultPrinter.
XmlUnitTestResultPrinter::XmlUnitTestResultPrinter(const char *output_file)
    : output_file_(output_file)
{
    if (output_file_.empty()) {
        GTEST_LOG_(FATAL) << "XML output file may not be null";
    }
}

// Called after the unit test ends.
void XmlUnitTestResultPrinter::OnTestIterationEnd(const UnitTest &unit_test,
                                                  int /*iteration*/)
{
    FILE *xmlout = OpenFileForWriting(output_file_);
    std::stringstream stream;
    PrintXmlUnitTest(&stream, unit_test);
    fprintf(xmlout, "%s", StringStreamToString(&stream).c_str());
    fclose(xmlout);
}

void XmlUnitTestResultPrinter::ListTestsMatchingFilter(
    const std::vector<TestSuite *> &test_suites)
{
    FILE *xmlout = OpenFileForWriting(output_file_);
    std::stringstream stream;
    PrintXmlTestsList(&stream, test_suites);
    fprintf(xmlout, "%s", StringStreamToString(&stream).c_str());
    fclose(xmlout);
}

// Returns an XML-escaped copy of the input string str.  If is_attribute
// is true, the text is meant to appear as an attribute value, and
// normalizable whitespace is preserved by replacing it with character
// references.
//
// Invalid XML characters in str, if any, are stripped from the output.
// It is expected that most, if not all, of the text processed by this
// module will consist of ordinary English text.
// If this module is ever modified to produce version 1.1 XML output,
// most invalid characters can be retained using character references.
std::string XmlUnitTestResultPrinter::EscapeXml(
    const std::string &str, bool is_attribute)
{
    Message m;

    for (size_t i = 0; i < str.size(); ++i) {
        const char ch = str[i];
        switch (ch) {
        case '<':
            m << "&lt;";
            break;
        case '>':
            m << "&gt;";
            break;
        case '&':
            m << "&amp;";
            break;
        case '\'':
            if (is_attribute)
                m << "&apos;";
            else
                m << '\'';
            break;
        case '"':
            if (is_attribute)
                m << "&quot;";
            else
                m << '"';
            break;
        default:
            if (IsValidXmlCharacter(ch)) {
                if (is_attribute && IsNormalizableWhitespace(ch))
                    m << "&#x" << String::FormatByte(static_cast<unsigned char>(ch))
                      << ";";
                else
                    m << ch;
            }
            break;
        }
    }

    return m.GetString();
}

// Returns the given string with all characters invalid in XML removed.
// Currently invalid characters are dropped from the string. An
// alternative is to replace them with certain characters such as . or ?.
std::string XmlUnitTestResultPrinter::RemoveInvalidXmlCharacters(
    const std::string &str)
{
    std::string output;
    output.reserve(str.size());
    for (std::string::const_iterator it = str.begin(); it != str.end(); ++it)
        if (IsValidXmlCharacter(*it))
            output.push_back(*it);

    return output;
}

// The following routines generate an XML representation of a UnitTest
// object.
// GOOGLETEST_CM0009 DO NOT DELETE
//
// This is how Google Test concepts map to the DTD:
//
// <testsuites name="AllTests">        <-- corresponds to a UnitTest object
//   <testsuite name="testcase-name">  <-- corresponds to a TestSuite object
//     <testcase name="test-name">     <-- corresponds to a TestInfo object
//       <failure message="...">...</failure>
//       <failure message="...">...</failure>
//       <failure message="...">...</failure>
//                                     <-- individual assertion failures
//     </testcase>
//   </testsuite>
// </testsuites>

// Formats the given time in milliseconds as seconds.
std::string FormatTimeInMillisAsSeconds(TimeInMillis ms)
{
    ::std::stringstream ss;
    ss << (static_cast<double>(ms) * 1e-3);
    return ss.str();
}

static bool PortableLocaltime(time_t seconds, struct tm *out)
{
#if defined(_MSC_VER)
    return localtime_s(out, &seconds) == 0;
#elif defined(__MINGW32__) || defined(__MINGW64__)
    // MINGW <time.h> provides neither localtime_r nor localtime_s, but uses
    // Windows' localtime(), which has a thread-local tm buffer.
    struct tm *tm_ptr = localtime(&seconds); // NOLINT
    if (tm_ptr == nullptr)
        return false;
    *out = *tm_ptr;
    return true;
#else
    return localtime_r(&seconds, out) != nullptr;
#endif
}

// Converts the given epoch time in milliseconds to a date string in the ISO
// 8601 format, without the timezone information.
std::string FormatEpochTimeInMillisAsIso8601(TimeInMillis ms)
{
    struct tm time_struct;
    if (!PortableLocaltime(static_cast<time_t>(ms / 1000), &time_struct))
        return "";
    // YYYY-MM-DDThh:mm:ss
    return StreamableToString(time_struct.tm_year + 1900) + "-" + String::FormatIntWidth2(time_struct.tm_mon + 1) + "-" + String::FormatIntWidth2(time_struct.tm_mday) + "T" + String::FormatIntWidth2(time_struct.tm_hour) + ":" + String::FormatIntWidth2(time_struct.tm_min) + ":" + String::FormatIntWidth2(time_struct.tm_sec);
}

// Streams an XML CDATA section, escaping invalid CDATA sequences as needed.
void XmlUnitTestResultPrinter::OutputXmlCDataSection(::std::ostream *stream,
                                                     const char *data)
{
    const char *segment = data;
    *stream << "<![CDATA[";
    for (;;) {
        const char *const next_segment = strstr(segment, "]]>");
        if (next_segment != nullptr) {
            stream->write(
                segment, static_cast<std::streamsize>(next_segment - segment));
            *stream << "]]>]]&gt;<![CDATA[";
            segment = next_segment + strlen("]]>");
        } else {
            *stream << segment;
            break;
        }
    }
    *stream << "]]>";
}

void XmlUnitTestResultPrinter::OutputXmlAttribute(
    std::ostream *stream,
    const std::string &element_name,
    const std::string &name,
    const std::string &value)
{
    const std::vector<std::string> &allowed_names =
        GetReservedOutputAttributesForElement(element_name);

    GTEST_CHECK_(std::find(allowed_names.begin(), allowed_names.end(), name) != allowed_names.end())
        << "Attribute " << name << " is not allowed for element <" << element_name
        << ">.";

    *stream << " " << name << "=\"" << EscapeXmlAttribute(value) << "\"";
}

// Prints an XML representation of a TestInfo object.
void XmlUnitTestResultPrinter::OutputXmlTestInfo(::std::ostream *stream,
                                                 const char *test_suite_name,
                                                 const TestInfo &test_info)
{
    const TestResult &result = *test_info.result();
    const std::string kTestsuite = "testcase";

    if (test_info.is_in_another_shard()) {
        return;
    }

    *stream << "    <testcase";
    OutputXmlAttribute(stream, kTestsuite, "name", test_info.name());

    if (test_info.value_param() != nullptr) {
        OutputXmlAttribute(stream, kTestsuite, "value_param",
                           test_info.value_param());
    }
    if (test_info.type_param() != nullptr) {
        OutputXmlAttribute(stream, kTestsuite, "type_param",
                           test_info.type_param());
    }
    if (GTEST_FLAG(list_tests)) {
        OutputXmlAttribute(stream, kTestsuite, "file", test_info.file());
        OutputXmlAttribute(stream, kTestsuite, "line",
                           StreamableToString(test_info.line()));
        *stream << " />\n";
        return;
    }

    OutputXmlAttribute(stream, kTestsuite, "status",
                       test_info.should_run() ? "run" : "notrun");
    OutputXmlAttribute(stream, kTestsuite, "result",
                       test_info.should_run()
                           ? (result.Skipped() ? "skipped" : "completed")
                           : "suppressed");
    OutputXmlAttribute(stream, kTestsuite, "time",
                       FormatTimeInMillisAsSeconds(result.elapsed_time()));
    OutputXmlAttribute(
        stream, kTestsuite, "timestamp",
        FormatEpochTimeInMillisAsIso8601(result.start_timestamp()));
    OutputXmlAttribute(stream, kTestsuite, "classname", test_suite_name);

    int failures = 0;
    int skips = 0;
    for (int i = 0; i < result.total_part_count(); ++i) {
        const TestPartResult &part = result.GetTestPartResult(i);
        if (part.failed()) {
            if (++failures == 1 && skips == 0) {
                *stream << ">\n";
            }
            const std::string location =
                internal::FormatCompilerIndependentFileLocation(part.file_name(),
                                                                part.line_number());
            const std::string summary = location + "\n" + part.summary();
            *stream << "      <failure message=\""
                    << EscapeXmlAttribute(summary)
                    << "\" type=\"\">";
            const std::string detail = location + "\n" + part.message();
            OutputXmlCDataSection(stream, RemoveInvalidXmlCharacters(detail).c_str());
            *stream << "</failure>\n";
        } else if (part.skipped()) {
            if (++skips == 1 && failures == 0) {
                *stream << ">\n";
            }
            const std::string location =
                internal::FormatCompilerIndependentFileLocation(part.file_name(),
                                                                part.line_number());
            const std::string summary = location + "\n" + part.summary();
            *stream << "      <skipped message=\""
                    << EscapeXmlAttribute(summary.c_str()) << "\">";
            const std::string detail = location + "\n" + part.message();
            OutputXmlCDataSection(stream, RemoveInvalidXmlCharacters(detail).c_str());
            *stream << "</skipped>\n";
        }
    }

    if (failures == 0 && skips == 0 && result.test_property_count() == 0) {
        *stream << " />\n";
    } else {
        if (failures == 0 && skips == 0) {
            *stream << ">\n";
        }
        OutputXmlTestProperties(stream, result);
        *stream << "    </testcase>\n";
    }
}

// Prints an XML representation of a TestSuite object
void XmlUnitTestResultPrinter::PrintXmlTestSuite(std::ostream *stream,
                                                 const TestSuite &test_suite)
{
    const std::string kTestsuite = "testsuite";
    *stream << "  <" << kTestsuite;
    OutputXmlAttribute(stream, kTestsuite, "name", test_suite.name());
    OutputXmlAttribute(stream, kTestsuite, "tests",
                       StreamableToString(test_suite.reportable_test_count()));
    if (!GTEST_FLAG(list_tests)) {
        OutputXmlAttribute(stream, kTestsuite, "failures",
                           StreamableToString(test_suite.failed_test_count()));
        OutputXmlAttribute(
            stream, kTestsuite, "disabled",
            StreamableToString(test_suite.reportable_disabled_test_count()));
        OutputXmlAttribute(stream, kTestsuite, "skipped",
                           StreamableToString(test_suite.skipped_test_count()));

        OutputXmlAttribute(stream, kTestsuite, "errors", "0");

        OutputXmlAttribute(stream, kTestsuite, "time",
                           FormatTimeInMillisAsSeconds(test_suite.elapsed_time()));
        OutputXmlAttribute(
            stream, kTestsuite, "timestamp",
            FormatEpochTimeInMillisAsIso8601(test_suite.start_timestamp()));
        *stream << TestPropertiesAsXmlAttributes(test_suite.ad_hoc_test_result());
    }
    *stream << ">\n";
    for (int i = 0; i < test_suite.total_test_count(); ++i) {
        if (test_suite.GetTestInfo(i)->is_reportable())
            OutputXmlTestInfo(stream, test_suite.name(), *test_suite.GetTestInfo(i));
    }
    *stream << "  </" << kTestsuite << ">\n";
}

// Prints an XML summary of unit_test to output stream out.
void XmlUnitTestResultPrinter::PrintXmlUnitTest(std::ostream *stream,
                                                const UnitTest &unit_test)
{
    const std::string kTestsuites = "testsuites";

    *stream << "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n";
    *stream << "<" << kTestsuites;

    OutputXmlAttribute(stream, kTestsuites, "tests",
                       StreamableToString(unit_test.reportable_test_count()));
    OutputXmlAttribute(stream, kTestsuites, "failures",
                       StreamableToString(unit_test.failed_test_count()));
    OutputXmlAttribute(
        stream, kTestsuites, "disabled",
        StreamableToString(unit_test.reportable_disabled_test_count()));
    OutputXmlAttribute(stream, kTestsuites, "errors", "0");
    OutputXmlAttribute(stream, kTestsuites, "time",
                       FormatTimeInMillisAsSeconds(unit_test.elapsed_time()));
    OutputXmlAttribute(
        stream, kTestsuites, "timestamp",
        FormatEpochTimeInMillisAsIso8601(unit_test.start_timestamp()));

    if (GTEST_FLAG(shuffle)) {
        OutputXmlAttribute(stream, kTestsuites, "random_seed",
                           StreamableToString(unit_test.random_seed()));
    }
    *stream << TestPropertiesAsXmlAttributes(unit_test.ad_hoc_test_result());

    OutputXmlAttribute(stream, kTestsuites, "name", "AllTests");
    *stream << ">\n";

    for (int i = 0; i < unit_test.total_test_suite_count(); ++i) {
        if (unit_test.GetTestSuite(i)->reportable_test_count() > 0)
            PrintXmlTestSuite(stream, *unit_test.GetTestSuite(i));
    }
    *stream << "</" << kTestsuites << ">\n";
}

void XmlUnitTestResultPrinter::PrintXmlTestsList(
    std::ostream *stream, const std::vector<TestSuite *> &test_suites)
{
    const std::string kTestsuites = "testsuites";

    *stream << "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n";
    *stream << "<" << kTestsuites;

    int total_tests = 0;
    for (auto test_suite : test_suites) {
        total_tests += test_suite->total_test_count();
    }
    OutputXmlAttribute(stream, kTestsuites, "tests",
                       StreamableToString(total_tests));
    OutputXmlAttribute(stream, kTestsuites, "name", "AllTests");
    *stream << ">\n";

    for (auto test_suite : test_suites) {
        PrintXmlTestSuite(stream, *test_suite);
    }
    *stream << "</" << kTestsuites << ">\n";
}

// Produces a string representing the test properties in a result as space
// delimited XML attributes based on the property key="value" pairs.
std::string XmlUnitTestResultPrinter::TestPropertiesAsXmlAttributes(
    const TestResult &result)
{
    Message attributes;
    for (int i = 0; i < result.test_property_count(); ++i) {
        const TestProperty &property = result.GetTestProperty(i);
        attributes << " " << property.key() << "="
                   << "\"" << EscapeXmlAttribute(property.value()) << "\"";
    }
    return attributes.GetString();
}

void XmlUnitTestResultPrinter::OutputXmlTestProperties(
    std::ostream *stream, const TestResult &result)
{
    const std::string kProperties = "properties";
    const std::string kProperty = "property";

    if (result.test_property_count() <= 0) {
        return;
    }

    *stream << "<" << kProperties << ">\n";
    for (int i = 0; i < result.test_property_count(); ++i) {
        const TestProperty &property = result.GetTestProperty(i);
        *stream << "<" << kProperty;
        *stream << " name=\"" << EscapeXmlAttribute(property.key()) << "\"";
        *stream << " value=\"" << EscapeXmlAttribute(property.value()) << "\"";
        *stream << "/>\n";
    }
    *stream << "</" << kProperties << ">\n";
}

// End XmlUnitTestResultPrinter

// This class generates an JSON output file.
class JsonUnitTestResultPrinter : public EmptyTestEventListener
{
public:
    explicit JsonUnitTestResultPrinter(const char *output_file);

    void OnTestIterationEnd(const UnitTest &unit_test, int iteration) override;

    // Prints an JSON summary of all unit tests.
    static void PrintJsonTestList(::std::ostream *stream,
                                  const std::vector<TestSuite *> &test_suites);

private:
    // Returns an JSON-escaped copy of the input string str.
    static std::string EscapeJson(const std::string &str);

    //// Verifies that the given attribute belongs to the given element and
    //// streams the attribute as JSON.
    static void OutputJsonKey(std::ostream *stream,
                              const std::string &element_name,
                              const std::string &name,
                              const std::string &value,
                              const std::string &indent,
                              bool comma = true);
    static void OutputJsonKey(std::ostream *stream,
                              const std::string &element_name,
                              const std::string &name,
                              int value,
                              const std::string &indent,
                              bool comma = true);

    // Streams a JSON representation of a TestInfo object.
    static void OutputJsonTestInfo(::std::ostream *stream,
                                   const char *test_suite_name,
                                   const TestInfo &test_info);

    // Prints a JSON representation of a TestSuite object
    static void PrintJsonTestSuite(::std::ostream *stream,
                                   const TestSuite &test_suite);

    // Prints a JSON summary of unit_test to output stream out.
    static void PrintJsonUnitTest(::std::ostream *stream,
                                  const UnitTest &unit_test);

    // Produces a string representing the test properties in a result as
    // a JSON dictionary.
    static std::string TestPropertiesAsJson(const TestResult &result,
                                            const std::string &indent);

    // The output file.
    const std::string output_file_;

    GTEST_DISALLOW_COPY_AND_ASSIGN_(JsonUnitTestResultPrinter);
};

// Creates a new JsonUnitTestResultPrinter.
JsonUnitTestResultPrinter::JsonUnitTestResultPrinter(const char *output_file)
    : output_file_(output_file)
{
    if (output_file_.empty()) {
        GTEST_LOG_(FATAL) << "JSON output file may not be null";
    }
}

void JsonUnitTestResultPrinter::OnTestIterationEnd(const UnitTest &unit_test,
                                                   int /*iteration*/)
{
    FILE *jsonout = OpenFileForWriting(output_file_);
    std::stringstream stream;
    PrintJsonUnitTest(&stream, unit_test);
    fprintf(jsonout, "%s", StringStreamToString(&stream).c_str());
    fclose(jsonout);
}

// Returns an JSON-escaped copy of the input string str.
std::string JsonUnitTestResultPrinter::EscapeJson(const std::string &str)
{
    Message m;

    for (size_t i = 0; i < str.size(); ++i) {
        const char ch = str[i];
        switch (ch) {
        case '\\':
        case '"':
        case '/':
            m << '\\' << ch;
            break;
        case '\b':
            m << "\\b";
            break;
        case '\t':
            m << "\\t";
            break;
        case '\n':
            m << "\\n";
            break;
        case '\f':
            m << "\\f";
            break;
        case '\r':
            m << "\\r";
            break;
        default:
            if (ch < ' ') {
                m << "\\u00" << String::FormatByte(static_cast<unsigned char>(ch));
            } else {
                m << ch;
            }
            break;
        }
    }

    return m.GetString();
}

// The following routines generate an JSON representation of a UnitTest
// object.

// Formats the given time in milliseconds as seconds.
static std::string FormatTimeInMillisAsDuration(TimeInMillis ms)
{
    ::std::stringstream ss;
    ss << (static_cast<double>(ms) * 1e-3) << "s";
    return ss.str();
}

// Converts the given epoch time in milliseconds to a date string in the
// RFC3339 format, without the timezone information.
static std::string FormatEpochTimeInMillisAsRFC3339(TimeInMillis ms)
{
    struct tm time_struct;
    if (!PortableLocaltime(static_cast<time_t>(ms / 1000), &time_struct))
        return "";
    // YYYY-MM-DDThh:mm:ss
    return StreamableToString(time_struct.tm_year + 1900) + "-" + String::FormatIntWidth2(time_struct.tm_mon + 1) + "-" + String::FormatIntWidth2(time_struct.tm_mday) + "T" + String::FormatIntWidth2(time_struct.tm_hour) + ":" + String::FormatIntWidth2(time_struct.tm_min) + ":" + String::FormatIntWidth2(time_struct.tm_sec) + "Z";
}

static inline std::string Indent(size_t width)
{
    return std::string(width, ' ');
}

void JsonUnitTestResultPrinter::OutputJsonKey(
    std::ostream *stream,
    const std::string &element_name,
    const std::string &name,
    const std::string &value,
    const std::string &indent,
    bool comma)
{
    const std::vector<std::string> &allowed_names =
        GetReservedOutputAttributesForElement(element_name);

    GTEST_CHECK_(std::find(allowed_names.begin(), allowed_names.end(), name) != allowed_names.end())
        << "Key \"" << name << "\" is not allowed for value \"" << element_name
        << "\".";

    *stream << indent << "\"" << name << "\": \"" << EscapeJson(value) << "\"";
    if (comma)
        *stream << ",\n";
}

void JsonUnitTestResultPrinter::OutputJsonKey(
    std::ostream *stream,
    const std::string &element_name,
    const std::string &name,
    int value,
    const std::string &indent,
    bool comma)
{
    const std::vector<std::string> &allowed_names =
        GetReservedOutputAttributesForElement(element_name);

    GTEST_CHECK_(std::find(allowed_names.begin(), allowed_names.end(), name) != allowed_names.end())
        << "Key \"" << name << "\" is not allowed for value \"" << element_name
        << "\".";

    *stream << indent << "\"" << name << "\": " << StreamableToString(value);
    if (comma)
        *stream << ",\n";
}

// Prints a JSON representation of a TestInfo object.
void JsonUnitTestResultPrinter::OutputJsonTestInfo(::std::ostream *stream,
                                                   const char *test_suite_name,
                                                   const TestInfo &test_info)
{
    const TestResult &result = *test_info.result();
    const std::string kTestsuite = "testcase";
    const std::string kIndent = Indent(10);

    *stream << Indent(8) << "{\n";
    OutputJsonKey(stream, kTestsuite, "name", test_info.name(), kIndent);

    if (test_info.value_param() != nullptr) {
        OutputJsonKey(stream, kTestsuite, "value_param", test_info.value_param(),
                      kIndent);
    }
    if (test_info.type_param() != nullptr) {
        OutputJsonKey(stream, kTestsuite, "type_param", test_info.type_param(),
                      kIndent);
    }
    if (GTEST_FLAG(list_tests)) {
        OutputJsonKey(stream, kTestsuite, "file", test_info.file(), kIndent);
        OutputJsonKey(stream, kTestsuite, "line", test_info.line(), kIndent, false);
        *stream << "\n"
                << Indent(8) << "}";
        return;
    }

    OutputJsonKey(stream, kTestsuite, "status",
                  test_info.should_run() ? "RUN" : "NOTRUN", kIndent);
    OutputJsonKey(stream, kTestsuite, "result",
                  test_info.should_run()
                      ? (result.Skipped() ? "SKIPPED" : "COMPLETED")
                      : "SUPPRESSED",
                  kIndent);
    OutputJsonKey(stream, kTestsuite, "timestamp",
                  FormatEpochTimeInMillisAsRFC3339(result.start_timestamp()),
                  kIndent);
    OutputJsonKey(stream, kTestsuite, "time",
                  FormatTimeInMillisAsDuration(result.elapsed_time()), kIndent);
    OutputJsonKey(stream, kTestsuite, "classname", test_suite_name, kIndent,
                  false);
    *stream << TestPropertiesAsJson(result, kIndent);

    int failures = 0;
    for (int i = 0; i < result.total_part_count(); ++i) {
        const TestPartResult &part = result.GetTestPartResult(i);
        if (part.failed()) {
            *stream << ",\n";
            if (++failures == 1) {
                *stream << kIndent << "\""
                        << "failures"
                        << "\": [\n";
            }
            const std::string location =
                internal::FormatCompilerIndependentFileLocation(part.file_name(),
                                                                part.line_number());
            const std::string message = EscapeJson(location + "\n" + part.message());
            *stream << kIndent << "  {\n"
                    << kIndent << "    \"failure\": \"" << message << "\",\n"
                    << kIndent << "    \"type\": \"\"\n"
                    << kIndent << "  }";
        }
    }

    if (failures > 0)
        *stream << "\n"
                << kIndent << "]";
    *stream << "\n"
            << Indent(8) << "}";
}

// Prints an JSON representation of a TestSuite object
void JsonUnitTestResultPrinter::PrintJsonTestSuite(
    std::ostream *stream, const TestSuite &test_suite)
{
    const std::string kTestsuite = "testsuite";
    const std::string kIndent = Indent(6);

    *stream << Indent(4) << "{\n";
    OutputJsonKey(stream, kTestsuite, "name", test_suite.name(), kIndent);
    OutputJsonKey(stream, kTestsuite, "tests", test_suite.reportable_test_count(),
                  kIndent);
    if (!GTEST_FLAG(list_tests)) {
        OutputJsonKey(stream, kTestsuite, "failures",
                      test_suite.failed_test_count(), kIndent);
        OutputJsonKey(stream, kTestsuite, "disabled",
                      test_suite.reportable_disabled_test_count(), kIndent);
        OutputJsonKey(stream, kTestsuite, "errors", 0, kIndent);
        OutputJsonKey(
            stream, kTestsuite, "timestamp",
            FormatEpochTimeInMillisAsRFC3339(test_suite.start_timestamp()),
            kIndent);
        OutputJsonKey(stream, kTestsuite, "time",
                      FormatTimeInMillisAsDuration(test_suite.elapsed_time()),
                      kIndent, false);
        *stream << TestPropertiesAsJson(test_suite.ad_hoc_test_result(), kIndent)
                << ",\n";
    }

    *stream << kIndent << "\"" << kTestsuite << "\": [\n";

    bool comma = false;
    for (int i = 0; i < test_suite.total_test_count(); ++i) {
        if (test_suite.GetTestInfo(i)->is_reportable()) {
            if (comma) {
                *stream << ",\n";
            } else {
                comma = true;
            }
            OutputJsonTestInfo(stream, test_suite.name(), *test_suite.GetTestInfo(i));
        }
    }
    *stream << "\n"
            << kIndent << "]\n"
            << Indent(4) << "}";
}

// Prints a JSON summary of unit_test to output stream out.
void JsonUnitTestResultPrinter::PrintJsonUnitTest(std::ostream *stream,
                                                  const UnitTest &unit_test)
{
    const std::string kTestsuites = "testsuites";
    const std::string kIndent = Indent(2);
    *stream << "{\n";

    OutputJsonKey(stream, kTestsuites, "tests", unit_test.reportable_test_count(),
                  kIndent);
    OutputJsonKey(stream, kTestsuites, "failures", unit_test.failed_test_count(),
                  kIndent);
    OutputJsonKey(stream, kTestsuites, "disabled",
                  unit_test.reportable_disabled_test_count(), kIndent);
    OutputJsonKey(stream, kTestsuites, "errors", 0, kIndent);
    if (GTEST_FLAG(shuffle)) {
        OutputJsonKey(stream, kTestsuites, "random_seed", unit_test.random_seed(),
                      kIndent);
    }
    OutputJsonKey(stream, kTestsuites, "timestamp",
                  FormatEpochTimeInMillisAsRFC3339(unit_test.start_timestamp()),
                  kIndent);
    OutputJsonKey(stream, kTestsuites, "time",
                  FormatTimeInMillisAsDuration(unit_test.elapsed_time()), kIndent,
                  false);

    *stream << TestPropertiesAsJson(unit_test.ad_hoc_test_result(), kIndent)
            << ",\n";

    OutputJsonKey(stream, kTestsuites, "name", "AllTests", kIndent);
    *stream << kIndent << "\"" << kTestsuites << "\": [\n";

    bool comma = false;
    for (int i = 0; i < unit_test.total_test_suite_count(); ++i) {
        if (unit_test.GetTestSuite(i)->reportable_test_count() > 0) {
            if (comma) {
                *stream << ",\n";
            } else {
                comma = true;
            }
            PrintJsonTestSuite(stream, *unit_test.GetTestSuite(i));
        }
    }

    *stream << "\n"
            << kIndent << "]\n"
            << "}\n";
}

void JsonUnitTestResultPrinter::PrintJsonTestList(
    std::ostream *stream, const std::vector<TestSuite *> &test_suites)
{
    const std::string kTestsuites = "testsuites";
    const std::string kIndent = Indent(2);
    *stream << "{\n";
    int total_tests = 0;
    for (auto test_suite : test_suites) {
        total_tests += test_suite->total_test_count();
    }
    OutputJsonKey(stream, kTestsuites, "tests", total_tests, kIndent);

    OutputJsonKey(stream, kTestsuites, "name", "AllTests", kIndent);
    *stream << kIndent << "\"" << kTestsuites << "\": [\n";

    for (size_t i = 0; i < test_suites.size(); ++i) {
        if (i != 0) {
            *stream << ",\n";
        }
        PrintJsonTestSuite(stream, *test_suites[i]);
    }

    *stream << "\n"
            << kIndent << "]\n"
            << "}\n";
}
// Produces a string representing the test properties in a result as
// a JSON dictionary.
std::string JsonUnitTestResultPrinter::TestPropertiesAsJson(
    const TestResult &result, const std::string &indent)
{
    Message attributes;
    for (int i = 0; i < result.test_property_count(); ++i) {
        const TestProperty &property = result.GetTestProperty(i);
        attributes << ",\n"
                   << indent << "\"" << property.key() << "\": "
                   << "\"" << EscapeJson(property.value()) << "\"";
    }
    return attributes.GetString();
}

// End JsonUnitTestResultPrinter

#if GTEST_CAN_STREAM_RESULTS_

// Checks if str contains '=', '&', '%' or '\n' characters. If yes,
// replaces them by "%xx" where xx is their hexadecimal value. For
// example, replaces "=" with "%3D".  This algorithm is O(strlen(str))
// in both time and space -- important as the input str may contain an
// arbitrarily long test failure message and stack trace.
std::string StreamingListener::UrlEncode(const char *str)
{
    std::string result;
    result.reserve(strlen(str) + 1);
    for (char ch = *str; ch != '\0'; ch = *++str) {
        switch (ch) {
        case '%':
        case '=':
        case '&':
        case '\n':
            result.append("%" + String::FormatByte(static_cast<unsigned char>(ch)));
            break;
        default:
            result.push_back(ch);
            break;
        }
    }
    return result;
}

void StreamingListener::SocketWriter::MakeConnection()
{
    GTEST_CHECK_(sockfd_ == -1)
        << "MakeConnection() can't be called when there is already a connection.";

    addrinfo hints;
    memset(&hints, 0, sizeof(hints));
    hints.ai_family = AF_UNSPEC; // To allow both IPv4 and IPv6 addresses.
    hints.ai_socktype = SOCK_STREAM;
    addrinfo *servinfo = nullptr;

    // Use the getaddrinfo() to get a linked list of IP addresses for
    // the given host name.
    const int error_num = getaddrinfo(
        host_name_.c_str(), port_num_.c_str(), &hints, &servinfo);
    if (error_num != 0) {
        GTEST_LOG_(WARNING) << "stream_result_to: getaddrinfo() failed: "
                            << gai_strerror(error_num);
    }

    // Loop through all the results and connect to the first we can.
    for (addrinfo *cur_addr = servinfo; sockfd_ == -1 && cur_addr != nullptr;
         cur_addr = cur_addr->ai_next) {
        sockfd_ = socket(
            cur_addr->ai_family, cur_addr->ai_socktype, cur_addr->ai_protocol);
        if (sockfd_ != -1) {
            // Connect the client socket to the server socket.
            if (connect(sockfd_, cur_addr->ai_addr, cur_addr->ai_addrlen) == -1) {
                close(sockfd_);
                sockfd_ = -1;
            }
        }
    }

    freeaddrinfo(servinfo); // all done with this structure

    if (sockfd_ == -1) {
        GTEST_LOG_(WARNING) << "stream_result_to: failed to connect to "
                            << host_name_ << ":" << port_num_;
    }
}

// End of class Streaming Listener
#endif // GTEST_CAN_STREAM_RESULTS__

// class OsStackTraceGetter

const char *const OsStackTraceGetterInterface::kElidedFramesMarker =
    "... " GTEST_NAME_ " internal frames ...";

std::string OsStackTraceGetter::CurrentStackTrace(int max_depth, int skip_count)
    GTEST_LOCK_EXCLUDED_(mutex_)
{
#if GTEST_HAS_ABSL
    std::string result;

    if (max_depth <= 0) {
        return result;
    }

    max_depth = std::min(max_depth, kMaxStackTraceDepth);

    std::vector<void *> raw_stack(max_depth);
    // Skips the frames requested by the caller, plus this function.
    const int raw_stack_size =
        absl::GetStackTrace(&raw_stack[0], max_depth, skip_count + 1);

    void *caller_frame = nullptr;
    {
        MutexLock lock(&mutex_);
        caller_frame = caller_frame_;
    }

    for (int i = 0; i < raw_stack_size; ++i) {
        if (raw_stack[i] == caller_frame && !GTEST_FLAG(show_internal_stack_frames)) {
            // Add a marker to the trace and stop adding frames.
            absl::StrAppend(&result, kElidedFramesMarker, "\n");
            break;
        }

        char tmp[1024];
        const char *symbol = "(unknown)";
        if (absl::Symbolize(raw_stack[i], tmp, sizeof(tmp))) {
            symbol = tmp;
        }

        char line[1024];
        snprintf(line, sizeof(line), "  %p: %s\n", raw_stack[i], symbol);
        result += line;
    }

    return result;

#else // !GTEST_HAS_ABSL
    static_cast<void>(max_depth);
    static_cast<void>(skip_count);
    return "";
#endif // GTEST_HAS_ABSL
}

void OsStackTraceGetter::UponLeavingGTest() GTEST_LOCK_EXCLUDED_(mutex_)
{
#if GTEST_HAS_ABSL
    void *caller_frame = nullptr;
    if (absl::GetStackTrace(&caller_frame, 1, 3) <= 0) {
        caller_frame = nullptr;
    }

    MutexLock lock(&mutex_);
    caller_frame_ = caller_frame;
#endif // GTEST_HAS_ABSL
}

// A helper class that creates the premature-exit file in its
// constructor and deletes the file in its destructor.
class ScopedPrematureExitFile
{
public:
    explicit ScopedPrematureExitFile(const char *premature_exit_filepath)
        : premature_exit_filepath_(premature_exit_filepath ? premature_exit_filepath : "")
    {
        // If a path to the premature-exit file is specified...
        if (!premature_exit_filepath_.empty()) {
            // create the file with a single "0" character in it.  I/O
            // errors are ignored as there's nothing better we can do and we
            // don't want to fail the test because of this.
            FILE *pfile = posix::FOpen(premature_exit_filepath, "w");
            fwrite("0", 1, 1, pfile);
            fclose(pfile);
        }
    }

    ~ScopedPrematureExitFile()
    {
#if !defined GTEST_OS_ESP8266
        if (!premature_exit_filepath_.empty()) {
            int retval = remove(premature_exit_filepath_.c_str());
            if (retval) {
                GTEST_LOG_(ERROR) << "Failed to remove premature exit filepath \""
                                  << premature_exit_filepath_ << "\" with error "
                                  << retval;
            }
        }
#endif
    }

private:
    const std::string premature_exit_filepath_;

    GTEST_DISALLOW_COPY_AND_ASSIGN_(ScopedPrematureExitFile);
};

} // namespace internal

// class TestEventListeners

TestEventListeners::TestEventListeners()
    : repeater_(new internal::TestEventRepeater())
    , default_result_printer_(nullptr)
    , default_xml_generator_(nullptr)
{
}

TestEventListeners::~TestEventListeners()
{
    delete repeater_;
}

// Returns the standard listener responsible for the default console
// output.  Can be removed from the listeners list to shut down default
// console output.  Note that removing this object from the listener list
// with Release transfers its ownership to the user.
void TestEventListeners::Append(TestEventListener *listener)
{
    repeater_->Append(listener);
}

// Removes the given event listener from the list and returns it.  It then
// becomes the caller's responsibility to delete the listener. Returns
// NULL if the listener is not found in the list.
TestEventListener *TestEventListeners::Release(TestEventListener *listener)
{
    if (listener == default_result_printer_)
        default_result_printer_ = nullptr;
    else if (listener == default_xml_generator_)
        default_xml_generator_ = nullptr;
    return repeater_->Release(listener);
}

// Returns repeater that broadcasts the TestEventListener events to all
// subscribers.
TestEventListener *TestEventListeners::repeater()
{
    return repeater_;
}

// Sets the default_result_printer attribute to the provided listener.
// The listener is also added to the listener list and previous
// default_result_printer is removed from it and deleted. The listener can
// also be NULL in which case it will not be added to the list. Does
// nothing if the previous and the current listener objects are the same.
void TestEventListeners::SetDefaultResultPrinter(TestEventListener *listener)
{
    if (default_result_printer_ != listener) {
        // It is an error to pass this method a listener that is already in the
        // list.
        delete Release(default_result_printer_);
        default_result_printer_ = listener;
        if (listener != nullptr)
            Append(listener);
    }
}

// Sets the default_xml_generator attribute to the provided listener.  The
// listener is also added to the listener list and previous
// default_xml_generator is removed from it and deleted. The listener can
// also be NULL in which case it will not be added to the list. Does
// nothing if the previous and the current listener objects are the same.
void TestEventListeners::SetDefaultXmlGenerator(TestEventListener *listener)
{
    if (default_xml_generator_ != listener) {
        // It is an error to pass this method a listener that is already in the
        // list.
        delete Release(default_xml_generator_);
        default_xml_generator_ = listener;
        if (listener != nullptr)
            Append(listener);
    }
}

// Controls whether events will be forwarded by the repeater to the
// listeners in the list.
bool TestEventListeners::EventForwardingEnabled() const
{
    return repeater_->forwarding_enabled();
}

void TestEventListeners::SuppressEventForwarding()
{
    repeater_->set_forwarding_enabled(false);
}

// class UnitTest

// Gets the singleton UnitTest object.  The first time this method is
// called, a UnitTest object is constructed and returned.  Consecutive
// calls will return the same object.
//
// We don't protect this under mutex_ as a user is not supposed to
// call this before main() starts, from which point on the return
// value will never change.
UnitTest *UnitTest::GetInstance()
{
    // CodeGear C++Builder insists on a public destructor for the
    // default implementation.  Use this implementation to keep good OO
    // design with private destructor.

#if defined(__BORLANDC__)
    static UnitTest *const instance = new UnitTest;
    return instance;
#else
    static UnitTest instance;
    return &instance;
#endif // defined(__BORLANDC__)
}

// Gets the number of successful test suites.
int UnitTest::successful_test_suite_count() const
{
    return impl()->successful_test_suite_count();
}

// Gets the number of failed test suites.
int UnitTest::failed_test_suite_count() const
{
    return impl()->failed_test_suite_count();
}

// Gets the number of all test suites.
int UnitTest::total_test_suite_count() const
{
    return impl()->total_test_suite_count();
}

// Gets the number of all test suites that contain at least one test
// that should run.
int UnitTest::test_suite_to_run_count() const
{
    return impl()->test_suite_to_run_count();
}

//  Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
int UnitTest::successful_test_case_count() const
{
    return impl()->successful_test_suite_count();
}
int UnitTest::failed_test_case_count() const
{
    return impl()->failed_test_suite_count();
}
int UnitTest::total_test_case_count() const
{
    return impl()->total_test_suite_count();
}
int UnitTest::test_case_to_run_count() const
{
    return impl()->test_suite_to_run_count();
}
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI_

// Gets the number of successful tests.
int UnitTest::successful_test_count() const
{
    return impl()->successful_test_count();
}

// Gets the number of skipped tests.
int UnitTest::skipped_test_count() const
{
    return impl()->skipped_test_count();
}

// Gets the number of failed tests.
int UnitTest::failed_test_count() const
{
    return impl()->failed_test_count();
}

// Gets the number of disabled tests that will be reported in the XML report.
int UnitTest::reportable_disabled_test_count() const
{
    return impl()->reportable_disabled_test_count();
}

// Gets the number of disabled tests.
int UnitTest::disabled_test_count() const
{
    return impl()->disabled_test_count();
}

// Gets the number of tests to be printed in the XML report.
int UnitTest::reportable_test_count() const
{
    return impl()->reportable_test_count();
}

// Gets the number of all tests.
int UnitTest::total_test_count() const
{
    return impl()->total_test_count();
}

// Gets the number of tests that should run.
int UnitTest::test_to_run_count() const
{
    return impl()->test_to_run_count();
}

// Gets the time of the test program start, in ms from the start of the
// UNIX epoch.
internal::TimeInMillis UnitTest::start_timestamp() const
{
    return impl()->start_timestamp();
}

// Gets the elapsed time, in milliseconds.
internal::TimeInMillis UnitTest::elapsed_time() const
{
    return impl()->elapsed_time();
}

// Returns true if and only if the unit test passed (i.e. all test suites
// passed).
bool UnitTest::Passed() const
{
    return impl()->Passed();
}

// Returns true if and only if the unit test failed (i.e. some test suite
// failed or something outside of all tests failed).
bool UnitTest::Failed() const
{
    return impl()->Failed();
}

// Gets the i-th test suite among all the test suites. i can range from 0 to
// total_test_suite_count() - 1. If i is not in that range, returns NULL.
const TestSuite *UnitTest::GetTestSuite(int i) const
{
    return impl()->GetTestSuite(i);
}

//  Legacy API is deprecated but still available
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
const TestCase *UnitTest::GetTestCase(int i) const
{
    return impl()->GetTestCase(i);
}
#endif //  GTEST_REMOVE_LEGACY_TEST_CASEAPI_

// Returns the TestResult containing information on test failures and
// properties logged outside of individual test suites.
const TestResult &UnitTest::ad_hoc_test_result() const
{
    return *impl()->ad_hoc_test_result();
}

// Gets the i-th test suite among all the test suites. i can range from 0 to
// total_test_suite_count() - 1. If i is not in that range, returns NULL.
TestSuite *UnitTest::GetMutableTestSuite(int i)
{
    return impl()->GetMutableSuiteCase(i);
}

// Returns the list of event listeners that can be used to track events
// inside Google Test.
TestEventListeners &UnitTest::listeners()
{
    return *impl()->listeners();
}

// Registers and returns a global test environment.  When a test
// program is run, all global test environments will be set-up in the
// order they were registered.  After all tests in the program have
// finished, all global test environments will be torn-down in the
// *reverse* order they were registered.
//
// The UnitTest object takes ownership of the given environment.
//
// We don't protect this under mutex_, as we only support calling it
// from the main thread.
Environment *UnitTest::AddEnvironment(Environment *env)
{
    if (env == nullptr) {
        return nullptr;
    }

    impl_->environments().push_back(env);
    return env;
}

// Adds a TestPartResult to the current TestResult object.  All Google Test
// assertion macros (e.g. ASSERT_TRUE, EXPECT_EQ, etc) eventually call
// this to report their results.  The user code should use the
// assertion macros instead of calling this directly.
void UnitTest::AddTestPartResult(
    TestPartResult::Type result_type,
    const char *file_name,
    int line_number,
    const std::string &message,
    const std::string &os_stack_trace) GTEST_LOCK_EXCLUDED_(mutex_)
{
    Message msg;
    msg << message;

    internal::MutexLock lock(&mutex_);
    if (impl_->gtest_trace_stack().size() > 0) {
        msg << "\n"
            << GTEST_NAME_ << " trace:";

        for (size_t i = impl_->gtest_trace_stack().size(); i > 0; --i) {
            const internal::TraceInfo &trace = impl_->gtest_trace_stack()[i - 1];
            msg << "\n"
                << internal::FormatFileLocation(trace.file, trace.line)
                << " " << trace.message;
        }
    }

    if (os_stack_trace.c_str() != nullptr && !os_stack_trace.empty()) {
        msg << internal::kStackTraceMarker << os_stack_trace;
    }

    const TestPartResult result = TestPartResult(
        result_type, file_name, line_number, msg.GetString().c_str());
    impl_->GetTestPartResultReporterForCurrentThread()->ReportTestPartResult(result);

    if (result_type != TestPartResult::kSuccess && result_type != TestPartResult::kSkip) {
        // gtest_break_on_failure takes precedence over
        // gtest_throw_on_failure.  This allows a user to set the latter
        // in the code (perhaps in order to use Google Test assertions
        // with another testing framework) and specify the former on the
        // command line for debugging.
        if (GTEST_FLAG(break_on_failure)) {
#if GTEST_OS_WINDOWS && !GTEST_OS_WINDOWS_PHONE && !GTEST_OS_WINDOWS_RT
            // Using DebugBreak on Windows allows gtest to still break into a debugger
            // when a failure happens and both the --gtest_break_on_failure and
            // the --gtest_catch_exceptions flags are specified.
            DebugBreak();
#elif (!defined(__native_client__)) && ((defined(__clang__) || defined(__GNUC__)) && (defined(__x86_64__) || defined(__i386__)))
            // with clang/gcc we can achieve the same effect on x86 by invoking int3
            asm("int3");
#else
            // Dereference nullptr through a volatile pointer to prevent the compiler
            // from removing. We use this rather than abort() or __builtin_trap() for
            // portability: some debuggers don't correctly trap abort().
            *static_cast<volatile int *>(nullptr) = 1;
#endif // GTEST_OS_WINDOWS
        } else if (GTEST_FLAG(throw_on_failure)) {
#if GTEST_HAS_EXCEPTIONS
            throw internal::GoogleTestFailureException(result);
#else
            // We cannot call abort() as it generates a pop-up in debug mode
            // that cannot be suppressed in VC 7.1 or below.
            exit(1);
#endif
        }
    }
}

// Adds a TestProperty to the current TestResult object when invoked from
// inside a test, to current TestSuite's ad_hoc_test_result_ when invoked
// from SetUpTestSuite or TearDownTestSuite, or to the global property set
// when invoked elsewhere.  If the result already contains a property with
// the same key, the value will be updated.
void UnitTest::RecordProperty(const std::string &key,
                              const std::string &value)
{
    impl_->RecordProperty(TestProperty(key, value));
}

// Runs all tests in this UnitTest object and prints the result.
// Returns 0 if successful, or 1 otherwise.
//
// We don't protect this under mutex_, as we only support calling it
// from the main thread.
int UnitTest::Run()
{
    const bool in_death_test_child_process =
        internal::GTEST_FLAG(internal_run_death_test).length() > 0;

    // Google Test implements this protocol for catching that a test
    // program exits before returning control to Google Test:
    //
    //   1. Upon start, Google Test creates a file whose absolute path
    //      is specified by the environment variable
    //      TEST_PREMATURE_EXIT_FILE.
    //   2. When Google Test has finished its work, it deletes the file.
    //
    // This allows a test runner to set TEST_PREMATURE_EXIT_FILE before
    // running a Google-Test-based test program and check the existence
    // of the file at the end of the test execution to see if it has
    // exited prematurely.

    // If we are in the child process of a death test, don't
    // create/delete the premature exit file, as doing so is unnecessary
    // and will confuse the parent process.  Otherwise, create/delete
    // the file upon entering/leaving this function.  If the program
    // somehow exits before this function has a chance to return, the
    // premature-exit file will be left undeleted, causing a test runner
    // that understands the premature-exit-file protocol to report the
    // test as having failed.
    const internal::ScopedPrematureExitFile premature_exit_file(
        in_death_test_child_process
            ? nullptr
            : internal::posix::GetEnv("TEST_PREMATURE_EXIT_FILE"));

    // Captures the value of GTEST_FLAG(catch_exceptions).  This value will be
    // used for the duration of the program.
    impl()->set_catch_exceptions(GTEST_FLAG(catch_exceptions));

#if GTEST_OS_WINDOWS
    // Either the user wants Google Test to catch exceptions thrown by the
    // tests or this is executing in the context of death test child
    // process. In either case the user does not want to see pop-up dialogs
    // about crashes - they are expected.
    if (impl()->catch_exceptions() || in_death_test_child_process) {
#if !GTEST_OS_WINDOWS_MOBILE && !GTEST_OS_WINDOWS_PHONE && !GTEST_OS_WINDOWS_RT
        // SetErrorMode doesn't exist on CE.
        SetErrorMode(SEM_FAILCRITICALERRORS | SEM_NOALIGNMENTFAULTEXCEPT | SEM_NOGPFAULTERRORBOX | SEM_NOOPENFILEERRORBOX);
#endif // !GTEST_OS_WINDOWS_MOBILE

#if (defined(_MSC_VER) || GTEST_OS_WINDOWS_MINGW) && !GTEST_OS_WINDOWS_MOBILE
        // Death test children can be terminated with _abort().  On Windows,
        // _abort() can show a dialog with a warning message.  This forces the
        // abort message to go to stderr instead.
        _set_error_mode(_OUT_TO_STDERR);
#endif

#if defined(_MSC_VER) && !GTEST_OS_WINDOWS_MOBILE
        // In the debug version, Visual Studio pops up a separate dialog
        // offering a choice to debug the aborted program. We need to suppress
        // this dialog or it will pop up for every EXPECT/ASSERT_DEATH statement
        // executed. Google Test will notify the user of any unexpected
        // failure via stderr.
        if (!GTEST_FLAG(break_on_failure))
            _set_abort_behavior(
                0x0, // Clear the following flags:
                _WRITE_ABORT_MSG | _CALL_REPORTFAULT); // pop-up window, core dump.

        // In debug mode, the Windows CRT can crash with an assertion over invalid
        // input (e.g. passing an invalid file descriptor).  The default handling
        // for these assertions is to pop up a dialog and wait for user input.
        // Instead ask the CRT to dump such assertions to stderr non-interactively.
        if (!IsDebuggerPresent()) {
            (void)_CrtSetReportMode(_CRT_ASSERT,
                                    _CRTDBG_MODE_FILE | _CRTDBG_MODE_DEBUG);
            (void)_CrtSetReportFile(_CRT_ASSERT, _CRTDBG_FILE_STDERR);
        }
#endif
    }
#endif // GTEST_OS_WINDOWS

    return internal::HandleExceptionsInMethodIfSupported(
               impl(),
               &internal::UnitTestImpl::RunAllTests,
               "auxiliary test code (environments or event listeners)")
               ? 0
               : 1;
}

// Returns the working directory when the first TEST() or TEST_F() was
// executed.
const char *UnitTest::original_working_dir() const
{
    return impl_->original_working_dir_.c_str();
}

// Returns the TestSuite object for the test that's currently running,
// or NULL if no test is running.
const TestSuite *UnitTest::current_test_suite() const
    GTEST_LOCK_EXCLUDED_(mutex_)
{
    internal::MutexLock lock(&mutex_);
    return impl_->current_test_suite();
}

// Legacy API is still available but deprecated
#ifndef GTEST_REMOVE_LEGACY_TEST_CASEAPI_
const TestCase *UnitTest::current_test_case() const
    GTEST_LOCK_EXCLUDED_(mutex_)
{
    internal::MutexLock lock(&mutex_);
    return impl_->current_test_suite();
}
#endif

// Returns the TestInfo object for the test that's currently running,
// or NULL if no test is running.
const TestInfo *UnitTest::current_test_info() const
    GTEST_LOCK_EXCLUDED_(mutex_)
{
    internal::MutexLock lock(&mutex_);
    return impl_->current_test_info();
}

// Returns the random seed used at the start of the current test run.
int UnitTest::random_seed() const
{
    return impl_->random_seed();
}

// Returns ParameterizedTestSuiteRegistry object used to keep track of
// value-parameterized tests and instantiate and register them.
internal::ParameterizedTestSuiteRegistry &
UnitTest::parameterized_test_registry() GTEST_LOCK_EXCLUDED_(mutex_)
{
    return impl_->parameterized_test_registry();
}

// Creates an empty UnitTest.
UnitTest::UnitTest()
{
    impl_ = new internal::UnitTestImpl(this);
}

// Destructor of UnitTest.
UnitTest::~UnitTest()
{
    delete impl_;
}

// Pushes a trace defined by SCOPED_TRACE() on to the per-thread
// Google Test trace stack.
void UnitTest::PushGTestTrace(const internal::TraceInfo &trace)
    GTEST_LOCK_EXCLUDED_(mutex_)
{
    internal::MutexLock lock(&mutex_);
    impl_->gtest_trace_stack().push_back(trace);
}

// Pops a trace from the per-thread Google Test trace stack.
void UnitTest::PopGTestTrace()
    GTEST_LOCK_EXCLUDED_(mutex_)
{
    internal::MutexLock lock(&mutex_);
    impl_->gtest_trace_stack().pop_back();
}

namespace internal {

UnitTestImpl::UnitTestImpl(UnitTest *parent)
    : parent_(parent)
    , GTEST_DISABLE_MSC_WARNINGS_PUSH_(4355 /* using this in initializer */)
          default_global_test_part_result_reporter_(this)
    , default_per_thread_test_part_result_reporter_(this)
    , GTEST_DISABLE_MSC_WARNINGS_POP_() global_test_part_result_repoter_(
          &default_global_test_part_result_reporter_)
    , per_thread_test_part_result_reporter_(
          &default_per_thread_test_part_result_reporter_)
    , parameterized_test_registry_()
    , parameterized_tests_registered_(false)
    , last_death_test_suite_(-1)
    , current_test_suite_(nullptr)
    , current_test_info_(nullptr)
    , ad_hoc_test_result_()
    , os_stack_trace_getter_(nullptr)
    , post_flag_parse_init_performed_(false)
    , random_seed_(0)
    , // Will be overridden by the flag before first use.
    random_(0)
    , // Will be reseeded before first use.
    start_timestamp_(0)
    , elapsed_time_(0)
    ,
#if GTEST_HAS_DEATH_TEST
    death_test_factory_(new DefaultDeathTestFactory)
    ,
#endif
    // Will be overridden by the flag before first use.
    catch_exceptions_(false)
{
    listeners()->SetDefaultResultPrinter(new PrettyUnitTestResultPrinter);
}

UnitTestImpl::~UnitTestImpl()
{
    // Deletes every TestSuite.
    ForEach(test_suites_, internal::Delete<TestSuite>);

    // Deletes every Environment.
    ForEach(environments_, internal::Delete<Environment>);

    delete os_stack_trace_getter_;
}

// Adds a TestProperty to the current TestResult object when invoked in a
// context of a test, to current test suite's ad_hoc_test_result when invoke
// from SetUpTestSuite/TearDownTestSuite, or to the global property set
// otherwise.  If the result already contains a property with the same key,
// the value will be updated.
void UnitTestImpl::RecordProperty(const TestProperty &test_property)
{
    std::string xml_element;
    TestResult *test_result; // TestResult appropriate for property recording.

    if (current_test_info_ != nullptr) {
        xml_element = "testcase";
        test_result = &(current_test_info_->result_);
    } else if (current_test_suite_ != nullptr) {
        xml_element = "testsuite";
        test_result = &(current_test_suite_->ad_hoc_test_result_);
    } else {
        xml_element = "testsuites";
        test_result = &ad_hoc_test_result_;
    }
    test_result->RecordProperty(xml_element, test_property);
}

#if GTEST_HAS_DEATH_TEST
// Disables event forwarding if the control is currently in a death test
// subprocess. Must not be called before InitGoogleTest.
void UnitTestImpl::SuppressTestEventsIfInSubprocess()
{
    if (internal_run_death_test_flag_.get() != nullptr)
        listeners()->SuppressEventForwarding();
}
#endif // GTEST_HAS_DEATH_TEST

// Initializes event listeners performing XML output as specified by
// UnitTestOptions. Must not be called before InitGoogleTest.
void UnitTestImpl::ConfigureXmlOutput()
{
    const std::string &output_format = UnitTestOptions::GetOutputFormat();
    if (output_format == "xml") {
        listeners()->SetDefaultXmlGenerator(new XmlUnitTestResultPrinter(
            UnitTestOptions::GetAbsolutePathToOutputFile().c_str()));
    } else if (output_format == "json") {
        listeners()->SetDefaultXmlGenerator(new JsonUnitTestResultPrinter(
            UnitTestOptions::GetAbsolutePathToOutputFile().c_str()));
    } else if (output_format != "") {
        GTEST_LOG_(WARNING) << "WARNING: unrecognized output format \""
                            << output_format << "\" ignored.";
    }
}

#if GTEST_CAN_STREAM_RESULTS_
// Initializes event listeners for streaming test results in string form.
// Must not be called before InitGoogleTest.
void UnitTestImpl::ConfigureStreamingOutput()
{
    const std::string &target = GTEST_FLAG(stream_result_to);
    if (!target.empty()) {
        const size_t pos = target.find(':');
        if (pos != std::string::npos) {
            listeners()->Append(new StreamingListener(target.substr(0, pos),
                                                      target.substr(pos + 1)));
        } else {
            GTEST_LOG_(WARNING) << "unrecognized streaming target \"" << target
                                << "\" ignored.";
        }
    }
}
#endif // GTEST_CAN_STREAM_RESULTS_

// Performs initialization dependent upon flag values obtained in
// ParseGoogleTestFlagsOnly.  Is called from InitGoogleTest after the call to
// ParseGoogleTestFlagsOnly.  In case a user neglects to call InitGoogleTest
// this function is also called from RunAllTests.  Since this function can be
// called more than once, it has to be idempotent.
void UnitTestImpl::PostFlagParsingInit()
{
    // Ensures that this function does not execute more than once.
    if (!post_flag_parse_init_performed_) {
        post_flag_parse_init_performed_ = true;

#if defined(GTEST_CUSTOM_TEST_EVENT_LISTENER_)
        // Register to send notifications about key process state changes.
        listeners()->Append(new GTEST_CUSTOM_TEST_EVENT_LISTENER_());
#endif // defined(GTEST_CUSTOM_TEST_EVENT_LISTENER_)

#if GTEST_HAS_DEATH_TEST
        InitDeathTestSubprocessControlInfo();
        SuppressTestEventsIfInSubprocess();
#endif // GTEST_HAS_DEATH_TEST

        // Registers parameterized tests. This makes parameterized tests
        // available to the UnitTest reflection API without running
        // RUN_ALL_TESTS.
        RegisterParameterizedTests();

        // Configures listeners for XML output. This makes it possible for users
        // to shut down the default XML output before invoking RUN_ALL_TESTS.
        ConfigureXmlOutput();

        if (GTEST_FLAG(brief)) {
            listeners()->SetDefaultResultPrinter(new BriefUnitTestResultPrinter);
        }

#if GTEST_CAN_STREAM_RESULTS_
        // Configures listeners for streaming test results to the specified server.
        ConfigureStreamingOutput();
#endif // GTEST_CAN_STREAM_RESULTS_

#if GTEST_HAS_ABSL
        if (GTEST_FLAG(install_failure_signal_handler)) {
            absl::FailureSignalHandlerOptions options;
            absl::InstallFailureSignalHandler(options);
        }
#endif // GTEST_HAS_ABSL
    }
}

// A predicate that checks the name of a TestSuite against a known
// value.
//
// This is used for implementation of the UnitTest class only.  We put
// it in the anonymous namespace to prevent polluting the outer
// namespace.
//
// TestSuiteNameIs is copyable.
class TestSuiteNameIs
{
public:
    // Constructor.
    explicit TestSuiteNameIs(const std::string &name)
        : name_(name)
    {
    }

    // Returns true if and only if the name of test_suite matches name_.
    bool operator()(const TestSuite *test_suite) const
    {
        return test_suite != nullptr && strcmp(test_suite->name(), name_.c_str()) == 0;
    }

private:
    std::string name_;
};

// Finds and returns a TestSuite with the given name.  If one doesn't
// exist, creates one and returns it.  It's the CALLER'S
// RESPONSIBILITY to ensure that this function is only called WHEN THE
// TESTS ARE NOT SHUFFLED.
//
// Arguments:
//
//   test_suite_name: name of the test suite
//   type_param:     the name of the test suite's type parameter, or NULL if
//                   this is not a typed or a type-parameterized test suite.
//   set_up_tc:      pointer to the function that sets up the test suite
//   tear_down_tc:   pointer to the function that tears down the test suite
TestSuite *UnitTestImpl::GetTestSuite(
    const char *test_suite_name, const char *type_param,
    internal::SetUpTestSuiteFunc set_up_tc,
    internal::TearDownTestSuiteFunc tear_down_tc)
{
    // Can we find a TestSuite with the given name?
    const auto test_suite =
        std::find_if(test_suites_.rbegin(), test_suites_.rend(),
                     TestSuiteNameIs(test_suite_name));

    if (test_suite != test_suites_.rend())
        return *test_suite;

    // No.  Let's create one.
    auto *const new_test_suite =
        new TestSuite(test_suite_name, type_param, set_up_tc, tear_down_tc);

    // Is this a death test suite?
    if (internal::UnitTestOptions::MatchesFilter(test_suite_name,
                                                 kDeathTestSuiteFilter)) {
        // Yes.  Inserts the test suite after the last death test suite
        // defined so far.  This only works when the test suites haven't
        // been shuffled.  Otherwise we may end up running a death test
        // after a non-death test.
        ++last_death_test_suite_;
        test_suites_.insert(test_suites_.begin() + last_death_test_suite_,
                            new_test_suite);
    } else {
        // No.  Appends to the end of the list.
        test_suites_.push_back(new_test_suite);
    }

    test_suite_indices_.push_back(static_cast<int>(test_suite_indices_.size()));
    return new_test_suite;
}

// Helpers for setting up / tearing down the given environment.  They
// are for use in the ForEach() function.
static void SetUpEnvironment(Environment *env)
{
    env->SetUp();
}
static void TearDownEnvironment(Environment *env)
{
    env->TearDown();
}

// Runs all tests in this UnitTest object, prints the result, and
// returns true if all tests are successful.  If any exception is
// thrown during a test, the test is considered to be failed, but the
// rest of the tests will still be run.
//
// When parameterized tests are enabled, it expands and registers
// parameterized tests first in RegisterParameterizedTests().
// All other functions called from RunAllTests() may safely assume that
// parameterized tests are ready to be counted and run.
bool UnitTestImpl::RunAllTests()
{
    // True if and only if Google Test is initialized before RUN_ALL_TESTS() is
    // called.
    const bool gtest_is_initialized_before_run_all_tests = GTestIsInitialized();

    // Do not run any test if the --help flag was specified.
    if (g_help_flag)
        return true;

    // Repeats the call to the post-flag parsing initialization in case the
    // user didn't call InitGoogleTest.
    PostFlagParsingInit();

    // Even if sharding is not on, test runners may want to use the
    // GTEST_SHARD_STATUS_FILE to query whether the test supports the sharding
    // protocol.
    internal::WriteToShardStatusFileIfNeeded();

    // True if and only if we are in a subprocess for running a thread-safe-style
    // death test.
    bool in_subprocess_for_death_test = false;

#if GTEST_HAS_DEATH_TEST
    in_subprocess_for_death_test =
        (internal_run_death_test_flag_.get() != nullptr);
#if defined(GTEST_EXTRA_DEATH_TEST_CHILD_SETUP_)
    if (in_subprocess_for_death_test) {
        GTEST_EXTRA_DEATH_TEST_CHILD_SETUP_();
    }
#endif // defined(GTEST_EXTRA_DEATH_TEST_CHILD_SETUP_)
#endif // GTEST_HAS_DEATH_TEST

    const bool should_shard = ShouldShard(kTestTotalShards, kTestShardIndex,
                                          in_subprocess_for_death_test);

    // Compares the full test names with the filter to decide which
    // tests to run.
    const bool has_tests_to_run = FilterTests(should_shard
                                                  ? HONOR_SHARDING_PROTOCOL
                                                  : IGNORE_SHARDING_PROTOCOL)
                                  > 0;

    // Lists the tests and exits if the --gtest_list_tests flag was specified.
    if (GTEST_FLAG(list_tests)) {
        // This must be called *after* FilterTests() has been called.
        ListTestsMatchingFilter();
        return true;
    }

    random_seed_ = GTEST_FLAG(shuffle) ? GetRandomSeedFromFlag(GTEST_FLAG(random_seed)) : 0;

    // True if and only if at least one test has failed.
    bool failed = false;

    TestEventListener *repeater = listeners()->repeater();

    start_timestamp_ = GetTimeInMillis();
    repeater->OnTestProgramStart(*parent_);

    // How many times to repeat the tests?  We don't want to repeat them
    // when we are inside the subprocess of a death test.
    const int repeat = in_subprocess_for_death_test ? 1 : GTEST_FLAG(repeat);
    // Repeats forever if the repeat count is negative.
    const bool gtest_repeat_forever = repeat < 0;
    for (int i = 0; gtest_repeat_forever || i != repeat; i++) {
        // We want to preserve failures generated by ad-hoc test
        // assertions executed before RUN_ALL_TESTS().
        ClearNonAdHocTestResult();

        const TimeInMillis start = GetTimeInMillis();

        // Shuffles test suites and tests if requested.
        if (has_tests_to_run && GTEST_FLAG(shuffle)) {
            random()->Reseed(static_cast<uint32_t>(random_seed_));
            // This should be done before calling OnTestIterationStart(),
            // such that a test event listener can see the actual test order
            // in the event.
            ShuffleTests();
        }

        // Tells the unit test event listeners that the tests are about to start.
        repeater->OnTestIterationStart(*parent_, i);

        // Runs each test suite if there is at least one test to run.
        if (has_tests_to_run) {
            // Sets up all environments beforehand.
            repeater->OnEnvironmentsSetUpStart(*parent_);
            ForEach(environments_, SetUpEnvironment);
            repeater->OnEnvironmentsSetUpEnd(*parent_);

            // Runs the tests only if there was no fatal failure or skip triggered
            // during global set-up.
            if (Test::IsSkipped()) {
                // Emit diagnostics when global set-up calls skip, as it will not be
                // emitted by default.
                TestResult &test_result =
                    *internal::GetUnitTestImpl()->current_test_result();
                for (int j = 0; j < test_result.total_part_count(); ++j) {
                    const TestPartResult &test_part_result =
                        test_result.GetTestPartResult(j);
                    if (test_part_result.type() == TestPartResult::kSkip) {
                        const std::string &result = test_part_result.message();
                        printf("%s\n", result.c_str());
                    }
                }
                fflush(stdout);
            } else if (!Test::HasFatalFailure()) {
                for (int test_index = 0; test_index < total_test_suite_count();
                     test_index++) {
                    GetMutableSuiteCase(test_index)->Run();
                    if (GTEST_FLAG(fail_fast) && GetMutableSuiteCase(test_index)->Failed()) {
                        for (int j = test_index + 1; j < total_test_suite_count(); j++) {
                            GetMutableSuiteCase(j)->Skip();
                        }
                        break;
                    }
                }
            }

            // Tears down all environments in reverse order afterwards.
            repeater->OnEnvironmentsTearDownStart(*parent_);
            std::for_each(environments_.rbegin(), environments_.rend(),
                          TearDownEnvironment);
            repeater->OnEnvironmentsTearDownEnd(*parent_);
        }

        elapsed_time_ = GetTimeInMillis() - start;

        // Tells the unit test event listener that the tests have just finished.
        repeater->OnTestIterationEnd(*parent_, i);

        // Gets the result and clears it.
        if (!Passed()) {
            failed = true;
        }

        // Restores the original test order after the iteration.  This
        // allows the user to quickly repro a failure that happens in the
        // N-th iteration without repeating the first (N - 1) iterations.
        // This is not enclosed in "if (GTEST_FLAG(shuffle)) { ... }", in
        // case the user somehow changes the value of the flag somewhere
        // (it's always safe to unshuffle the tests).
        UnshuffleTests();

        if (GTEST_FLAG(shuffle)) {
            // Picks a new random seed for each iteration.
            random_seed_ = GetNextRandomSeed(random_seed_);
        }
    }

    repeater->OnTestProgramEnd(*parent_);

    if (!gtest_is_initialized_before_run_all_tests) {
        ColoredPrintf(
            GTestColor::kRed,
            "\nIMPORTANT NOTICE - DO NOT IGNORE:\n"
            "This test program did NOT call " GTEST_INIT_GOOGLE_TEST_NAME_
            "() before calling RUN_ALL_TESTS(). This is INVALID. Soon " GTEST_NAME_
            " will start to enforce the valid usage. "
            "Please fix it ASAP, or IT WILL START TO FAIL.\n"); // NOLINT
#if GTEST_FOR_GOOGLE_
        ColoredPrintf(GTestColor::kRed,
                      "For more details, see http://wiki/Main/ValidGUnitMain.\n");
#endif // GTEST_FOR_GOOGLE_
    }

    return !failed;
}

// Reads the GTEST_SHARD_STATUS_FILE environment variable, and creates the file
// if the variable is present. If a file already exists at this location, this
// function will write over it. If the variable is present, but the file cannot
// be created, prints an error and exits.
void WriteToShardStatusFileIfNeeded()
{
    const char *const test_shard_file = posix::GetEnv(kTestShardStatusFile);
    if (test_shard_file != nullptr) {
        FILE *const file = posix::FOpen(test_shard_file, "w");
        if (file == nullptr) {
            ColoredPrintf(GTestColor::kRed,
                          "Could not write to the test shard status file \"%s\" "
                          "specified by the %s environment variable.\n",
                          test_shard_file, kTestShardStatusFile);
            fflush(stdout);
            exit(EXIT_FAILURE);
        }
        fclose(file);
    }
}

// Checks whether sharding is enabled by examining the relevant
// environment variable values. If the variables are present,
// but inconsistent (i.e., shard_index >= total_shards), prints
// an error and exits. If in_subprocess_for_death_test, sharding is
// disabled because it must only be applied to the original test
// process. Otherwise, we could filter out death tests we intended to execute.
bool ShouldShard(const char *total_shards_env,
                 const char *shard_index_env,
                 bool in_subprocess_for_death_test)
{
    if (in_subprocess_for_death_test) {
        return false;
    }

    const int32_t total_shards = Int32FromEnvOrDie(total_shards_env, -1);
    const int32_t shard_index = Int32FromEnvOrDie(shard_index_env, -1);

    if (total_shards == -1 && shard_index == -1) {
        return false;
    } else if (total_shards == -1 && shard_index != -1) {
        const Message msg = Message()
                            << "Invalid environment variables: you have "
                            << kTestShardIndex << " = " << shard_index
                            << ", but have left " << kTestTotalShards << " unset.\n";
        ColoredPrintf(GTestColor::kRed, "%s", msg.GetString().c_str());
        fflush(stdout);
        exit(EXIT_FAILURE);
    } else if (total_shards != -1 && shard_index == -1) {
        const Message msg = Message()
                            << "Invalid environment variables: you have "
                            << kTestTotalShards << " = " << total_shards
                            << ", but have left " << kTestShardIndex << " unset.\n";
        ColoredPrintf(GTestColor::kRed, "%s", msg.GetString().c_str());
        fflush(stdout);
        exit(EXIT_FAILURE);
    } else if (shard_index < 0 || shard_index >= total_shards) {
        const Message msg = Message()
                            << "Invalid environment variables: we require 0 <= "
                            << kTestShardIndex << " < " << kTestTotalShards
                            << ", but you have " << kTestShardIndex << "=" << shard_index
                            << ", " << kTestTotalShards << "=" << total_shards << ".\n";
        ColoredPrintf(GTestColor::kRed, "%s", msg.GetString().c_str());
        fflush(stdout);
        exit(EXIT_FAILURE);
    }

    return total_shards > 1;
}

// Parses the environment variable var as an Int32. If it is unset,
// returns default_val. If it is not an Int32, prints an error
// and aborts.
int32_t Int32FromEnvOrDie(const char *var, int32_t default_val)
{
    const char *str_val = posix::GetEnv(var);
    if (str_val == nullptr) {
        return default_val;
    }

    int32_t result;
    if (!ParseInt32(Message() << "The value of environment variable " << var,
                    str_val, &result)) {
        exit(EXIT_FAILURE);
    }
    return result;
}

// Given the total number of shards, the shard index, and the test id,
// returns true if and only if the test should be run on this shard. The test id
// is some arbitrary but unique non-negative integer assigned to each test
// method. Assumes that 0 <= shard_index < total_shards.
bool ShouldRunTestOnShard(int total_shards, int shard_index, int test_id)
{
    return (test_id % total_shards) == shard_index;
}

// Compares the name of each test with the user-specified filter to
// decide whether the test should be run, then records the result in
// each TestSuite and TestInfo object.
// If shard_tests == true, further filters tests based on sharding
// variables in the environment - see
// https://github.com/google/googletest/blob/master/googletest/docs/advanced.md
// . Returns the number of tests that should run.
int UnitTestImpl::FilterTests(ReactionToSharding shard_tests)
{
    const int32_t total_shards = shard_tests == HONOR_SHARDING_PROTOCOL ? Int32FromEnvOrDie(kTestTotalShards, -1) : -1;
    const int32_t shard_index = shard_tests == HONOR_SHARDING_PROTOCOL ? Int32FromEnvOrDie(kTestShardIndex, -1) : -1;

    // num_runnable_tests are the number of tests that will
    // run across all shards (i.e., match filter and are not disabled).
    // num_selected_tests are the number of tests to be run on
    // this shard.
    int num_runnable_tests = 0;
    int num_selected_tests = 0;
    for (auto *test_suite : test_suites_) {
        const std::string &test_suite_name = test_suite->name();
        test_suite->set_should_run(false);

        for (size_t j = 0; j < test_suite->test_info_list().size(); j++) {
            TestInfo *const test_info = test_suite->test_info_list()[j];
            const std::string test_name(test_info->name());
            // A test is disabled if test suite name or test name matches
            // kDisableTestFilter.
            const bool is_disabled = internal::UnitTestOptions::MatchesFilter(
                                         test_suite_name, kDisableTestFilter)
                                     || internal::UnitTestOptions::MatchesFilter(
                                            test_name, kDisableTestFilter);
            test_info->is_disabled_ = is_disabled;

            const bool matches_filter = internal::UnitTestOptions::FilterMatchesTest(
                test_suite_name, test_name);
            test_info->matches_filter_ = matches_filter;

            const bool is_runnable =
                (GTEST_FLAG(also_run_disabled_tests) || !is_disabled) && matches_filter;

            const bool is_in_another_shard =
                shard_tests != IGNORE_SHARDING_PROTOCOL && !ShouldRunTestOnShard(total_shards, shard_index, num_runnable_tests);
            test_info->is_in_another_shard_ = is_in_another_shard;
            const bool is_selected = is_runnable && !is_in_another_shard;

            num_runnable_tests += is_runnable;
            num_selected_tests += is_selected;

            test_info->should_run_ = is_selected;
            test_suite->set_should_run(test_suite->should_run() || is_selected);
        }
    }
    return num_selected_tests;
}

// Prints the given C-string on a single line by replacing all '\n'
// characters with string "\\n".  If the output takes more than
// max_length characters, only prints the first max_length characters
// and "...".
static void PrintOnOneLine(const char *str, int max_length)
{
    if (str != nullptr) {
        for (int i = 0; *str != '\0'; ++str) {
            if (i >= max_length) {
                printf("...");
                break;
            }
            if (*str == '\n') {
                printf("\\n");
                i += 2;
            } else {
                printf("%c", *str);
                ++i;
            }
        }
    }
}

// Prints the names of the tests matching the user-specified filter flag.
void UnitTestImpl::ListTestsMatchingFilter()
{
    // Print at most this many characters for each type/value parameter.
    const int kMaxParamLength = 250;

    for (auto *test_suite : test_suites_) {
        bool printed_test_suite_name = false;

        for (size_t j = 0; j < test_suite->test_info_list().size(); j++) {
            const TestInfo *const test_info = test_suite->test_info_list()[j];
            if (test_info->matches_filter_) {
                if (!printed_test_suite_name) {
                    printed_test_suite_name = true;
                    printf("%s.", test_suite->name());
                    if (test_suite->type_param() != nullptr) {
                        printf("  # %s = ", kTypeParamLabel);
                        // We print the type parameter on a single line to make
                        // the output easy to parse by a program.
                        PrintOnOneLine(test_suite->type_param(), kMaxParamLength);
                    }
                    printf("\n");
                }
                printf("  %s", test_info->name());
                if (test_info->value_param() != nullptr) {
                    printf("  # %s = ", kValueParamLabel);
                    // We print the value parameter on a single line to make the
                    // output easy to parse by a program.
                    PrintOnOneLine(test_info->value_param(), kMaxParamLength);
                }
                printf("\n");
            }
        }
    }
    fflush(stdout);
    const std::string &output_format = UnitTestOptions::GetOutputFormat();
    if (output_format == "xml" || output_format == "json") {
        FILE *fileout = OpenFileForWriting(
            UnitTestOptions::GetAbsolutePathToOutputFile().c_str());
        std::stringstream stream;
        if (output_format == "xml") {
            XmlUnitTestResultPrinter(
                UnitTestOptions::GetAbsolutePathToOutputFile().c_str())
                .PrintXmlTestsList(&stream, test_suites_);
        } else if (output_format == "json") {
            JsonUnitTestResultPrinter(
                UnitTestOptions::GetAbsolutePathToOutputFile().c_str())
                .PrintJsonTestList(&stream, test_suites_);
        }
        fprintf(fileout, "%s", StringStreamToString(&stream).c_str());
        fclose(fileout);
    }
}

// Sets the OS stack trace getter.
//
// Does nothing if the input and the current OS stack trace getter are
// the same; otherwise, deletes the old getter and makes the input the
// current getter.
void UnitTestImpl::set_os_stack_trace_getter(
    OsStackTraceGetterInterface *getter)
{
    if (os_stack_trace_getter_ != getter) {
        delete os_stack_trace_getter_;
        os_stack_trace_getter_ = getter;
    }
}

// Returns the current OS stack trace getter if it is not NULL;
// otherwise, creates an OsStackTraceGetter, makes it the current
// getter, and returns it.
OsStackTraceGetterInterface *UnitTestImpl::os_stack_trace_getter()
{
    if (os_stack_trace_getter_ == nullptr) {
#ifdef GTEST_OS_STACK_TRACE_GETTER_
        os_stack_trace_getter_ = new GTEST_OS_STACK_TRACE_GETTER_;
#else
        os_stack_trace_getter_ = new OsStackTraceGetter;
#endif // GTEST_OS_STACK_TRACE_GETTER_
    }

    return os_stack_trace_getter_;
}

// Returns the most specific TestResult currently running.
TestResult *UnitTestImpl::current_test_result()
{
    if (current_test_info_ != nullptr) {
        return &current_test_info_->result_;
    }
    if (current_test_suite_ != nullptr) {
        return &current_test_suite_->ad_hoc_test_result_;
    }
    return &ad_hoc_test_result_;
}

// Shuffles all test suites, and the tests within each test suite,
// making sure that death tests are still run first.
void UnitTestImpl::ShuffleTests()
{
    // Shuffles the death test suites.
    ShuffleRange(random(), 0, last_death_test_suite_ + 1, &test_suite_indices_);

    // Shuffles the non-death test suites.
    ShuffleRange(random(), last_death_test_suite_ + 1,
                 static_cast<int>(test_suites_.size()), &test_suite_indices_);

    // Shuffles the tests inside each test suite.
    for (auto &test_suite : test_suites_) {
        test_suite->ShuffleTests(random());
    }
}

// Restores the test suites and tests to their order before the first shuffle.
void UnitTestImpl::UnshuffleTests()
{
    for (size_t i = 0; i < test_suites_.size(); i++) {
        // Unshuffles the tests in each test suite.
        test_suites_[i]->UnshuffleTests();
        // Resets the index of each test suite.
        test_suite_indices_[i] = static_cast<int>(i);
    }
}

// Returns the current OS stack trace as an std::string.
//
// The maximum number of stack frames to be included is specified by
// the gtest_stack_trace_depth flag.  The skip_count parameter
// specifies the number of top frames to be skipped, which doesn't
// count against the number of frames to be included.
//
// For example, if Foo() calls Bar(), which in turn calls
// GetCurrentOsStackTraceExceptTop(..., 1), Foo() will be included in
// the trace but Bar() and GetCurrentOsStackTraceExceptTop() won't.
std::string GetCurrentOsStackTraceExceptTop(UnitTest * /*unit_test*/,
                                            int skip_count)
{
    // We pass skip_count + 1 to skip this wrapper function in addition
    // to what the user really wants to skip.
    return GetUnitTestImpl()->CurrentOsStackTraceExceptTop(skip_count + 1);
}

// Used by the GTEST_SUPPRESS_UNREACHABLE_CODE_WARNING_BELOW_ macro to
// suppress unreachable code warnings.
namespace {
class ClassUniqueToAlwaysTrue
{
};
} // namespace

bool IsTrue(bool condition)
{
    return condition;
}

bool AlwaysTrue()
{
#if GTEST_HAS_EXCEPTIONS
    // This condition is always false so AlwaysTrue() never actually throws,
    // but it makes the compiler think that it may throw.
    if (IsTrue(false))
        throw ClassUniqueToAlwaysTrue();
#endif // GTEST_HAS_EXCEPTIONS
    return true;
}

// If *pstr starts with the given prefix, modifies *pstr to be right
// past the prefix and returns true; otherwise leaves *pstr unchanged
// and returns false.  None of pstr, *pstr, and prefix can be NULL.
bool SkipPrefix(const char *prefix, const char **pstr)
{
    const size_t prefix_len = strlen(prefix);
    if (strncmp(*pstr, prefix, prefix_len) == 0) {
        *pstr += prefix_len;
        return true;
    }
    return false;
}

// Parses a string as a command line flag.  The string should have
// the format "--flag=value".  When def_optional is true, the "=value"
// part can be omitted.
//
// Returns the value of the flag, or NULL if the parsing failed.
static const char *ParseFlagValue(const char *str, const char *flag,
                                  bool def_optional)
{
    // str and flag must not be NULL.
    if (str == nullptr || flag == nullptr)
        return nullptr;

    // The flag must start with "--" followed by GTEST_FLAG_PREFIX_.
    const std::string flag_str = std::string("--") + GTEST_FLAG_PREFIX_ + flag;
    const size_t flag_len = flag_str.length();
    if (strncmp(str, flag_str.c_str(), flag_len) != 0)
        return nullptr;

    // Skips the flag name.
    const char *flag_end = str + flag_len;

    // When def_optional is true, it's OK to not have a "=value" part.
    if (def_optional && (flag_end[0] == '\0')) {
        return flag_end;
    }

    // If def_optional is true and there are more characters after the
    // flag name, or if def_optional is false, there must be a '=' after
    // the flag name.
    if (flag_end[0] != '=')
        return nullptr;

    // Returns the string after "=".
    return flag_end + 1;
}

// Parses a string for a bool flag, in the form of either
// "--flag=value" or "--flag".
//
// In the former case, the value is taken as true as long as it does
// not start with '0', 'f', or 'F'.
//
// In the latter case, the value is taken as true.
//
// On success, stores the value of the flag in *value, and returns
// true.  On failure, returns false without changing *value.
static bool ParseBoolFlag(const char *str, const char *flag, bool *value)
{
    // Gets the value of the flag as a string.
    const char *const value_str = ParseFlagValue(str, flag, true);

    // Aborts if the parsing failed.
    if (value_str == nullptr)
        return false;

    // Converts the string value to a bool.
    *value = !(*value_str == '0' || *value_str == 'f' || *value_str == 'F');
    return true;
}

// Parses a string for an int32_t flag, in the form of "--flag=value".
//
// On success, stores the value of the flag in *value, and returns
// true.  On failure, returns false without changing *value.
bool ParseInt32Flag(const char *str, const char *flag, int32_t *value)
{
    // Gets the value of the flag as a string.
    const char *const value_str = ParseFlagValue(str, flag, false);

    // Aborts if the parsing failed.
    if (value_str == nullptr)
        return false;

    // Sets *value to the value of the flag.
    return ParseInt32(Message() << "The value of flag --" << flag,
                      value_str, value);
}

// Parses a string for a string flag, in the form of "--flag=value".
//
// On success, stores the value of the flag in *value, and returns
// true.  On failure, returns false without changing *value.
template<typename String>
static bool ParseStringFlag(const char *str, const char *flag, String *value)
{
    // Gets the value of the flag as a string.
    const char *const value_str = ParseFlagValue(str, flag, false);

    // Aborts if the parsing failed.
    if (value_str == nullptr)
        return false;

    // Sets *value to the value of the flag.
    *value = value_str;
    return true;
}

// Determines whether a string has a prefix that Google Test uses for its
// flags, i.e., starts with GTEST_FLAG_PREFIX_ or GTEST_FLAG_PREFIX_DASH_.
// If Google Test detects that a command line flag has its prefix but is not
// recognized, it will print its help message. Flags starting with
// GTEST_INTERNAL_PREFIX_ followed by "internal_" are considered Google Test
// internal flags and do not trigger the help message.
static bool HasGoogleTestFlagPrefix(const char *str)
{
    return (SkipPrefix("--", &str) || SkipPrefix("-", &str) || SkipPrefix("/", &str)) && !SkipPrefix(GTEST_FLAG_PREFIX_ "internal_", &str) && (SkipPrefix(GTEST_FLAG_PREFIX_, &str) || SkipPrefix(GTEST_FLAG_PREFIX_DASH_, &str));
}

// Prints a string containing code-encoded text.  The following escape
// sequences can be used in the string to control the text color:
//
//   @@    prints a single '@' character.
//   @R    changes the color to red.
//   @G    changes the color to green.
//   @Y    changes the color to yellow.
//   @D    changes to the default terminal text color.
//
static void PrintColorEncoded(const char *str)
{
    GTestColor color = GTestColor::kDefault; // The current color.

    // Conceptually, we split the string into segments divided by escape
    // sequences.  Then we print one segment at a time.  At the end of
    // each iteration, the str pointer advances to the beginning of the
    // next segment.
    for (;;) {
        const char *p = strchr(str, '@');
        if (p == nullptr) {
            ColoredPrintf(color, "%s", str);
            return;
        }

        ColoredPrintf(color, "%s", std::string(str, p).c_str());

        const char ch = p[1];
        str = p + 2;
        if (ch == '@') {
            ColoredPrintf(color, "@");
        } else if (ch == 'D') {
            color = GTestColor::kDefault;
        } else if (ch == 'R') {
            color = GTestColor::kRed;
        } else if (ch == 'G') {
            color = GTestColor::kGreen;
        } else if (ch == 'Y') {
            color = GTestColor::kYellow;
        } else {
            --str;
        }
    }
}

static const char kColorEncodedHelpMessage[] =
    "This program contains tests written using " GTEST_NAME_
    ". You can use the\n"
    "following command line flags to control its behavior:\n"
    "\n"
    "Test Selection:\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "list_tests@D\n"
    "      List the names of all tests instead of running them. The name of\n"
    "      TEST(Foo, Bar) is \"Foo.Bar\".\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "filter=@YPOSTIVE_PATTERNS"
    "[@G-@YNEGATIVE_PATTERNS]@D\n"
    "      Run only the tests whose name matches one of the positive patterns "
    "but\n"
    "      none of the negative patterns. '?' matches any single character; "
    "'*'\n"
    "      matches any substring; ':' separates two patterns.\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "also_run_disabled_tests@D\n"
    "      Run all disabled tests too.\n"
    "\n"
    "Test Execution:\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "repeat=@Y[COUNT]@D\n"
    "      Run the tests repeatedly; use a negative count to repeat forever.\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "shuffle@D\n"
    "      Randomize tests' orders on every iteration.\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "random_seed=@Y[NUMBER]@D\n"
    "      Random number seed to use for shuffling test orders (between 1 and\n"
    "      99999, or 0 to use a seed based on the current time).\n"
    "\n"
    "Test Output:\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "color=@Y(@Gyes@Y|@Gno@Y|@Gauto@Y)@D\n"
    "      Enable/disable colored output. The default is @Gauto@D.\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "brief=1@D\n"
    "      Only print test failures.\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "print_time=0@D\n"
    "      Don't print the elapsed time of each test.\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "output=@Y(@Gjson@Y|@Gxml@Y)[@G:@YDIRECTORY_PATH@G" GTEST_PATH_SEP_
    "@Y|@G:@YFILE_PATH]@D\n"
    "      Generate a JSON or XML report in the given directory or with the "
    "given\n"
    "      file name. @YFILE_PATH@D defaults to @Gtest_detail.xml@D.\n"
#if GTEST_CAN_STREAM_RESULTS_
    "  @G--" GTEST_FLAG_PREFIX_
    "stream_result_to=@YHOST@G:@YPORT@D\n"
    "      Stream test results to the given server.\n"
#endif // GTEST_CAN_STREAM_RESULTS_
    "\n"
    "Assertion Behavior:\n"
#if GTEST_HAS_DEATH_TEST && !GTEST_OS_WINDOWS
    "  @G--" GTEST_FLAG_PREFIX_
    "death_test_style=@Y(@Gfast@Y|@Gthreadsafe@Y)@D\n"
    "      Set the default death test style.\n"
#endif // GTEST_HAS_DEATH_TEST && !GTEST_OS_WINDOWS
    "  @G--" GTEST_FLAG_PREFIX_
    "break_on_failure@D\n"
    "      Turn assertion failures into debugger break-points.\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "throw_on_failure@D\n"
    "      Turn assertion failures into C++ exceptions for use by an external\n"
    "      test framework.\n"
    "  @G--" GTEST_FLAG_PREFIX_
    "catch_exceptions=0@D\n"
    "      Do not report exceptions as test failures. Instead, allow them\n"
    "      to crash the program or throw a pop-up (on Windows).\n"
    "\n"
    "Except for @G--" GTEST_FLAG_PREFIX_
    "list_tests@D, you can alternatively set "
    "the corresponding\n"
    "environment variable of a flag (all letters in upper-case). For example, "
    "to\n"
    "disable colored text output, you can either specify "
    "@G--" GTEST_FLAG_PREFIX_
    "color=no@D or set\n"
    "the @G" GTEST_FLAG_PREFIX_UPPER_
    "COLOR@D environment variable to @Gno@D.\n"
    "\n"
    "For more information, please read the " GTEST_NAME_
    " documentation at\n"
    "@G" GTEST_PROJECT_URL_ "@D. If you find a bug in " GTEST_NAME_
    "\n"
    "(not one in your own code or tests), please report it to\n"
    "@G<" GTEST_DEV_EMAIL_ ">@D.\n";

static bool ParseGoogleTestFlag(const char *const arg)
{
    return ParseBoolFlag(arg, kAlsoRunDisabledTestsFlag,
                         &GTEST_FLAG(also_run_disabled_tests))
           || ParseBoolFlag(arg, kBreakOnFailureFlag,
                            &GTEST_FLAG(break_on_failure))
           || ParseBoolFlag(arg, kCatchExceptionsFlag,
                            &GTEST_FLAG(catch_exceptions))
           || ParseStringFlag(arg, kColorFlag, &GTEST_FLAG(color)) || ParseStringFlag(arg, kDeathTestStyleFlag, &GTEST_FLAG(death_test_style)) || ParseBoolFlag(arg, kDeathTestUseFork, &GTEST_FLAG(death_test_use_fork)) || ParseBoolFlag(arg, kFailFast, &GTEST_FLAG(fail_fast)) || ParseStringFlag(arg, kFilterFlag, &GTEST_FLAG(filter)) || ParseStringFlag(arg, kInternalRunDeathTestFlag, &GTEST_FLAG(internal_run_death_test)) || ParseBoolFlag(arg, kListTestsFlag, &GTEST_FLAG(list_tests)) || ParseStringFlag(arg, kOutputFlag, &GTEST_FLAG(output)) || ParseBoolFlag(arg, kBriefFlag, &GTEST_FLAG(brief)) || ParseBoolFlag(arg, kPrintTimeFlag, &GTEST_FLAG(print_time)) || ParseBoolFlag(arg, kPrintUTF8Flag, &GTEST_FLAG(print_utf8)) || ParseInt32Flag(arg, kRandomSeedFlag, &GTEST_FLAG(random_seed)) || ParseInt32Flag(arg, kRepeatFlag, &GTEST_FLAG(repeat)) || ParseBoolFlag(arg, kShuffleFlag, &GTEST_FLAG(shuffle)) || ParseInt32Flag(arg, kStackTraceDepthFlag, &GTEST_FLAG(stack_trace_depth)) || ParseStringFlag(arg, kStreamResultToFlag, &GTEST_FLAG(stream_result_to)) || ParseBoolFlag(arg, kThrowOnFailureFlag, &GTEST_FLAG(throw_on_failure));
}

#if GTEST_USE_OWN_FLAGFILE_FLAG_
static void LoadFlagsFromFile(const std::string &path)
{
    FILE *flagfile = posix::FOpen(path.c_str(), "r");
    if (!flagfile) {
        GTEST_LOG_(FATAL) << "Unable to open file \"" << GTEST_FLAG(flagfile)
                          << "\"";
    }
    std::string contents(ReadEntireFile(flagfile));
    posix::FClose(flagfile);
    std::vector<std::string> lines;
    SplitString(contents, '\n', &lines);
    for (size_t i = 0; i < lines.size(); ++i) {
        if (lines[i].empty())
            continue;
        if (!ParseGoogleTestFlag(lines[i].c_str()))
            g_help_flag = true;
    }
}
#endif // GTEST_USE_OWN_FLAGFILE_FLAG_

// Parses the command line for Google Test flags, without initializing
// other parts of Google Test.  The type parameter CharType can be
// instantiated to either char or wchar_t.
template<typename CharType>
void ParseGoogleTestFlagsOnlyImpl(int *argc, CharType **argv)
{
    for (int i = 1; i < *argc; i++) {
        const std::string arg_string = StreamableToString(argv[i]);
        const char *const arg = arg_string.c_str();

        using internal::ParseBoolFlag;
        using internal::ParseInt32Flag;
        using internal::ParseStringFlag;

        bool remove_flag = false;
        if (ParseGoogleTestFlag(arg)) {
            remove_flag = true;
#if GTEST_USE_OWN_FLAGFILE_FLAG_
        } else if (ParseStringFlag(arg, kFlagfileFlag, &GTEST_FLAG(flagfile))) {
            LoadFlagsFromFile(GTEST_FLAG(flagfile));
            remove_flag = true;
#endif // GTEST_USE_OWN_FLAGFILE_FLAG_
        } else if (arg_string == "--help" || arg_string == "-h" || arg_string == "-?" || arg_string == "/?" || HasGoogleTestFlagPrefix(arg)) {
            // Both help flag and unrecognized Google Test flags (excluding
            // internal ones) trigger help display.
            g_help_flag = true;
        }

        if (remove_flag) {
            // Shift the remainder of the argv list left by one.  Note
            // that argv has (*argc + 1) elements, the last one always being
            // NULL.  The following loop moves the trailing NULL element as
            // well.
            for (int j = i; j != *argc; j++) {
                argv[j] = argv[j + 1];
            }

            // Decrements the argument count.
            (*argc)--;

            // We also need to decrement the iterator as we just removed
            // an element.
            i--;
        }
    }

    if (g_help_flag) {
        // We print the help here instead of in RUN_ALL_TESTS(), as the
        // latter may not be called at all if the user is using Google
        // Test with another testing framework.
        PrintColorEncoded(kColorEncodedHelpMessage);
    }
}

// Parses the command line for Google Test flags, without initializing
// other parts of Google Test.
void ParseGoogleTestFlagsOnly(int *argc, char **argv)
{
    ParseGoogleTestFlagsOnlyImpl(argc, argv);

    // Fix the value of *_NSGetArgc() on macOS, but if and only if
    // *_NSGetArgv() == argv
    // Only applicable to char** version of argv
#if GTEST_OS_MAC
#ifndef GTEST_OS_IOS
    if (*_NSGetArgv() == argv) {
        *_NSGetArgc() = *argc;
    }
#endif
#endif
}
void ParseGoogleTestFlagsOnly(int *argc, wchar_t **argv)
{
    ParseGoogleTestFlagsOnlyImpl(argc, argv);
}

// The internal implementation of InitGoogleTest().
//
// The type parameter CharType can be instantiated to either char or
// wchar_t.
template<typename CharType>
void InitGoogleTestImpl(int *argc, CharType **argv)
{
    // We don't want to run the initialization code twice.
    if (GTestIsInitialized())
        return;

    if (*argc <= 0)
        return;

    g_argvs.clear();
    for (int i = 0; i != *argc; i++) {
        g_argvs.push_back(StreamableToString(argv[i]));
    }

#if GTEST_HAS_ABSL
    absl::InitializeSymbolizer(g_argvs[0].c_str());
#endif // GTEST_HAS_ABSL

    ParseGoogleTestFlagsOnly(argc, argv);
    GetUnitTestImpl()->PostFlagParsingInit();
}

} // namespace internal

// Initializes Google Test.  This must be called before calling
// RUN_ALL_TESTS().  In particular, it parses a command line for the
// flags that Google Test recognizes.  Whenever a Google Test flag is
// seen, it is removed from argv, and *argc is decremented.
//
// No value is returned.  Instead, the Google Test flag variables are
// updated.
//
// Calling the function for the second time has no user-visible effect.
void InitGoogleTest(int *argc, char **argv)
{
#if defined(GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_)
    GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_(argc, argv);
#else // defined(GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_)
    internal::InitGoogleTestImpl(argc, argv);
#endif // defined(GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_)
}

// This overloaded version can be used in Windows programs compiled in
// UNICODE mode.
void InitGoogleTest(int *argc, wchar_t **argv)
{
#if defined(GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_)
    GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_(argc, argv);
#else // defined(GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_)
    internal::InitGoogleTestImpl(argc, argv);
#endif // defined(GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_)
}

// This overloaded version can be used on Arduino/embedded platforms where
// there is no argc/argv.
void InitGoogleTest()
{
    // Since Arduino doesn't have a command line, fake out the argc/argv arguments
    int argc = 1;
    const auto arg0 = "dummy";
    char *argv0 = const_cast<char *>(arg0);
    char **argv = &argv0;

#if defined(GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_)
    GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_(&argc, argv);
#else // defined(GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_)
    internal::InitGoogleTestImpl(&argc, argv);
#endif // defined(GTEST_CUSTOM_INIT_GOOGLE_TEST_FUNCTION_)
}

std::string TempDir()
{
#if defined(GTEST_CUSTOM_TEMPDIR_FUNCTION_)
    return GTEST_CUSTOM_TEMPDIR_FUNCTION_();
#endif

#if GTEST_OS_WINDOWS_MOBILE
    return "\\temp\\";
#elif GTEST_OS_WINDOWS
    const char *temp_dir = internal::posix::GetEnv("TEMP");
    if (temp_dir == nullptr || temp_dir[0] == '\0')
        return "\\temp\\";
    else if (temp_dir[strlen(temp_dir) - 1] == '\\')
        return temp_dir;
    else
        return std::string(temp_dir) + "\\";
#elif GTEST_OS_LINUX_ANDROID
    const char *temp_dir = internal::posix::GetEnv("TEST_TMPDIR");
    if (temp_dir == nullptr || temp_dir[0] == '\0')
        return "/data/local/tmp/";
    else
        return temp_dir;
#else
    return "/tmp/";
#endif // GTEST_OS_WINDOWS_MOBILE
}

// Class ScopedTrace

// Pushes the given source file location and message onto a per-thread
// trace stack maintained by Google Test.
void ScopedTrace::PushTrace(const char *file, int line, std::string message)
{
    internal::TraceInfo trace;
    trace.file = file;
    trace.line = line;
    trace.message.swap(message);

    UnitTest::GetInstance()->PushGTestTrace(trace);
}

// Pops the info pushed by the c'tor.
ScopedTrace::~ScopedTrace()
    GTEST_LOCK_EXCLUDED_(&UnitTest::mutex_)
{
    UnitTest::GetInstance()->PopGTestTrace();
}

} // namespace testing
