// Copyright 2013 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "base/test/launcher/test_launcher.h" #include #include #include #include #include #include "base/at_exit.h" #include "base/bind.h" #include "base/command_line.h" #include "base/environment.h" #include "base/files/file_path.h" #include "base/files/file_util.h" #include "base/files/scoped_file.h" #include "base/format_macros.h" #include "base/hash.h" #include "base/lazy_instance.h" #include "base/location.h" #include "base/logging.h" #include "base/macros.h" #include "base/memory/ptr_util.h" #include "base/numerics/safe_conversions.h" #include "base/process/kill.h" #include "base/process/launch.h" #include "base/run_loop.h" #include "base/single_thread_task_runner.h" #include "base/strings/pattern.h" #include "base/strings/string_number_conversions.h" #include "base/strings/string_piece.h" #include "base/strings/string_split.h" #include "base/strings/string_util.h" #include "base/strings/stringize_macros.h" #include "base/strings/stringprintf.h" #include "base/strings/utf_string_conversions.h" #include "base/sys_info.h" #include "base/task_scheduler/post_task.h" #include "base/task_scheduler/task_scheduler.h" #include "base/test/gtest_util.h" #include "base/test/launcher/test_launcher_tracer.h" #include "base/test/launcher/test_results_tracker.h" #include "base/test/test_switches.h" #include "base/test/test_timeouts.h" #include "base/threading/thread_restrictions.h" #include "base/threading/thread_task_runner_handle.h" #include "base/time/time.h" #include "build/build_config.h" #include "testing/gtest/include/gtest/gtest.h" #if defined(OS_POSIX) #include #include "base/files/file_descriptor_watcher_posix.h" #endif #if defined(OS_MACOSX) #include "base/mac/scoped_nsautorelease_pool.h" #endif #if defined(OS_WIN) #include "base/win/windows_version.h" #endif #if defined(OS_FUCHSIA) #include #include "base/fuchsia/default_job.h" #include "base/fuchsia/fuchsia_logging.h" #endif namespace base { // See https://groups.google.com/a/chromium.org/d/msg/chromium-dev/nkdTP7sstSc/uT3FaE_sgkAJ . using ::operator<<; // The environment variable name for the total number of test shards. const char kTestTotalShards[] = "GTEST_TOTAL_SHARDS"; // The environment variable name for the test shard index. const char kTestShardIndex[] = "GTEST_SHARD_INDEX"; namespace { // Global tag for test runs where the results are incomplete or unreliable // for any reason, e.g. early exit because of too many broken tests. const char kUnreliableResultsTag[] = "UNRELIABLE_RESULTS"; // Maximum time of no output after which we print list of processes still // running. This deliberately doesn't use TestTimeouts (which is otherwise // a recommended solution), because they can be increased. This would defeat // the purpose of this timeout, which is 1) to avoid buildbot "no output for // X seconds" timeout killing the process 2) help communicate status of // the test launcher to people looking at the output (no output for a long // time is mysterious and gives no info about what is happening) 3) help // debugging in case the process hangs anyway. constexpr TimeDelta kOutputTimeout = TimeDelta::FromSeconds(15); // Limit of output snippet lines when printing to stdout. // Avoids flooding the logs with amount of output that gums up // the infrastructure. const size_t kOutputSnippetLinesLimit = 5000; // Limit of output snippet size. Exceeding this limit // results in truncating the output and failing the test. const size_t kOutputSnippetBytesLimit = 300 * 1024; // Limit of seed values for gtest shuffling. Arbitrary, but based on // gtest's similarly arbitrary choice. const uint32_t kRandomSeedUpperBound = 100000; // Set of live launch test processes with corresponding lock (it is allowed // for callers to launch processes on different threads). Lock* GetLiveProcessesLock() { static auto* lock = new Lock; return lock; } std::map* GetLiveProcesses() { static auto* map = new std::map; return map; } // Performance trace generator. TestLauncherTracer* GetTestLauncherTracer() { static auto* tracer = new TestLauncherTracer; return tracer; } // Creates and starts a TaskScheduler with |num_parallel_jobs| dedicated to // foreground blocking tasks (corresponds to the traits used to launch and wait // for child processes). void CreateAndStartTaskScheduler(int num_parallel_jobs) { // These values are taken from TaskScheduler::StartWithDefaultParams(), which // is not used directly to allow a custom number of threads in the foreground // blocking pool. constexpr int kMaxBackgroundThreads = 1; constexpr int kMaxBackgroundBlockingThreads = 2; const int max_foreground_threads = std::max(1, base::SysInfo::NumberOfProcessors()); constexpr base::TimeDelta kSuggestedReclaimTime = base::TimeDelta::FromSeconds(30); base::TaskScheduler::Create("TestLauncher"); base::TaskScheduler::GetInstance()->Start( {{kMaxBackgroundThreads, kSuggestedReclaimTime}, {kMaxBackgroundBlockingThreads, kSuggestedReclaimTime}, {max_foreground_threads, kSuggestedReclaimTime}, {num_parallel_jobs, kSuggestedReclaimTime}}); } #if defined(OS_POSIX) // Self-pipe that makes it possible to do complex shutdown handling // outside of the signal handler. int g_shutdown_pipe[2] = { -1, -1 }; void ShutdownPipeSignalHandler(int signal) { HANDLE_EINTR(write(g_shutdown_pipe[1], "q", 1)); } void KillSpawnedTestProcesses() { // Keep the lock until exiting the process to prevent further processes // from being spawned. AutoLock lock(*GetLiveProcessesLock()); fprintf(stdout, "Sending SIGTERM to %" PRIuS " child processes... ", GetLiveProcesses()->size()); fflush(stdout); for (const auto& pair : *GetLiveProcesses()) { // Send the signal to entire process group. kill((-1) * (pair.first), SIGTERM); } fprintf(stdout, "done.\nGiving processes a chance to terminate cleanly... "); fflush(stdout); PlatformThread::Sleep(TimeDelta::FromMilliseconds(500)); fprintf(stdout, "done.\n"); fflush(stdout); fprintf(stdout, "Sending SIGKILL to %" PRIuS " child processes... ", GetLiveProcesses()->size()); fflush(stdout); for (const auto& pair : *GetLiveProcesses()) { // Send the signal to entire process group. kill((-1) * (pair.first), SIGKILL); } fprintf(stdout, "done.\n"); fflush(stdout); } #endif // defined(OS_POSIX) // Parses the environment variable var as an Int32. If it is unset, returns // true. If it is set, unsets it then converts it to Int32 before // returning it in |result|. Returns true on success. bool TakeInt32FromEnvironment(const char* const var, int32_t* result) { std::unique_ptr env(Environment::Create()); std::string str_val; if (!env->GetVar(var, &str_val)) return true; if (!env->UnSetVar(var)) { LOG(ERROR) << "Invalid environment: we could not unset " << var << ".\n"; return false; } if (!StringToInt(str_val, result)) { LOG(ERROR) << "Invalid environment: " << var << " is not an integer.\n"; return false; } return true; } // Unsets the environment variable |name| and returns true on success. // Also returns true if the variable just doesn't exist. bool UnsetEnvironmentVariableIfExists(const std::string& name) { std::unique_ptr env(Environment::Create()); std::string str_val; if (!env->GetVar(name, &str_val)) return true; return env->UnSetVar(name); } // Returns true if bot mode has been requested, i.e. defaults optimized // for continuous integration bots. This way developers don't have to remember // special command-line flags. bool BotModeEnabled() { std::unique_ptr env(Environment::Create()); return CommandLine::ForCurrentProcess()->HasSwitch( switches::kTestLauncherBotMode) || env->HasVar("CHROMIUM_TEST_LAUNCHER_BOT_MODE"); } // Returns command line command line after gtest-specific processing // and applying |wrapper|. CommandLine PrepareCommandLineForGTest(const CommandLine& command_line, const std::string& wrapper) { CommandLine new_command_line(command_line.GetProgram()); CommandLine::SwitchMap switches = command_line.GetSwitches(); // Handled by the launcher process. switches.erase(kGTestRepeatFlag); switches.erase(kGTestShuffleFlag); switches.erase(kGTestRandomSeedFlag); // Don't try to write the final XML report in child processes. switches.erase(kGTestOutputFlag); for (CommandLine::SwitchMap::const_iterator iter = switches.begin(); iter != switches.end(); ++iter) { new_command_line.AppendSwitchNative((*iter).first, (*iter).second); } // Prepend wrapper after last CommandLine quasi-copy operation. CommandLine // does not really support removing switches well, and trying to do that // on a CommandLine with a wrapper is known to break. // TODO(phajdan.jr): Give it a try to support CommandLine removing switches. #if defined(OS_WIN) new_command_line.PrependWrapper(ASCIIToUTF16(wrapper)); #else new_command_line.PrependWrapper(wrapper); #endif return new_command_line; } // Launches a child process using |command_line|. If the child process is still // running after |timeout|, it is terminated and |*was_timeout| is set to true. // Returns exit code of the process. int LaunchChildTestProcessWithOptions(const CommandLine& command_line, const LaunchOptions& options, int flags, TimeDelta timeout, ProcessLifetimeObserver* observer, bool* was_timeout) { TimeTicks start_time(TimeTicks::Now()); #if defined(OS_POSIX) // Make sure an option we rely on is present - see LaunchChildGTestProcess. DCHECK(options.new_process_group); #endif LaunchOptions new_options(options); #if defined(OS_WIN) DCHECK(!new_options.job_handle); win::ScopedHandle job_handle; if (flags & TestLauncher::USE_JOB_OBJECTS) { job_handle.Set(CreateJobObject(NULL, NULL)); if (!job_handle.IsValid()) { LOG(ERROR) << "Could not create JobObject."; return -1; } DWORD job_flags = JOB_OBJECT_LIMIT_KILL_ON_JOB_CLOSE; // Allow break-away from job since sandbox and few other places rely on it // on Windows versions prior to Windows 8 (which supports nested jobs). if (win::GetVersion() < win::VERSION_WIN8 && flags & TestLauncher::ALLOW_BREAKAWAY_FROM_JOB) { job_flags |= JOB_OBJECT_LIMIT_BREAKAWAY_OK; } if (!SetJobObjectLimitFlags(job_handle.Get(), job_flags)) { LOG(ERROR) << "Could not SetJobObjectLimitFlags."; return -1; } new_options.job_handle = job_handle.Get(); } #elif defined(OS_FUCHSIA) DCHECK(!new_options.job_handle); zx::job job_handle; zx_status_t result = zx::job::create(*GetDefaultJob(), 0, &job_handle); ZX_CHECK(ZX_OK == result, result) << "zx_job_create"; new_options.job_handle = job_handle.get(); #endif // defined(OS_FUCHSIA) #if defined(OS_LINUX) // To prevent accidental privilege sharing to an untrusted child, processes // are started with PR_SET_NO_NEW_PRIVS. Do not set that here, since this // new child will be privileged and trusted. new_options.allow_new_privs = true; #endif Process process; { // Note how we grab the lock before the process possibly gets created. // This ensures that when the lock is held, ALL the processes are registered // in the set. AutoLock lock(*GetLiveProcessesLock()); #if defined(OS_WIN) // Allow the handle used to capture stdio and stdout to be inherited by the // child. Note that this is done under GetLiveProcessesLock() to ensure that // only the desired child receives the handle. if (new_options.stdout_handle) { ::SetHandleInformation(new_options.stdout_handle, HANDLE_FLAG_INHERIT, HANDLE_FLAG_INHERIT); } #endif process = LaunchProcess(command_line, new_options); #if defined(OS_WIN) // Revoke inheritance so that the handle isn't leaked into other children. // Note that this is done under GetLiveProcessesLock() to ensure that only // the desired child receives the handle. if (new_options.stdout_handle) ::SetHandleInformation(new_options.stdout_handle, HANDLE_FLAG_INHERIT, 0); #endif if (!process.IsValid()) return -1; // TODO(rvargas) crbug.com/417532: Don't store process handles. GetLiveProcesses()->insert(std::make_pair(process.Handle(), command_line)); } if (observer) observer->OnLaunched(process.Handle(), process.Pid()); int exit_code = 0; bool did_exit = false; { base::ScopedAllowBaseSyncPrimitivesForTesting allow_base_sync_primitives; did_exit = process.WaitForExitWithTimeout(timeout, &exit_code); } if (!did_exit) { if (observer) observer->OnTimedOut(command_line); *was_timeout = true; exit_code = -1; // Set a non-zero exit code to signal a failure. { base::ScopedAllowBaseSyncPrimitivesForTesting allow_base_sync_primitives; // Ensure that the process terminates. process.Terminate(-1, true); } } { // Note how we grab the log before issuing a possibly broad process kill. // Other code parts that grab the log kill processes, so avoid trying // to do that twice and trigger all kinds of log messages. AutoLock lock(*GetLiveProcessesLock()); #if defined(OS_FUCHSIA) zx_status_t status = job_handle.kill(); ZX_CHECK(status == ZX_OK, status); #elif defined(OS_POSIX) if (exit_code != 0) { // On POSIX, in case the test does not exit cleanly, either due to a crash // or due to it timing out, we need to clean up any child processes that // it might have created. On Windows, child processes are automatically // cleaned up using JobObjects. KillProcessGroup(process.Handle()); } #endif GetLiveProcesses()->erase(process.Handle()); } GetTestLauncherTracer()->RecordProcessExecution( start_time, TimeTicks::Now() - start_time); return exit_code; } void DoLaunchChildTestProcess( const CommandLine& command_line, TimeDelta timeout, const TestLauncher::LaunchOptions& test_launch_options, bool redirect_stdio, SingleThreadTaskRunner* task_runner, std::unique_ptr observer) { TimeTicks start_time = TimeTicks::Now(); ScopedFILE output_file; FilePath output_filename; if (redirect_stdio) { FILE* raw_output_file = CreateAndOpenTemporaryFile(&output_filename); output_file.reset(raw_output_file); CHECK(output_file); } LaunchOptions options; #if defined(OS_WIN) options.inherit_mode = test_launch_options.inherit_mode; options.handles_to_inherit = test_launch_options.handles_to_inherit; if (redirect_stdio) { HANDLE handle = reinterpret_cast(_get_osfhandle(_fileno(output_file.get()))); CHECK_NE(INVALID_HANDLE_VALUE, handle); options.stdin_handle = INVALID_HANDLE_VALUE; options.stdout_handle = handle; options.stderr_handle = handle; // See LaunchOptions.stdout_handle comments for why this compares against // FILE_TYPE_CHAR. if (options.inherit_mode == base::LaunchOptions::Inherit::kSpecific && GetFileType(handle) != FILE_TYPE_CHAR) { options.handles_to_inherit.push_back(handle); } } #else // if !defined(OS_WIN) options.fds_to_remap = test_launch_options.fds_to_remap; if (redirect_stdio) { int output_file_fd = fileno(output_file.get()); CHECK_LE(0, output_file_fd); options.fds_to_remap.push_back( std::make_pair(output_file_fd, STDOUT_FILENO)); options.fds_to_remap.push_back( std::make_pair(output_file_fd, STDERR_FILENO)); } #if !defined(OS_FUCHSIA) options.new_process_group = true; #endif #if defined(OS_LINUX) options.kill_on_parent_death = true; #endif #endif // !defined(OS_WIN) bool was_timeout = false; int exit_code = LaunchChildTestProcessWithOptions( command_line, options, test_launch_options.flags, timeout, observer.get(), &was_timeout); std::string output_file_contents; if (redirect_stdio) { fflush(output_file.get()); output_file.reset(); // Reading the file can sometimes fail when the process was killed midflight // (e.g. on test suite timeout): https://crbug.com/826408. Attempt to read // the output file anyways, but do not crash on failure in this case. CHECK(ReadFileToString(output_filename, &output_file_contents) || exit_code != 0); if (!DeleteFile(output_filename, false)) { // This needs to be non-fatal at least for Windows. LOG(WARNING) << "Failed to delete " << output_filename.AsUTF8Unsafe(); } } // Invoke OnCompleted on the thread it was originating from, not on a worker // pool thread. task_runner->PostTask( FROM_HERE, BindOnce(&ProcessLifetimeObserver::OnCompleted, std::move(observer), exit_code, TimeTicks::Now() - start_time, was_timeout, output_file_contents)); } } // namespace const char kGTestBreakOnFailure[] = "gtest_break_on_failure"; const char kGTestFilterFlag[] = "gtest_filter"; const char kGTestFlagfileFlag[] = "gtest_flagfile"; const char kGTestHelpFlag[] = "gtest_help"; const char kGTestListTestsFlag[] = "gtest_list_tests"; const char kGTestRepeatFlag[] = "gtest_repeat"; const char kGTestRunDisabledTestsFlag[] = "gtest_also_run_disabled_tests"; const char kGTestOutputFlag[] = "gtest_output"; const char kGTestShuffleFlag[] = "gtest_shuffle"; const char kGTestRandomSeedFlag[] = "gtest_random_seed"; TestLauncherDelegate::~TestLauncherDelegate() = default; TestLauncher::LaunchOptions::LaunchOptions() = default; TestLauncher::LaunchOptions::LaunchOptions(const LaunchOptions& other) = default; TestLauncher::LaunchOptions::~LaunchOptions() = default; TestLauncher::TestLauncher(TestLauncherDelegate* launcher_delegate, size_t parallel_jobs) : launcher_delegate_(launcher_delegate), total_shards_(1), shard_index_(0), cycles_(1), test_found_count_(0), test_started_count_(0), test_finished_count_(0), test_success_count_(0), test_broken_count_(0), retry_count_(0), retry_limit_(0), force_run_broken_tests_(false), run_result_(true), shuffle_(false), shuffle_seed_(0), watchdog_timer_(FROM_HERE, kOutputTimeout, this, &TestLauncher::OnOutputTimeout), parallel_jobs_(parallel_jobs) {} TestLauncher::~TestLauncher() { if (base::TaskScheduler::GetInstance()) { base::TaskScheduler::GetInstance()->Shutdown(); } } bool TestLauncher::Run() { if (!Init()) return false; // Value of |cycles_| changes after each iteration. Keep track of the // original value. int requested_cycles = cycles_; #if defined(OS_POSIX) CHECK_EQ(0, pipe(g_shutdown_pipe)); struct sigaction action; memset(&action, 0, sizeof(action)); sigemptyset(&action.sa_mask); action.sa_handler = &ShutdownPipeSignalHandler; CHECK_EQ(0, sigaction(SIGINT, &action, nullptr)); CHECK_EQ(0, sigaction(SIGQUIT, &action, nullptr)); CHECK_EQ(0, sigaction(SIGTERM, &action, nullptr)); auto controller = base::FileDescriptorWatcher::WatchReadable( g_shutdown_pipe[0], base::Bind(&TestLauncher::OnShutdownPipeReadable, Unretained(this))); #endif // defined(OS_POSIX) // Start the watchdog timer. watchdog_timer_.Reset(); ThreadTaskRunnerHandle::Get()->PostTask( FROM_HERE, BindOnce(&TestLauncher::RunTestIteration, Unretained(this))); RunLoop().Run(); if (requested_cycles != 1) results_tracker_.PrintSummaryOfAllIterations(); MaybeSaveSummaryAsJSON(std::vector()); return run_result_; } void TestLauncher::LaunchChildGTestProcess( const CommandLine& command_line, const std::string& wrapper, TimeDelta timeout, const LaunchOptions& options, std::unique_ptr observer) { DCHECK(thread_checker_.CalledOnValidThread()); // Record the exact command line used to launch the child. CommandLine new_command_line( PrepareCommandLineForGTest(command_line, wrapper)); // When running in parallel mode we need to redirect stdio to avoid mixed-up // output. We also always redirect on the bots to get the test output into // JSON summary. bool redirect_stdio = (parallel_jobs_ > 1) || BotModeEnabled(); PostTaskWithTraits( FROM_HERE, {MayBlock(), TaskShutdownBehavior::BLOCK_SHUTDOWN}, BindOnce(&DoLaunchChildTestProcess, new_command_line, timeout, options, redirect_stdio, RetainedRef(ThreadTaskRunnerHandle::Get()), std::move(observer))); } void TestLauncher::OnTestFinished(const TestResult& original_result) { ++test_finished_count_; TestResult result(original_result); if (result.output_snippet.length() > kOutputSnippetBytesLimit) { if (result.status == TestResult::TEST_SUCCESS) result.status = TestResult::TEST_EXCESSIVE_OUTPUT; // Keep the top and bottom of the log and truncate the middle part. result.output_snippet = result.output_snippet.substr(0, kOutputSnippetBytesLimit / 2) + "\n" + StringPrintf("\n", result.output_snippet.length()) + result.output_snippet.substr(result.output_snippet.length() - kOutputSnippetBytesLimit / 2) + "\n"; } bool print_snippet = false; std::string print_test_stdio("auto"); if (CommandLine::ForCurrentProcess()->HasSwitch( switches::kTestLauncherPrintTestStdio)) { print_test_stdio = CommandLine::ForCurrentProcess()->GetSwitchValueASCII( switches::kTestLauncherPrintTestStdio); } if (print_test_stdio == "auto") { print_snippet = (result.status != TestResult::TEST_SUCCESS); } else if (print_test_stdio == "always") { print_snippet = true; } else if (print_test_stdio == "never") { print_snippet = false; } else { LOG(WARNING) << "Invalid value of " << switches::kTestLauncherPrintTestStdio << ": " << print_test_stdio; } if (print_snippet) { std::vector snippet_lines = SplitStringPiece(result.output_snippet, "\n", base::KEEP_WHITESPACE, base::SPLIT_WANT_ALL); if (snippet_lines.size() > kOutputSnippetLinesLimit) { size_t truncated_size = snippet_lines.size() - kOutputSnippetLinesLimit; snippet_lines.erase( snippet_lines.begin(), snippet_lines.begin() + truncated_size); snippet_lines.insert(snippet_lines.begin(), ""); } fprintf(stdout, "%s", base::JoinString(snippet_lines, "\n").c_str()); fflush(stdout); } if (result.status == TestResult::TEST_SUCCESS) { ++test_success_count_; } else { tests_to_retry_.insert(result.full_name); } results_tracker_.AddTestResult(result); // TODO(phajdan.jr): Align counter (padding). std::string status_line( StringPrintf("[%" PRIuS "/%" PRIuS "] %s ", test_finished_count_, test_started_count_, result.full_name.c_str())); if (result.completed()) { status_line.append(StringPrintf("(%" PRId64 " ms)", result.elapsed_time.InMilliseconds())); } else if (result.status == TestResult::TEST_TIMEOUT) { status_line.append("(TIMED OUT)"); } else if (result.status == TestResult::TEST_CRASH) { status_line.append("(CRASHED)"); } else if (result.status == TestResult::TEST_SKIPPED) { status_line.append("(SKIPPED)"); } else if (result.status == TestResult::TEST_UNKNOWN) { status_line.append("(UNKNOWN)"); } else { // Fail very loudly so it's not ignored. CHECK(false) << "Unhandled test result status: " << result.status; } fprintf(stdout, "%s\n", status_line.c_str()); fflush(stdout); // We just printed a status line, reset the watchdog timer. watchdog_timer_.Reset(); // Do not waste time on timeouts. We include tests with unknown results here // because sometimes (e.g. hang in between unit tests) that's how a timeout // gets reported. if (result.status == TestResult::TEST_TIMEOUT || result.status == TestResult::TEST_UNKNOWN) { test_broken_count_++; } size_t broken_threshold = std::max(static_cast(20), test_found_count_ / 10); if (!force_run_broken_tests_ && test_broken_count_ >= broken_threshold) { fprintf(stdout, "Too many badly broken tests (%" PRIuS "), exiting now.\n", test_broken_count_); fflush(stdout); #if defined(OS_POSIX) KillSpawnedTestProcesses(); #endif // defined(OS_POSIX) MaybeSaveSummaryAsJSON({"BROKEN_TEST_EARLY_EXIT", kUnreliableResultsTag}); exit(1); } if (test_finished_count_ != test_started_count_) return; if (tests_to_retry_.empty() || retry_count_ >= retry_limit_) { OnTestIterationFinished(); return; } if (!force_run_broken_tests_ && tests_to_retry_.size() >= broken_threshold) { fprintf(stdout, "Too many failing tests (%" PRIuS "), skipping retries.\n", tests_to_retry_.size()); fflush(stdout); results_tracker_.AddGlobalTag("BROKEN_TEST_SKIPPED_RETRIES"); results_tracker_.AddGlobalTag(kUnreliableResultsTag); OnTestIterationFinished(); return; } retry_count_++; std::vector test_names(tests_to_retry_.begin(), tests_to_retry_.end()); tests_to_retry_.clear(); size_t retry_started_count = launcher_delegate_->RetryTests(this, test_names); if (retry_started_count == 0) { // Signal failure, but continue to run all requested test iterations. // With the summary of all iterations at the end this is a good default. run_result_ = false; OnTestIterationFinished(); return; } fprintf(stdout, "Retrying %" PRIuS " test%s (retry #%" PRIuS ")\n", retry_started_count, retry_started_count > 1 ? "s" : "", retry_count_); fflush(stdout); test_started_count_ += retry_started_count; } // Helper used to parse test filter files. Syntax is documented in // //testing/buildbot/filters/README.md . bool LoadFilterFile(const FilePath& file_path, std::vector* positive_filter, std::vector* negative_filter) { std::string file_content; if (!ReadFileToString(file_path, &file_content)) { LOG(ERROR) << "Failed to read the filter file."; return false; } std::vector filter_lines = SplitString( file_content, "\n", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL); int line_num = 0; for (const std::string& filter_line : filter_lines) { line_num++; size_t hash_pos = filter_line.find('#'); // In case when # symbol is not in the beginning of the line and is not // proceeded with a space then it's likely that the comment was // unintentional. if (hash_pos != std::string::npos && hash_pos > 0 && filter_line[hash_pos - 1] != ' ') { LOG(WARNING) << "Content of line " << line_num << " in " << file_path << " after # is treated as a comment, " << filter_line; } // Strip comments and whitespace from each line. std::string trimmed_line = TrimWhitespaceASCII(filter_line.substr(0, hash_pos), TRIM_ALL) .as_string(); if (trimmed_line.substr(0, 2) == "//") { LOG(ERROR) << "Line " << line_num << " in " << file_path << " starts with //, use # for comments."; return false; } // Treat a line starting with '//' as a comment. if (trimmed_line.empty()) continue; if (trimmed_line[0] == '-') negative_filter->push_back(trimmed_line.substr(1)); else positive_filter->push_back(trimmed_line); } return true; } bool TestLauncher::Init() { const CommandLine* command_line = CommandLine::ForCurrentProcess(); // Initialize sharding. Command line takes precedence over legacy environment // variables. if (command_line->HasSwitch(switches::kTestLauncherTotalShards) && command_line->HasSwitch(switches::kTestLauncherShardIndex)) { if (!StringToInt( command_line->GetSwitchValueASCII( switches::kTestLauncherTotalShards), &total_shards_)) { LOG(ERROR) << "Invalid value for " << switches::kTestLauncherTotalShards; return false; } if (!StringToInt( command_line->GetSwitchValueASCII( switches::kTestLauncherShardIndex), &shard_index_)) { LOG(ERROR) << "Invalid value for " << switches::kTestLauncherShardIndex; return false; } fprintf(stdout, "Using sharding settings from command line. This is shard %d/%d\n", shard_index_, total_shards_); fflush(stdout); } else { if (!TakeInt32FromEnvironment(kTestTotalShards, &total_shards_)) return false; if (!TakeInt32FromEnvironment(kTestShardIndex, &shard_index_)) return false; fprintf(stdout, "Using sharding settings from environment. This is shard %d/%d\n", shard_index_, total_shards_); fflush(stdout); } if (shard_index_ < 0 || total_shards_ < 0 || shard_index_ >= total_shards_) { LOG(ERROR) << "Invalid sharding settings: we require 0 <= " << kTestShardIndex << " < " << kTestTotalShards << ", but you have " << kTestShardIndex << "=" << shard_index_ << ", " << kTestTotalShards << "=" << total_shards_ << ".\n"; return false; } // Make sure we don't pass any sharding-related environment to the child // processes. This test launcher implements the sharding completely. CHECK(UnsetEnvironmentVariableIfExists("GTEST_TOTAL_SHARDS")); CHECK(UnsetEnvironmentVariableIfExists("GTEST_SHARD_INDEX")); if (command_line->HasSwitch(kGTestRepeatFlag) && !StringToInt(command_line->GetSwitchValueASCII(kGTestRepeatFlag), &cycles_)) { LOG(ERROR) << "Invalid value for " << kGTestRepeatFlag; return false; } if (command_line->HasSwitch(switches::kTestLauncherRetryLimit)) { int retry_limit = -1; if (!StringToInt(command_line->GetSwitchValueASCII( switches::kTestLauncherRetryLimit), &retry_limit) || retry_limit < 0) { LOG(ERROR) << "Invalid value for " << switches::kTestLauncherRetryLimit; return false; } retry_limit_ = retry_limit; } else if (!command_line->HasSwitch(kGTestFilterFlag) || BotModeEnabled()) { // Retry failures 3 times by default if we are running all of the tests or // in bot mode. retry_limit_ = 3; } if (command_line->HasSwitch(switches::kTestLauncherForceRunBrokenTests)) force_run_broken_tests_ = true; // Some of the TestLauncherDelegate implementations don't call into gtest // until they've already split into test-specific processes. This results // in gtest's native shuffle implementation attempting to shuffle one test. // Shuffling the list of tests in the test launcher (before the delegate // gets involved) ensures that the entire shard is shuffled. if (command_line->HasSwitch(kGTestShuffleFlag)) { shuffle_ = true; if (command_line->HasSwitch(kGTestRandomSeedFlag)) { const std::string custom_seed_str = command_line->GetSwitchValueASCII(kGTestRandomSeedFlag); uint32_t custom_seed = 0; if (!StringToUint(custom_seed_str, &custom_seed)) { LOG(ERROR) << "Unable to parse seed \"" << custom_seed_str << "\"."; return false; } if (custom_seed >= kRandomSeedUpperBound) { LOG(ERROR) << "Seed " << custom_seed << " outside of expected range " << "[0, " << kRandomSeedUpperBound << ")"; return false; } shuffle_seed_ = custom_seed; } else { std::uniform_int_distribution dist(0, kRandomSeedUpperBound); std::random_device random_dev; shuffle_seed_ = dist(random_dev); } } else if (command_line->HasSwitch(kGTestRandomSeedFlag)) { LOG(ERROR) << kGTestRandomSeedFlag << " requires " << kGTestShuffleFlag; return false; } fprintf(stdout, "Using %" PRIuS " parallel jobs.\n", parallel_jobs_); fflush(stdout); CreateAndStartTaskScheduler(static_cast(parallel_jobs_)); std::vector positive_file_filter; std::vector positive_gtest_filter; if (command_line->HasSwitch(switches::kTestLauncherFilterFile)) { base::FilePath filter_file_path = base::MakeAbsoluteFilePath( command_line->GetSwitchValuePath(switches::kTestLauncherFilterFile)); if (!LoadFilterFile(filter_file_path, &positive_file_filter, &negative_test_filter_)) return false; } // Split --gtest_filter at '-', if there is one, to separate into // positive filter and negative filter portions. std::string filter = command_line->GetSwitchValueASCII(kGTestFilterFlag); size_t dash_pos = filter.find('-'); if (dash_pos == std::string::npos) { positive_gtest_filter = SplitString(filter, ":", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL); } else { // Everything up to the dash. positive_gtest_filter = SplitString(filter.substr(0, dash_pos), ":", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL); // Everything after the dash. for (std::string pattern : SplitString(filter.substr(dash_pos + 1), ":", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL)) { negative_test_filter_.push_back(pattern); } } if (!launcher_delegate_->GetTests(&tests_)) { LOG(ERROR) << "Failed to get list of tests."; return false; } CombinePositiveTestFilters(std::move(positive_gtest_filter), std::move(positive_file_filter)); if (!results_tracker_.Init(*command_line)) { LOG(ERROR) << "Failed to initialize test results tracker."; return 1; } #if defined(NDEBUG) results_tracker_.AddGlobalTag("MODE_RELEASE"); #else results_tracker_.AddGlobalTag("MODE_DEBUG"); #endif // Operating systems (sorted alphabetically). // Note that they can deliberately overlap, e.g. OS_LINUX is a subset // of OS_POSIX. #if defined(OS_ANDROID) results_tracker_.AddGlobalTag("OS_ANDROID"); #endif #if defined(OS_BSD) results_tracker_.AddGlobalTag("OS_BSD"); #endif #if defined(OS_FREEBSD) results_tracker_.AddGlobalTag("OS_FREEBSD"); #endif #if defined(OS_FUCHSIA) results_tracker_.AddGlobalTag("OS_FUCHSIA"); #endif #if defined(OS_IOS) results_tracker_.AddGlobalTag("OS_IOS"); #endif #if defined(OS_LINUX) results_tracker_.AddGlobalTag("OS_LINUX"); #endif #if defined(OS_MACOSX) results_tracker_.AddGlobalTag("OS_MACOSX"); #endif #if defined(OS_NACL) results_tracker_.AddGlobalTag("OS_NACL"); #endif #if defined(OS_OPENBSD) results_tracker_.AddGlobalTag("OS_OPENBSD"); #endif #if defined(OS_POSIX) results_tracker_.AddGlobalTag("OS_POSIX"); #endif #if defined(OS_SOLARIS) results_tracker_.AddGlobalTag("OS_SOLARIS"); #endif #if defined(OS_WIN) results_tracker_.AddGlobalTag("OS_WIN"); #endif // CPU-related tags. #if defined(ARCH_CPU_32_BITS) results_tracker_.AddGlobalTag("CPU_32_BITS"); #endif #if defined(ARCH_CPU_64_BITS) results_tracker_.AddGlobalTag("CPU_64_BITS"); #endif return true; } void TestLauncher::CombinePositiveTestFilters( std::vector filter_a, std::vector filter_b) { has_at_least_one_positive_filter_ = !filter_a.empty() || !filter_b.empty(); if (!has_at_least_one_positive_filter_) { return; } // If two positive filters are present, only run tests that match a pattern // in both filters. if (!filter_a.empty() && !filter_b.empty()) { for (size_t i = 0; i < tests_.size(); i++) { std::string test_name = FormatFullTestName(tests_[i].test_case_name, tests_[i].test_name); bool found_a = false; bool found_b = false; for (size_t k = 0; k < filter_a.size(); ++k) { found_a = found_a || MatchPattern(test_name, filter_a[k]); } for (size_t k = 0; k < filter_b.size(); ++k) { found_b = found_b || MatchPattern(test_name, filter_b[k]); } if (found_a && found_b) { positive_test_filter_.push_back(test_name); } } } else if (!filter_a.empty()) { positive_test_filter_ = std::move(filter_a); } else { positive_test_filter_ = std::move(filter_b); } } void TestLauncher::RunTests() { std::vector test_names; const CommandLine* command_line = CommandLine::ForCurrentProcess(); for (const TestIdentifier& test_id : tests_) { std::string test_name = FormatFullTestName(test_id.test_case_name, test_id.test_name); results_tracker_.AddTest(test_name); if (test_name.find("DISABLED") != std::string::npos) { results_tracker_.AddDisabledTest(test_name); // Skip disabled tests unless explicitly requested. if (!command_line->HasSwitch(kGTestRunDisabledTestsFlag)) continue; } if (!launcher_delegate_->ShouldRunTest(test_id.test_case_name, test_id.test_name)) { continue; } // Count tests in the binary, before we apply filter and sharding. test_found_count_++; std::string test_name_no_disabled = TestNameWithoutDisabledPrefix(test_name); // Skip the test that doesn't match the filter (if given). if (has_at_least_one_positive_filter_) { bool found = false; for (auto filter : positive_test_filter_) { if (MatchPattern(test_name, filter) || MatchPattern(test_name_no_disabled, filter)) { found = true; break; } } if (!found) continue; } if (!negative_test_filter_.empty()) { bool excluded = false; for (auto filter : negative_test_filter_) { if (MatchPattern(test_name, filter) || MatchPattern(test_name_no_disabled, filter)) { excluded = true; break; } } if (excluded) continue; } if (Hash(test_name) % total_shards_ != static_cast(shard_index_)) continue; // Report test locations after applying all filters, so that we report test // locations only for those tests that were run as part of this shard. results_tracker_.AddTestLocation(test_name, test_id.file, test_id.line); test_names.push_back(test_name); } if (shuffle_) { std::mt19937 randomizer; randomizer.seed(shuffle_seed_); std::shuffle(test_names.begin(), test_names.end(), randomizer); fprintf(stdout, "Randomizing with seed %u\n", shuffle_seed_); fflush(stdout); } // Save an early test summary in case the launcher crashes or gets killed. MaybeSaveSummaryAsJSON({"EARLY_SUMMARY", kUnreliableResultsTag}); test_started_count_ = launcher_delegate_->RunTests(this, test_names); if (test_started_count_ == 0) { fprintf(stdout, "0 tests run\n"); fflush(stdout); // No tests have actually been started, so kick off the next iteration. ThreadTaskRunnerHandle::Get()->PostTask( FROM_HERE, BindOnce(&TestLauncher::RunTestIteration, Unretained(this))); } } void TestLauncher::RunTestIteration() { const bool stop_on_failure = CommandLine::ForCurrentProcess()->HasSwitch(kGTestBreakOnFailure); if (cycles_ == 0 || (stop_on_failure && test_success_count_ != test_finished_count_)) { RunLoop::QuitCurrentWhenIdleDeprecated(); return; } // Special value "-1" means "repeat indefinitely". cycles_ = (cycles_ == -1) ? cycles_ : cycles_ - 1; test_found_count_ = 0; test_started_count_ = 0; test_finished_count_ = 0; test_success_count_ = 0; test_broken_count_ = 0; retry_count_ = 0; tests_to_retry_.clear(); results_tracker_.OnTestIterationStarting(); ThreadTaskRunnerHandle::Get()->PostTask( FROM_HERE, BindOnce(&TestLauncher::RunTests, Unretained(this))); } #if defined(OS_POSIX) // I/O watcher for the reading end of the self-pipe above. // Terminates any launched child processes and exits the process. void TestLauncher::OnShutdownPipeReadable() { fprintf(stdout, "\nCaught signal. Killing spawned test processes...\n"); fflush(stdout); KillSpawnedTestProcesses(); MaybeSaveSummaryAsJSON({"CAUGHT_TERMINATION_SIGNAL", kUnreliableResultsTag}); // The signal would normally kill the process, so exit now. _exit(1); } #endif // defined(OS_POSIX) void TestLauncher::MaybeSaveSummaryAsJSON( const std::vector& additional_tags) { const CommandLine* command_line = CommandLine::ForCurrentProcess(); if (command_line->HasSwitch(switches::kTestLauncherSummaryOutput)) { FilePath summary_path(command_line->GetSwitchValuePath( switches::kTestLauncherSummaryOutput)); if (!results_tracker_.SaveSummaryAsJSON(summary_path, additional_tags)) { LOG(ERROR) << "Failed to save test launcher output summary."; } } if (command_line->HasSwitch(switches::kTestLauncherTrace)) { FilePath trace_path( command_line->GetSwitchValuePath(switches::kTestLauncherTrace)); if (!GetTestLauncherTracer()->Dump(trace_path)) { LOG(ERROR) << "Failed to save test launcher trace."; } } } void TestLauncher::OnTestIterationFinished() { TestResultsTracker::TestStatusMap tests_by_status( results_tracker_.GetTestStatusMapForCurrentIteration()); if (!tests_by_status[TestResult::TEST_UNKNOWN].empty()) results_tracker_.AddGlobalTag(kUnreliableResultsTag); // When we retry tests, success is determined by having nothing more // to retry (everything eventually passed), as opposed to having // no failures at all. if (tests_to_retry_.empty()) { fprintf(stdout, "SUCCESS: all tests passed.\n"); fflush(stdout); } else { // Signal failure, but continue to run all requested test iterations. // With the summary of all iterations at the end this is a good default. run_result_ = false; } results_tracker_.PrintSummaryOfCurrentIteration(); // Kick off the next iteration. ThreadTaskRunnerHandle::Get()->PostTask( FROM_HERE, BindOnce(&TestLauncher::RunTestIteration, Unretained(this))); } void TestLauncher::OnOutputTimeout() { DCHECK(thread_checker_.CalledOnValidThread()); AutoLock lock(*GetLiveProcessesLock()); fprintf(stdout, "Still waiting for the following processes to finish:\n"); for (const auto& pair : *GetLiveProcesses()) { #if defined(OS_WIN) fwprintf(stdout, L"\t%s\n", pair.second.GetCommandLineString().c_str()); #else fprintf(stdout, "\t%s\n", pair.second.GetCommandLineString().c_str()); #endif } fflush(stdout); // Arm the timer again - otherwise it would fire only once. watchdog_timer_.Reset(); } size_t NumParallelJobs() { const CommandLine* command_line = CommandLine::ForCurrentProcess(); if (command_line->HasSwitch(switches::kTestLauncherJobs)) { // If the number of test launcher jobs was specified, return that number. size_t jobs = 0U; if (!StringToSizeT( command_line->GetSwitchValueASCII(switches::kTestLauncherJobs), &jobs) || !jobs) { LOG(ERROR) << "Invalid value for " << switches::kTestLauncherJobs; return 0U; } return jobs; } if (command_line->HasSwitch(kGTestFilterFlag) && !BotModeEnabled()) { // Do not run jobs in parallel by default if we are running a subset of // the tests and if bot mode is off. return 1U; } // Default to the number of processor cores. return base::checked_cast(SysInfo::NumberOfProcessors()); } std::string GetTestOutputSnippet(const TestResult& result, const std::string& full_output) { size_t run_pos = full_output.find(std::string("[ RUN ] ") + result.full_name); if (run_pos == std::string::npos) return std::string(); size_t end_pos = full_output.find(std::string("[ FAILED ] ") + result.full_name, run_pos); // Only clip the snippet to the "OK" message if the test really // succeeded. It still might have e.g. crashed after printing it. if (end_pos == std::string::npos && result.status == TestResult::TEST_SUCCESS) { end_pos = full_output.find(std::string("[ OK ] ") + result.full_name, run_pos); } if (end_pos != std::string::npos) { size_t newline_pos = full_output.find("\n", end_pos); if (newline_pos != std::string::npos) end_pos = newline_pos + 1; } std::string snippet(full_output.substr(run_pos)); if (end_pos != std::string::npos) snippet = full_output.substr(run_pos, end_pos - run_pos); return snippet; } } // namespace base