Spaces:
No application file
No application file
| # Copyright 2017 The Abseil Authors. | |
| # | |
| # Licensed under the Apache License, Version 2.0 (the "License"); | |
| # you may not use this file except in compliance with the License. | |
| # You may obtain a copy of the License at | |
| # | |
| # http://www.apache.org/licenses/LICENSE-2.0 | |
| # | |
| # Unless required by applicable law or agreed to in writing, software | |
| # distributed under the License is distributed on an "AS IS" BASIS, | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
| # See the License for the specific language governing permissions and | |
| # limitations under the License. | |
| """Base functionality for Abseil Python tests. | |
| This module contains base classes and high-level functions for Abseil-style | |
| tests. | |
| """ | |
| from collections import abc | |
| import contextlib | |
| import difflib | |
| import enum | |
| import errno | |
| import faulthandler | |
| import getpass | |
| import inspect | |
| import io | |
| import itertools | |
| import json | |
| import os | |
| import random | |
| import re | |
| import shlex | |
| import shutil | |
| import signal | |
| import stat | |
| import subprocess | |
| import sys | |
| import tempfile | |
| import textwrap | |
| import typing | |
| from typing import Any, AnyStr, BinaryIO, Callable, ContextManager, IO, Iterator, List, Mapping, MutableMapping, MutableSequence, NoReturn, Optional, Sequence, Text, TextIO, Tuple, Type, Union | |
| import unittest | |
| from unittest import mock # pylint: disable=unused-import Allow absltest.mock. | |
| from urllib import parse | |
| from absl import app # pylint: disable=g-import-not-at-top | |
| from absl import flags | |
| from absl import logging | |
| from absl.testing import _pretty_print_reporter | |
| from absl.testing import xml_reporter | |
| # Use an if-type-checking block to prevent leakage of type-checking only | |
| # symbols. We don't want people relying on these at runtime. | |
| if typing.TYPE_CHECKING: | |
| # Unbounded TypeVar for general usage | |
| _T = typing.TypeVar('_T') | |
| import unittest.case # pylint: disable=g-import-not-at-top,g-bad-import-order | |
| _OutcomeType = unittest.case._Outcome # pytype: disable=module-attr | |
| # pylint: enable=g-import-not-at-top | |
| # Re-export a bunch of unittest functions we support so that people don't | |
| # have to import unittest to get them | |
| # pylint: disable=invalid-name | |
| skip = unittest.skip | |
| skipIf = unittest.skipIf | |
| skipUnless = unittest.skipUnless | |
| SkipTest = unittest.SkipTest | |
| expectedFailure = unittest.expectedFailure | |
| # pylint: enable=invalid-name | |
| # End unittest re-exports | |
| FLAGS = flags.FLAGS | |
| _TEXT_OR_BINARY_TYPES = (str, bytes) | |
| # Suppress surplus entries in AssertionError stack traces. | |
| __unittest = True # pylint: disable=invalid-name | |
| def expectedFailureIf(condition, reason): # pylint: disable=invalid-name | |
| """Expects the test to fail if the run condition is True. | |
| Example usage:: | |
| @expectedFailureIf(sys.version.major == 2, "Not yet working in py2") | |
| def test_foo(self): | |
| ... | |
| Args: | |
| condition: bool, whether to expect failure or not. | |
| reason: Text, the reason to expect failure. | |
| Returns: | |
| Decorator function | |
| """ | |
| del reason # Unused | |
| if condition: | |
| return unittest.expectedFailure | |
| else: | |
| return lambda f: f | |
| class TempFileCleanup(enum.Enum): | |
| # Always cleanup temp files when the test completes. | |
| ALWAYS = 'always' | |
| # Only cleanup temp file if the test passes. This allows easier inspection | |
| # of tempfile contents on test failure. absltest.TEST_TMPDIR.value determines | |
| # where tempfiles are created. | |
| SUCCESS = 'success' | |
| # Never cleanup temp files. | |
| OFF = 'never' | |
| # Many of the methods in this module have names like assertSameElements. | |
| # This kind of name does not comply with PEP8 style, | |
| # but it is consistent with the naming of methods in unittest.py. | |
| # pylint: disable=invalid-name | |
| def _get_default_test_random_seed(): | |
| # type: () -> int | |
| random_seed = 301 | |
| value = os.environ.get('TEST_RANDOM_SEED', '') | |
| try: | |
| random_seed = int(value) | |
| except ValueError: | |
| pass | |
| return random_seed | |
| def get_default_test_srcdir(): | |
| # type: () -> Text | |
| """Returns default test source dir.""" | |
| return os.environ.get('TEST_SRCDIR', '') | |
| def get_default_test_tmpdir(): | |
| # type: () -> Text | |
| """Returns default test temp dir.""" | |
| tmpdir = os.environ.get('TEST_TMPDIR', '') | |
| if not tmpdir: | |
| tmpdir = os.path.join(tempfile.gettempdir(), 'absl_testing') | |
| return tmpdir | |
| def _get_default_randomize_ordering_seed(): | |
| # type: () -> int | |
| """Returns default seed to use for randomizing test order. | |
| This function first checks the --test_randomize_ordering_seed flag, and then | |
| the TEST_RANDOMIZE_ORDERING_SEED environment variable. If the first value | |
| we find is: | |
| * (not set): disable test randomization | |
| * 0: disable test randomization | |
| * 'random': choose a random seed in [1, 4294967295] for test order | |
| randomization | |
| * positive integer: use this seed for test order randomization | |
| (The values used are patterned after | |
| https://docs.python.org/3/using/cmdline.html#envvar-PYTHONHASHSEED). | |
| In principle, it would be simpler to return None if no override is provided; | |
| however, the python random module has no `get_seed()`, only `getstate()`, | |
| which returns far more data than we want to pass via an environment variable | |
| or flag. | |
| Returns: | |
| A default value for test case randomization (int). 0 means do not randomize. | |
| Raises: | |
| ValueError: Raised when the flag or env value is not one of the options | |
| above. | |
| """ | |
| if FLAGS['test_randomize_ordering_seed'].present: | |
| randomize = FLAGS.test_randomize_ordering_seed | |
| elif 'TEST_RANDOMIZE_ORDERING_SEED' in os.environ: | |
| randomize = os.environ['TEST_RANDOMIZE_ORDERING_SEED'] | |
| else: | |
| randomize = '' | |
| if not randomize: | |
| return 0 | |
| if randomize == 'random': | |
| return random.Random().randint(1, 4294967295) | |
| if randomize == '0': | |
| return 0 | |
| try: | |
| seed = int(randomize) | |
| if seed > 0: | |
| return seed | |
| except ValueError: | |
| pass | |
| raise ValueError( | |
| 'Unknown test randomization seed value: {}'.format(randomize)) | |
| TEST_SRCDIR = flags.DEFINE_string( | |
| 'test_srcdir', | |
| get_default_test_srcdir(), | |
| 'Root of directory tree where source files live', | |
| allow_override_cpp=True) | |
| TEST_TMPDIR = flags.DEFINE_string( | |
| 'test_tmpdir', | |
| get_default_test_tmpdir(), | |
| 'Directory for temporary testing files', | |
| allow_override_cpp=True) | |
| flags.DEFINE_integer( | |
| 'test_random_seed', | |
| _get_default_test_random_seed(), | |
| 'Random seed for testing. Some test frameworks may ' | |
| 'change the default value of this flag between runs, so ' | |
| 'it is not appropriate for seeding probabilistic tests.', | |
| allow_override_cpp=True) | |
| flags.DEFINE_string( | |
| 'test_randomize_ordering_seed', | |
| '', | |
| 'If positive, use this as a seed to randomize the ' | |
| 'execution order for test cases. If "random", pick a ' | |
| 'random seed to use. If 0 or not set, do not randomize ' | |
| 'test case execution order. This flag also overrides ' | |
| 'the TEST_RANDOMIZE_ORDERING_SEED environment variable.', | |
| allow_override_cpp=True) | |
| flags.DEFINE_string('xml_output_file', '', 'File to store XML test results') | |
| # We might need to monkey-patch TestResult so that it stops considering an | |
| # unexpected pass as a as a "successful result". For details, see | |
| # http://bugs.python.org/issue20165 | |
| def _monkey_patch_test_result_for_unexpected_passes(): | |
| # type: () -> None | |
| """Workaround for <http://bugs.python.org/issue20165>.""" | |
| def wasSuccessful(self): | |
| # type: () -> bool | |
| """Tells whether or not this result was a success. | |
| Any unexpected pass is to be counted as a non-success. | |
| Args: | |
| self: The TestResult instance. | |
| Returns: | |
| Whether or not this result was a success. | |
| """ | |
| return (len(self.failures) == len(self.errors) == | |
| len(self.unexpectedSuccesses) == 0) | |
| test_result = unittest.TestResult() | |
| test_result.addUnexpectedSuccess(unittest.FunctionTestCase(lambda: None)) | |
| if test_result.wasSuccessful(): # The bug is present. | |
| unittest.TestResult.wasSuccessful = wasSuccessful | |
| if test_result.wasSuccessful(): # Warn the user if our hot-fix failed. | |
| sys.stderr.write('unittest.result.TestResult monkey patch to report' | |
| ' unexpected passes as failures did not work.\n') | |
| _monkey_patch_test_result_for_unexpected_passes() | |
| def _open(filepath, mode, _open_func=open): | |
| # type: (Text, Text, Callable[..., IO]) -> IO | |
| """Opens a file. | |
| Like open(), but ensure that we can open real files even if tests stub out | |
| open(). | |
| Args: | |
| filepath: A filepath. | |
| mode: A mode. | |
| _open_func: A built-in open() function. | |
| Returns: | |
| The opened file object. | |
| """ | |
| return _open_func(filepath, mode, encoding='utf-8') | |
| class _TempDir(object): | |
| """Represents a temporary directory for tests. | |
| Creation of this class is internal. Using its public methods is OK. | |
| This class implements the `os.PathLike` interface (specifically, | |
| `os.PathLike[str]`). This means, in Python 3, it can be directly passed | |
| to e.g. `os.path.join()`. | |
| """ | |
| def __init__(self, path): | |
| # type: (Text) -> None | |
| """Module-private: do not instantiate outside module.""" | |
| self._path = path | |
| def full_path(self): | |
| # type: () -> Text | |
| """Returns the path, as a string, for the directory. | |
| TIP: Instead of e.g. `os.path.join(temp_dir.full_path)`, you can simply | |
| do `os.path.join(temp_dir)` because `__fspath__()` is implemented. | |
| """ | |
| return self._path | |
| def __fspath__(self): | |
| # type: () -> Text | |
| """See os.PathLike.""" | |
| return self.full_path | |
| def create_file(self, file_path=None, content=None, mode='w', encoding='utf8', | |
| errors='strict'): | |
| # type: (Optional[Text], Optional[AnyStr], Text, Text, Text) -> _TempFile | |
| """Create a file in the directory. | |
| NOTE: If the file already exists, it will be made writable and overwritten. | |
| Args: | |
| file_path: Optional file path for the temp file. If not given, a unique | |
| file name will be generated and used. Slashes are allowed in the name; | |
| any missing intermediate directories will be created. NOTE: This path | |
| is the path that will be cleaned up, including any directories in the | |
| path, e.g., 'foo/bar/baz.txt' will `rm -r foo` | |
| content: Optional string or bytes to initially write to the file. If not | |
| specified, then an empty file is created. | |
| mode: Mode string to use when writing content. Only used if `content` is | |
| non-empty. | |
| encoding: Encoding to use when writing string content. Only used if | |
| `content` is text. | |
| errors: How to handle text to bytes encoding errors. Only used if | |
| `content` is text. | |
| Returns: | |
| A _TempFile representing the created file. | |
| """ | |
| tf, _ = _TempFile._create(self._path, file_path, content, mode, encoding, | |
| errors) | |
| return tf | |
| def mkdir(self, dir_path=None): | |
| # type: (Optional[Text]) -> _TempDir | |
| """Create a directory in the directory. | |
| Args: | |
| dir_path: Optional path to the directory to create. If not given, | |
| a unique name will be generated and used. | |
| Returns: | |
| A _TempDir representing the created directory. | |
| """ | |
| if dir_path: | |
| path = os.path.join(self._path, dir_path) | |
| else: | |
| path = tempfile.mkdtemp(dir=self._path) | |
| # Note: there's no need to clear the directory since the containing | |
| # dir was cleared by the tempdir() function. | |
| os.makedirs(path, exist_ok=True) | |
| return _TempDir(path) | |
| class _TempFile(object): | |
| """Represents a tempfile for tests. | |
| Creation of this class is internal. Using its public methods is OK. | |
| This class implements the `os.PathLike` interface (specifically, | |
| `os.PathLike[str]`). This means, in Python 3, it can be directly passed | |
| to e.g. `os.path.join()`. | |
| """ | |
| def __init__(self, path): | |
| # type: (Text) -> None | |
| """Private: use _create instead.""" | |
| self._path = path | |
| # pylint: disable=line-too-long | |
| def _create(cls, base_path, file_path, content, mode, encoding, errors): | |
| # type: (Text, Optional[Text], AnyStr, Text, Text, Text) -> Tuple[_TempFile, Text] | |
| # pylint: enable=line-too-long | |
| """Module-private: create a tempfile instance.""" | |
| if file_path: | |
| cleanup_path = os.path.join(base_path, _get_first_part(file_path)) | |
| path = os.path.join(base_path, file_path) | |
| os.makedirs(os.path.dirname(path), exist_ok=True) | |
| # The file may already exist, in which case, ensure it's writable so that | |
| # it can be truncated. | |
| if os.path.exists(path) and not os.access(path, os.W_OK): | |
| stat_info = os.stat(path) | |
| os.chmod(path, stat_info.st_mode | stat.S_IWUSR) | |
| else: | |
| os.makedirs(base_path, exist_ok=True) | |
| fd, path = tempfile.mkstemp(dir=str(base_path)) | |
| os.close(fd) | |
| cleanup_path = path | |
| tf = cls(path) | |
| if content: | |
| if isinstance(content, str): | |
| tf.write_text(content, mode=mode, encoding=encoding, errors=errors) | |
| else: | |
| tf.write_bytes(content, mode) | |
| else: | |
| tf.write_bytes(b'') | |
| return tf, cleanup_path | |
| def full_path(self): | |
| # type: () -> Text | |
| """Returns the path, as a string, for the file. | |
| TIP: Instead of e.g. `os.path.join(temp_file.full_path)`, you can simply | |
| do `os.path.join(temp_file)` because `__fspath__()` is implemented. | |
| """ | |
| return self._path | |
| def __fspath__(self): | |
| # type: () -> Text | |
| """See os.PathLike.""" | |
| return self.full_path | |
| def read_text(self, encoding='utf8', errors='strict'): | |
| # type: (Text, Text) -> Text | |
| """Return the contents of the file as text.""" | |
| with self.open_text(encoding=encoding, errors=errors) as fp: | |
| return fp.read() | |
| def read_bytes(self): | |
| # type: () -> bytes | |
| """Return the content of the file as bytes.""" | |
| with self.open_bytes() as fp: | |
| return fp.read() | |
| def write_text(self, text, mode='w', encoding='utf8', errors='strict'): | |
| # type: (Text, Text, Text, Text) -> None | |
| """Write text to the file. | |
| Args: | |
| text: Text to write. In Python 2, it can be bytes, which will be | |
| decoded using the `encoding` arg (this is as an aid for code that | |
| is 2 and 3 compatible). | |
| mode: The mode to open the file for writing. | |
| encoding: The encoding to use when writing the text to the file. | |
| errors: The error handling strategy to use when converting text to bytes. | |
| """ | |
| with self.open_text(mode, encoding=encoding, errors=errors) as fp: | |
| fp.write(text) | |
| def write_bytes(self, data, mode='wb'): | |
| # type: (bytes, Text) -> None | |
| """Write bytes to the file. | |
| Args: | |
| data: bytes to write. | |
| mode: Mode to open the file for writing. The "b" flag is implicit if | |
| not already present. It must not have the "t" flag. | |
| """ | |
| with self.open_bytes(mode) as fp: | |
| fp.write(data) | |
| def open_text(self, mode='rt', encoding='utf8', errors='strict'): | |
| # type: (Text, Text, Text) -> ContextManager[TextIO] | |
| """Return a context manager for opening the file in text mode. | |
| Args: | |
| mode: The mode to open the file in. The "t" flag is implicit if not | |
| already present. It must not have the "b" flag. | |
| encoding: The encoding to use when opening the file. | |
| errors: How to handle decoding errors. | |
| Returns: | |
| Context manager that yields an open file. | |
| Raises: | |
| ValueError: if invalid inputs are provided. | |
| """ | |
| if 'b' in mode: | |
| raise ValueError('Invalid mode {!r}: "b" flag not allowed when opening ' | |
| 'file in text mode'.format(mode)) | |
| if 't' not in mode: | |
| mode += 't' | |
| cm = self._open(mode, encoding, errors) | |
| return cm | |
| def open_bytes(self, mode='rb'): | |
| # type: (Text) -> ContextManager[BinaryIO] | |
| """Return a context manager for opening the file in binary mode. | |
| Args: | |
| mode: The mode to open the file in. The "b" mode is implicit if not | |
| already present. It must not have the "t" flag. | |
| Returns: | |
| Context manager that yields an open file. | |
| Raises: | |
| ValueError: if invalid inputs are provided. | |
| """ | |
| if 't' in mode: | |
| raise ValueError('Invalid mode {!r}: "t" flag not allowed when opening ' | |
| 'file in binary mode'.format(mode)) | |
| if 'b' not in mode: | |
| mode += 'b' | |
| cm = self._open(mode, encoding=None, errors=None) | |
| return cm | |
| # TODO(b/123775699): Once pytype supports typing.Literal, use overload and | |
| # Literal to express more precise return types. The contained type is | |
| # currently `Any` to avoid [bad-return-type] errors in the open_* methods. | |
| def _open( | |
| self, | |
| mode: str, | |
| encoding: Optional[str] = 'utf8', | |
| errors: Optional[str] = 'strict', | |
| ) -> Iterator[Any]: | |
| with io.open( | |
| self.full_path, mode=mode, encoding=encoding, errors=errors) as fp: | |
| yield fp | |
| class _method(object): | |
| """A decorator that supports both instance and classmethod invocations. | |
| Using similar semantics to the @property builtin, this decorator can augment | |
| an instance method to support conditional logic when invoked on a class | |
| object. This breaks support for invoking an instance method via the class | |
| (e.g. Cls.method(self, ...)) but is still situationally useful. | |
| """ | |
| def __init__(self, finstancemethod): | |
| # type: (Callable[..., Any]) -> None | |
| self._finstancemethod = finstancemethod | |
| self._fclassmethod = None | |
| def classmethod(self, fclassmethod): | |
| # type: (Callable[..., Any]) -> _method | |
| self._fclassmethod = classmethod(fclassmethod) | |
| return self | |
| def __doc__(self): | |
| # type: () -> str | |
| if getattr(self._finstancemethod, '__doc__'): | |
| return self._finstancemethod.__doc__ | |
| elif getattr(self._fclassmethod, '__doc__'): | |
| return self._fclassmethod.__doc__ | |
| return '' | |
| def __get__(self, obj, type_): | |
| # type: (Optional[Any], Optional[Type[Any]]) -> Callable[..., Any] | |
| func = self._fclassmethod if obj is None else self._finstancemethod | |
| return func.__get__(obj, type_) # pytype: disable=attribute-error | |
| class TestCase(unittest.TestCase): | |
| """Extension of unittest.TestCase providing more power.""" | |
| # When to cleanup files/directories created by our `create_tempfile()` and | |
| # `create_tempdir()` methods after each test case completes. This does *not* | |
| # affect e.g., files created outside of those methods, e.g., using the stdlib | |
| # tempfile module. This can be overridden at the class level, instance level, | |
| # or with the `cleanup` arg of `create_tempfile()` and `create_tempdir()`. See | |
| # `TempFileCleanup` for details on the different values. | |
| # TODO(b/70517332): Remove the type comment and the disable once pytype has | |
| # better support for enums. | |
| tempfile_cleanup = TempFileCleanup.ALWAYS # type: TempFileCleanup # pytype: disable=annotation-type-mismatch | |
| maxDiff = 80 * 20 | |
| longMessage = True | |
| # Exit stacks for per-test and per-class scopes. | |
| if sys.version_info < (3, 11): | |
| _exit_stack = None | |
| _cls_exit_stack = None | |
| def __init__(self, *args, **kwargs): | |
| super(TestCase, self).__init__(*args, **kwargs) | |
| # This is to work around missing type stubs in unittest.pyi | |
| self._outcome = getattr(self, '_outcome') # type: Optional[_OutcomeType] | |
| def setUp(self): | |
| super(TestCase, self).setUp() | |
| # NOTE: Only Python 3 contextlib has ExitStack and | |
| # Python 3.11+ already has enterContext. | |
| if hasattr(contextlib, 'ExitStack') and sys.version_info < (3, 11): | |
| self._exit_stack = contextlib.ExitStack() | |
| self.addCleanup(self._exit_stack.close) | |
| def setUpClass(cls): | |
| super(TestCase, cls).setUpClass() | |
| # NOTE: Only Python 3 contextlib has ExitStack, only Python 3.8+ has | |
| # addClassCleanup and Python 3.11+ already has enterClassContext. | |
| if ( | |
| hasattr(contextlib, 'ExitStack') | |
| and hasattr(cls, 'addClassCleanup') | |
| and sys.version_info < (3, 11) | |
| ): | |
| cls._cls_exit_stack = contextlib.ExitStack() | |
| cls.addClassCleanup(cls._cls_exit_stack.close) | |
| def create_tempdir(self, name=None, cleanup=None): | |
| # type: (Optional[Text], Optional[TempFileCleanup]) -> _TempDir | |
| """Create a temporary directory specific to the test. | |
| NOTE: The directory and its contents will be recursively cleared before | |
| creation. This ensures that there is no pre-existing state. | |
| This creates a named directory on disk that is isolated to this test, and | |
| will be properly cleaned up by the test. This avoids several pitfalls of | |
| creating temporary directories for test purposes, as well as makes it easier | |
| to setup directories and verify their contents. For example:: | |
| def test_foo(self): | |
| out_dir = self.create_tempdir() | |
| out_log = out_dir.create_file('output.log') | |
| expected_outputs = [ | |
| os.path.join(out_dir, 'data-0.txt'), | |
| os.path.join(out_dir, 'data-1.txt'), | |
| ] | |
| code_under_test(out_dir) | |
| self.assertTrue(os.path.exists(expected_paths[0])) | |
| self.assertTrue(os.path.exists(expected_paths[1])) | |
| self.assertEqual('foo', out_log.read_text()) | |
| See also: :meth:`create_tempfile` for creating temporary files. | |
| Args: | |
| name: Optional name of the directory. If not given, a unique | |
| name will be generated and used. | |
| cleanup: Optional cleanup policy on when/if to remove the directory (and | |
| all its contents) at the end of the test. If None, then uses | |
| :attr:`tempfile_cleanup`. | |
| Returns: | |
| A _TempDir representing the created directory; see _TempDir class docs | |
| for usage. | |
| """ | |
| test_path = self._get_tempdir_path_test() | |
| if name: | |
| path = os.path.join(test_path, name) | |
| cleanup_path = os.path.join(test_path, _get_first_part(name)) | |
| else: | |
| os.makedirs(test_path, exist_ok=True) | |
| path = tempfile.mkdtemp(dir=test_path) | |
| cleanup_path = path | |
| _rmtree_ignore_errors(cleanup_path) | |
| os.makedirs(path, exist_ok=True) | |
| self._maybe_add_temp_path_cleanup(cleanup_path, cleanup) | |
| return _TempDir(path) | |
| # pylint: disable=line-too-long | |
| def create_tempfile(self, file_path=None, content=None, mode='w', | |
| encoding='utf8', errors='strict', cleanup=None): | |
| # type: (Optional[Text], Optional[AnyStr], Text, Text, Text, Optional[TempFileCleanup]) -> _TempFile | |
| # pylint: enable=line-too-long | |
| """Create a temporary file specific to the test. | |
| This creates a named file on disk that is isolated to this test, and will | |
| be properly cleaned up by the test. This avoids several pitfalls of | |
| creating temporary files for test purposes, as well as makes it easier | |
| to setup files, their data, read them back, and inspect them when | |
| a test fails. For example:: | |
| def test_foo(self): | |
| output = self.create_tempfile() | |
| code_under_test(output) | |
| self.assertGreater(os.path.getsize(output), 0) | |
| self.assertEqual('foo', output.read_text()) | |
| NOTE: This will zero-out the file. This ensures there is no pre-existing | |
| state. | |
| NOTE: If the file already exists, it will be made writable and overwritten. | |
| See also: :meth:`create_tempdir` for creating temporary directories, and | |
| ``_TempDir.create_file`` for creating files within a temporary directory. | |
| Args: | |
| file_path: Optional file path for the temp file. If not given, a unique | |
| file name will be generated and used. Slashes are allowed in the name; | |
| any missing intermediate directories will be created. NOTE: This path is | |
| the path that will be cleaned up, including any directories in the path, | |
| e.g., ``'foo/bar/baz.txt'`` will ``rm -r foo``. | |
| content: Optional string or | |
| bytes to initially write to the file. If not | |
| specified, then an empty file is created. | |
| mode: Mode string to use when writing content. Only used if `content` is | |
| non-empty. | |
| encoding: Encoding to use when writing string content. Only used if | |
| `content` is text. | |
| errors: How to handle text to bytes encoding errors. Only used if | |
| `content` is text. | |
| cleanup: Optional cleanup policy on when/if to remove the directory (and | |
| all its contents) at the end of the test. If None, then uses | |
| :attr:`tempfile_cleanup`. | |
| Returns: | |
| A _TempFile representing the created file; see _TempFile class docs for | |
| usage. | |
| """ | |
| test_path = self._get_tempdir_path_test() | |
| tf, cleanup_path = _TempFile._create(test_path, file_path, content=content, | |
| mode=mode, encoding=encoding, | |
| errors=errors) | |
| self._maybe_add_temp_path_cleanup(cleanup_path, cleanup) | |
| return tf | |
| def enter_context(self, manager): | |
| # type: (ContextManager[_T]) -> _T | |
| """Returns the CM's value after registering it with the exit stack. | |
| Entering a context pushes it onto a stack of contexts. When `enter_context` | |
| is called on the test instance (e.g. `self.enter_context`), the context is | |
| exited after the test case's tearDown call. When called on the test class | |
| (e.g. `TestCase.enter_context`), the context is exited after the test | |
| class's tearDownClass call. | |
| Contexts are exited in the reverse order of entering. They will always | |
| be exited, regardless of test failure/success. | |
| This is useful to eliminate per-test boilerplate when context managers | |
| are used. For example, instead of decorating every test with `@mock.patch`, | |
| simply do `self.foo = self.enter_context(mock.patch(...))' in `setUp()`. | |
| NOTE: The context managers will always be exited without any error | |
| information. This is an unfortunate implementation detail due to some | |
| internals of how unittest runs tests. | |
| Args: | |
| manager: The context manager to enter. | |
| """ | |
| if sys.version_info >= (3, 11): | |
| return self.enterContext(manager) | |
| if not self._exit_stack: | |
| raise AssertionError( | |
| 'self._exit_stack is not set: enter_context is Py3-only; also make ' | |
| 'sure that AbslTest.setUp() is called.') | |
| return self._exit_stack.enter_context(manager) | |
| def enter_context(cls, manager): # pylint: disable=no-self-argument | |
| # type: (ContextManager[_T]) -> _T | |
| if sys.version_info >= (3, 11): | |
| return cls.enterClassContext(manager) | |
| if not cls._cls_exit_stack: | |
| raise AssertionError( | |
| 'cls._cls_exit_stack is not set: cls.enter_context requires ' | |
| 'Python 3.8+; also make sure that AbslTest.setUpClass() is called.') | |
| return cls._cls_exit_stack.enter_context(manager) | |
| def _get_tempdir_path_cls(cls): | |
| # type: () -> Text | |
| return os.path.join(TEST_TMPDIR.value, | |
| cls.__qualname__.replace('__main__.', '')) | |
| def _get_tempdir_path_test(self): | |
| # type: () -> Text | |
| return os.path.join(self._get_tempdir_path_cls(), self._testMethodName) | |
| def _get_tempfile_cleanup(self, override): | |
| # type: (Optional[TempFileCleanup]) -> TempFileCleanup | |
| if override is not None: | |
| return override | |
| return self.tempfile_cleanup | |
| def _maybe_add_temp_path_cleanup(self, path, cleanup): | |
| # type: (Text, Optional[TempFileCleanup]) -> None | |
| cleanup = self._get_tempfile_cleanup(cleanup) | |
| if cleanup == TempFileCleanup.OFF: | |
| return | |
| elif cleanup == TempFileCleanup.ALWAYS: | |
| self.addCleanup(_rmtree_ignore_errors, path) | |
| elif cleanup == TempFileCleanup.SUCCESS: | |
| self._internal_add_cleanup_on_success(_rmtree_ignore_errors, path) | |
| else: | |
| raise AssertionError('Unexpected cleanup value: {}'.format(cleanup)) | |
| def _internal_add_cleanup_on_success( | |
| self, | |
| function: Callable[..., Any], | |
| *args: Any, | |
| **kwargs: Any, | |
| ) -> None: | |
| """Adds `function` as cleanup when the test case succeeds.""" | |
| outcome = self._outcome | |
| assert outcome is not None | |
| previous_failure_count = ( | |
| len(outcome.result.failures) | |
| + len(outcome.result.errors) | |
| + len(outcome.result.unexpectedSuccesses) | |
| ) | |
| def _call_cleaner_on_success(*args, **kwargs): | |
| if not self._internal_ran_and_passed_when_called_during_cleanup( | |
| previous_failure_count): | |
| return | |
| function(*args, **kwargs) | |
| self.addCleanup(_call_cleaner_on_success, *args, **kwargs) | |
| def _internal_ran_and_passed_when_called_during_cleanup( | |
| self, | |
| previous_failure_count: int, | |
| ) -> bool: | |
| """Returns whether test is passed. Expected to be called during cleanup.""" | |
| outcome = self._outcome | |
| if sys.version_info[:2] >= (3, 11): | |
| assert outcome is not None | |
| current_failure_count = ( | |
| len(outcome.result.failures) | |
| + len(outcome.result.errors) | |
| + len(outcome.result.unexpectedSuccesses) | |
| ) | |
| return current_failure_count == previous_failure_count | |
| else: | |
| # Before Python 3.11 https://github.com/python/cpython/pull/28180, errors | |
| # were bufferred in _Outcome before calling cleanup. | |
| result = self.defaultTestResult() | |
| self._feedErrorsToResult(result, outcome.errors) # pytype: disable=attribute-error | |
| return result.wasSuccessful() | |
| def shortDescription(self): | |
| # type: () -> Text | |
| """Formats both the test method name and the first line of its docstring. | |
| If no docstring is given, only returns the method name. | |
| This method overrides unittest.TestCase.shortDescription(), which | |
| only returns the first line of the docstring, obscuring the name | |
| of the test upon failure. | |
| Returns: | |
| desc: A short description of a test method. | |
| """ | |
| desc = self.id() | |
| # Omit the main name so that test name can be directly copy/pasted to | |
| # the command line. | |
| if desc.startswith('__main__.'): | |
| desc = desc[len('__main__.'):] | |
| # NOTE: super() is used here instead of directly invoking | |
| # unittest.TestCase.shortDescription(self), because of the | |
| # following line that occurs later on: | |
| # unittest.TestCase = TestCase | |
| # Because of this, direct invocation of what we think is the | |
| # superclass will actually cause infinite recursion. | |
| doc_first_line = super(TestCase, self).shortDescription() | |
| if doc_first_line is not None: | |
| desc = '\n'.join((desc, doc_first_line)) | |
| return desc | |
| def assertStartsWith(self, actual, expected_start, msg=None): | |
| """Asserts that actual.startswith(expected_start) is True. | |
| Args: | |
| actual: str | |
| expected_start: str | |
| msg: Optional message to report on failure. | |
| """ | |
| if not actual.startswith(expected_start): | |
| self.fail('%r does not start with %r' % (actual, expected_start), msg) | |
| def assertNotStartsWith(self, actual, unexpected_start, msg=None): | |
| """Asserts that actual.startswith(unexpected_start) is False. | |
| Args: | |
| actual: str | |
| unexpected_start: str | |
| msg: Optional message to report on failure. | |
| """ | |
| if actual.startswith(unexpected_start): | |
| self.fail('%r does start with %r' % (actual, unexpected_start), msg) | |
| def assertEndsWith(self, actual, expected_end, msg=None): | |
| """Asserts that actual.endswith(expected_end) is True. | |
| Args: | |
| actual: str | |
| expected_end: str | |
| msg: Optional message to report on failure. | |
| """ | |
| if not actual.endswith(expected_end): | |
| self.fail('%r does not end with %r' % (actual, expected_end), msg) | |
| def assertNotEndsWith(self, actual, unexpected_end, msg=None): | |
| """Asserts that actual.endswith(unexpected_end) is False. | |
| Args: | |
| actual: str | |
| unexpected_end: str | |
| msg: Optional message to report on failure. | |
| """ | |
| if actual.endswith(unexpected_end): | |
| self.fail('%r does end with %r' % (actual, unexpected_end), msg) | |
| def assertSequenceStartsWith(self, prefix, whole, msg=None): | |
| """An equality assertion for the beginning of ordered sequences. | |
| If prefix is an empty sequence, it will raise an error unless whole is also | |
| an empty sequence. | |
| If prefix is not a sequence, it will raise an error if the first element of | |
| whole does not match. | |
| Args: | |
| prefix: A sequence expected at the beginning of the whole parameter. | |
| whole: The sequence in which to look for prefix. | |
| msg: Optional message to report on failure. | |
| """ | |
| try: | |
| prefix_len = len(prefix) | |
| except (TypeError, NotImplementedError): | |
| prefix = [prefix] | |
| prefix_len = 1 | |
| if isinstance(whole, abc.Mapping) or isinstance(whole, abc.Set): | |
| self.fail( | |
| 'For whole: Mapping or Set objects are not supported, found type: %s' | |
| % type(whole), | |
| msg, | |
| ) | |
| try: | |
| whole_len = len(whole) | |
| except (TypeError, NotImplementedError): | |
| self.fail('For whole: len(%s) is not supported, it appears to be type: ' | |
| '%s' % (whole, type(whole)), msg) | |
| assert prefix_len <= whole_len, self._formatMessage( | |
| msg, | |
| 'Prefix length (%d) is longer than whole length (%d).' % | |
| (prefix_len, whole_len) | |
| ) | |
| if not prefix_len and whole_len: | |
| self.fail('Prefix length is 0 but whole length is %d: %s' % | |
| (len(whole), whole), msg) | |
| try: | |
| self.assertSequenceEqual(prefix, whole[:prefix_len], msg) | |
| except AssertionError: | |
| self.fail('prefix: %s not found at start of whole: %s.' % | |
| (prefix, whole), msg) | |
| def assertEmpty(self, container, msg=None): | |
| """Asserts that an object has zero length. | |
| Args: | |
| container: Anything that implements the collections.abc.Sized interface. | |
| msg: Optional message to report on failure. | |
| """ | |
| if not isinstance(container, abc.Sized): | |
| self.fail('Expected a Sized object, got: ' | |
| '{!r}'.format(type(container).__name__), msg) | |
| # explicitly check the length since some Sized objects (e.g. numpy.ndarray) | |
| # have strange __nonzero__/__bool__ behavior. | |
| if len(container): # pylint: disable=g-explicit-length-test | |
| self.fail('{!r} has length of {}.'.format(container, len(container)), msg) | |
| def assertNotEmpty(self, container, msg=None): | |
| """Asserts that an object has non-zero length. | |
| Args: | |
| container: Anything that implements the collections.abc.Sized interface. | |
| msg: Optional message to report on failure. | |
| """ | |
| if not isinstance(container, abc.Sized): | |
| self.fail('Expected a Sized object, got: ' | |
| '{!r}'.format(type(container).__name__), msg) | |
| # explicitly check the length since some Sized objects (e.g. numpy.ndarray) | |
| # have strange __nonzero__/__bool__ behavior. | |
| if not len(container): # pylint: disable=g-explicit-length-test | |
| self.fail('{!r} has length of 0.'.format(container), msg) | |
| def assertLen(self, container, expected_len, msg=None): | |
| """Asserts that an object has the expected length. | |
| Args: | |
| container: Anything that implements the collections.abc.Sized interface. | |
| expected_len: The expected length of the container. | |
| msg: Optional message to report on failure. | |
| """ | |
| if not isinstance(container, abc.Sized): | |
| self.fail('Expected a Sized object, got: ' | |
| '{!r}'.format(type(container).__name__), msg) | |
| if len(container) != expected_len: | |
| container_repr = unittest.util.safe_repr(container) # pytype: disable=module-attr | |
| self.fail('{} has length of {}, expected {}.'.format( | |
| container_repr, len(container), expected_len), msg) | |
| def assertSequenceAlmostEqual(self, expected_seq, actual_seq, places=None, | |
| msg=None, delta=None): | |
| """An approximate equality assertion for ordered sequences. | |
| Fail if the two sequences are unequal as determined by their value | |
| differences rounded to the given number of decimal places (default 7) and | |
| comparing to zero, or by comparing that the difference between each value | |
| in the two sequences is more than the given delta. | |
| Note that decimal places (from zero) are usually not the same as significant | |
| digits (measured from the most significant digit). | |
| If the two sequences compare equal then they will automatically compare | |
| almost equal. | |
| Args: | |
| expected_seq: A sequence containing elements we are expecting. | |
| actual_seq: The sequence that we are testing. | |
| places: The number of decimal places to compare. | |
| msg: The message to be printed if the test fails. | |
| delta: The OK difference between compared values. | |
| """ | |
| if len(expected_seq) != len(actual_seq): | |
| self.fail('Sequence size mismatch: {} vs {}'.format( | |
| len(expected_seq), len(actual_seq)), msg) | |
| err_list = [] | |
| for idx, (exp_elem, act_elem) in enumerate(zip(expected_seq, actual_seq)): | |
| try: | |
| # assertAlmostEqual should be called with at most one of `places` and | |
| # `delta`. However, it's okay for assertSequenceAlmostEqual to pass | |
| # both because we want the latter to fail if the former does. | |
| # pytype: disable=wrong-keyword-args | |
| self.assertAlmostEqual(exp_elem, act_elem, places=places, msg=msg, | |
| delta=delta) | |
| # pytype: enable=wrong-keyword-args | |
| except self.failureException as err: | |
| err_list.append('At index {}: {}'.format(idx, err)) | |
| if err_list: | |
| if len(err_list) > 30: | |
| err_list = err_list[:30] + ['...'] | |
| msg = self._formatMessage(msg, '\n'.join(err_list)) | |
| self.fail(msg) | |
| def assertContainsSubset(self, expected_subset, actual_set, msg=None): | |
| """Checks whether actual iterable is a superset of expected iterable.""" | |
| missing = set(expected_subset) - set(actual_set) | |
| if not missing: | |
| return | |
| self.fail('Missing elements %s\nExpected: %s\nActual: %s' % ( | |
| missing, expected_subset, actual_set), msg) | |
| def assertNoCommonElements(self, expected_seq, actual_seq, msg=None): | |
| """Checks whether actual iterable and expected iterable are disjoint.""" | |
| common = set(expected_seq) & set(actual_seq) | |
| if not common: | |
| return | |
| self.fail('Common elements %s\nExpected: %s\nActual: %s' % ( | |
| common, expected_seq, actual_seq), msg) | |
| def assertItemsEqual(self, expected_seq, actual_seq, msg=None): | |
| """Deprecated, please use assertCountEqual instead. | |
| This is equivalent to assertCountEqual. | |
| Args: | |
| expected_seq: A sequence containing elements we are expecting. | |
| actual_seq: The sequence that we are testing. | |
| msg: The message to be printed if the test fails. | |
| """ | |
| super().assertCountEqual(expected_seq, actual_seq, msg) | |
| def assertSameElements(self, expected_seq, actual_seq, msg=None): | |
| """Asserts that two sequences have the same elements (in any order). | |
| This method, unlike assertCountEqual, doesn't care about any | |
| duplicates in the expected and actual sequences:: | |
| # Doesn't raise an AssertionError | |
| assertSameElements([1, 1, 1, 0, 0, 0], [0, 1]) | |
| If possible, you should use assertCountEqual instead of | |
| assertSameElements. | |
| Args: | |
| expected_seq: A sequence containing elements we are expecting. | |
| actual_seq: The sequence that we are testing. | |
| msg: The message to be printed if the test fails. | |
| """ | |
| # `unittest2.TestCase` used to have assertSameElements, but it was | |
| # removed in favor of assertItemsEqual. As there's a unit test | |
| # that explicitly checks this behavior, I am leaving this method | |
| # alone. | |
| # Fail on strings: empirically, passing strings to this test method | |
| # is almost always a bug. If comparing the character sets of two strings | |
| # is desired, cast the inputs to sets or lists explicitly. | |
| if (isinstance(expected_seq, _TEXT_OR_BINARY_TYPES) or | |
| isinstance(actual_seq, _TEXT_OR_BINARY_TYPES)): | |
| self.fail('Passing string/bytes to assertSameElements is usually a bug. ' | |
| 'Did you mean to use assertEqual?\n' | |
| 'Expected: %s\nActual: %s' % (expected_seq, actual_seq)) | |
| try: | |
| expected = dict([(element, None) for element in expected_seq]) | |
| actual = dict([(element, None) for element in actual_seq]) | |
| missing = [element for element in expected if element not in actual] | |
| unexpected = [element for element in actual if element not in expected] | |
| missing.sort() | |
| unexpected.sort() | |
| except TypeError: | |
| # Fall back to slower list-compare if any of the objects are | |
| # not hashable. | |
| expected = list(expected_seq) | |
| actual = list(actual_seq) | |
| expected.sort() | |
| actual.sort() | |
| missing, unexpected = _sorted_list_difference(expected, actual) | |
| errors = [] | |
| if msg: | |
| errors.extend((msg, ':\n')) | |
| if missing: | |
| errors.append('Expected, but missing:\n %r\n' % missing) | |
| if unexpected: | |
| errors.append('Unexpected, but present:\n %r\n' % unexpected) | |
| if missing or unexpected: | |
| self.fail(''.join(errors)) | |
| # unittest.TestCase.assertMultiLineEqual works very similarly, but it | |
| # has a different error format. However, I find this slightly more readable. | |
| def assertMultiLineEqual(self, first, second, msg=None, **kwargs): | |
| """Asserts that two multi-line strings are equal.""" | |
| assert isinstance(first, | |
| str), ('First argument is not a string: %r' % (first,)) | |
| assert isinstance(second, | |
| str), ('Second argument is not a string: %r' % (second,)) | |
| line_limit = kwargs.pop('line_limit', 0) | |
| if kwargs: | |
| raise TypeError('Unexpected keyword args {}'.format(tuple(kwargs))) | |
| if first == second: | |
| return | |
| if msg: | |
| failure_message = [msg + ':\n'] | |
| else: | |
| failure_message = ['\n'] | |
| if line_limit: | |
| line_limit += len(failure_message) | |
| for line in difflib.ndiff(first.splitlines(True), second.splitlines(True)): | |
| failure_message.append(line) | |
| if not line.endswith('\n'): | |
| failure_message.append('\n') | |
| if line_limit and len(failure_message) > line_limit: | |
| n_omitted = len(failure_message) - line_limit | |
| failure_message = failure_message[:line_limit] | |
| failure_message.append( | |
| '(... and {} more delta lines omitted for brevity.)\n'.format( | |
| n_omitted)) | |
| raise self.failureException(''.join(failure_message)) | |
| def assertBetween(self, value, minv, maxv, msg=None): | |
| """Asserts that value is between minv and maxv (inclusive).""" | |
| msg = self._formatMessage(msg, | |
| '"%r" unexpectedly not between "%r" and "%r"' % | |
| (value, minv, maxv)) | |
| self.assertTrue(minv <= value, msg) | |
| self.assertTrue(maxv >= value, msg) | |
| def assertRegexMatch(self, actual_str, regexes, message=None): | |
| r"""Asserts that at least one regex in regexes matches str. | |
| If possible you should use `assertRegex`, which is a simpler | |
| version of this method. `assertRegex` takes a single regular | |
| expression (a string or re compiled object) instead of a list. | |
| Notes: | |
| 1. This function uses substring matching, i.e. the matching | |
| succeeds if *any* substring of the error message matches *any* | |
| regex in the list. This is more convenient for the user than | |
| full-string matching. | |
| 2. If regexes is the empty list, the matching will always fail. | |
| 3. Use regexes=[''] for a regex that will always pass. | |
| 4. '.' matches any single character *except* the newline. To | |
| match any character, use '(.|\n)'. | |
| 5. '^' matches the beginning of each line, not just the beginning | |
| of the string. Similarly, '$' matches the end of each line. | |
| 6. An exception will be thrown if regexes contains an invalid | |
| regex. | |
| Args: | |
| actual_str: The string we try to match with the items in regexes. | |
| regexes: The regular expressions we want to match against str. | |
| See "Notes" above for detailed notes on how this is interpreted. | |
| message: The message to be printed if the test fails. | |
| """ | |
| if isinstance(regexes, _TEXT_OR_BINARY_TYPES): | |
| self.fail('regexes is string or bytes; use assertRegex instead.', | |
| message) | |
| if not regexes: | |
| self.fail('No regexes specified.', message) | |
| regex_type = type(regexes[0]) | |
| for regex in regexes[1:]: | |
| if type(regex) is not regex_type: # pylint: disable=unidiomatic-typecheck | |
| self.fail('regexes list must all be the same type.', message) | |
| if regex_type is bytes and isinstance(actual_str, str): | |
| regexes = [regex.decode('utf-8') for regex in regexes] | |
| regex_type = str | |
| elif regex_type is str and isinstance(actual_str, bytes): | |
| regexes = [regex.encode('utf-8') for regex in regexes] | |
| regex_type = bytes | |
| if regex_type is str: | |
| regex = u'(?:%s)' % u')|(?:'.join(regexes) | |
| elif regex_type is bytes: | |
| regex = b'(?:' + (b')|(?:'.join(regexes)) + b')' | |
| else: | |
| self.fail('Only know how to deal with unicode str or bytes regexes.', | |
| message) | |
| if not re.search(regex, actual_str, re.MULTILINE): | |
| self.fail('"%s" does not contain any of these regexes: %s.' % | |
| (actual_str, regexes), message) | |
| def assertCommandSucceeds(self, command, regexes=(b'',), env=None, | |
| close_fds=True, msg=None): | |
| """Asserts that a shell command succeeds (i.e. exits with code 0). | |
| Args: | |
| command: List or string representing the command to run. | |
| regexes: List of regular expression byte strings that match success. | |
| env: Dictionary of environment variable settings. If None, no environment | |
| variables will be set for the child process. This is to make tests | |
| more hermetic. NOTE: this behavior is different than the standard | |
| subprocess module. | |
| close_fds: Whether or not to close all open fd's in the child after | |
| forking. | |
| msg: Optional message to report on failure. | |
| """ | |
| (ret_code, err) = get_command_stderr(command, env, close_fds) | |
| # We need bytes regexes here because `err` is bytes. | |
| # Accommodate code which listed their output regexes w/o the b'' prefix by | |
| # converting them to bytes for the user. | |
| if isinstance(regexes[0], str): | |
| regexes = [regex.encode('utf-8') for regex in regexes] | |
| command_string = get_command_string(command) | |
| self.assertEqual( | |
| ret_code, 0, | |
| self._formatMessage(msg, | |
| 'Running command\n' | |
| '%s failed with error code %s and message\n' | |
| '%s' % (_quote_long_string(command_string), | |
| ret_code, | |
| _quote_long_string(err))) | |
| ) | |
| self.assertRegexMatch( | |
| err, | |
| regexes, | |
| message=self._formatMessage( | |
| msg, | |
| 'Running command\n' | |
| '%s failed with error code %s and message\n' | |
| '%s which matches no regex in %s' % ( | |
| _quote_long_string(command_string), | |
| ret_code, | |
| _quote_long_string(err), | |
| regexes))) | |
| def assertCommandFails(self, command, regexes, env=None, close_fds=True, | |
| msg=None): | |
| """Asserts a shell command fails and the error matches a regex in a list. | |
| Args: | |
| command: List or string representing the command to run. | |
| regexes: the list of regular expression strings. | |
| env: Dictionary of environment variable settings. If None, no environment | |
| variables will be set for the child process. This is to make tests | |
| more hermetic. NOTE: this behavior is different than the standard | |
| subprocess module. | |
| close_fds: Whether or not to close all open fd's in the child after | |
| forking. | |
| msg: Optional message to report on failure. | |
| """ | |
| (ret_code, err) = get_command_stderr(command, env, close_fds) | |
| # We need bytes regexes here because `err` is bytes. | |
| # Accommodate code which listed their output regexes w/o the b'' prefix by | |
| # converting them to bytes for the user. | |
| if isinstance(regexes[0], str): | |
| regexes = [regex.encode('utf-8') for regex in regexes] | |
| command_string = get_command_string(command) | |
| self.assertNotEqual( | |
| ret_code, 0, | |
| self._formatMessage(msg, 'The following command succeeded ' | |
| 'while expected to fail:\n%s' % | |
| _quote_long_string(command_string))) | |
| self.assertRegexMatch( | |
| err, | |
| regexes, | |
| message=self._formatMessage( | |
| msg, | |
| 'Running command\n' | |
| '%s failed with error code %s and message\n' | |
| '%s which matches no regex in %s' % ( | |
| _quote_long_string(command_string), | |
| ret_code, | |
| _quote_long_string(err), | |
| regexes))) | |
| class _AssertRaisesContext(object): | |
| def __init__(self, expected_exception, test_case, test_func, msg=None): | |
| self.expected_exception = expected_exception | |
| self.test_case = test_case | |
| self.test_func = test_func | |
| self.msg = msg | |
| def __enter__(self): | |
| return self | |
| def __exit__(self, exc_type, exc_value, tb): | |
| if exc_type is None: | |
| self.test_case.fail(self.expected_exception.__name__ + ' not raised', | |
| self.msg) | |
| if not issubclass(exc_type, self.expected_exception): | |
| return False | |
| self.test_func(exc_value) | |
| if exc_value: | |
| self.exception = exc_value.with_traceback(None) | |
| return True | |
| def assertRaisesWithPredicateMatch( | |
| self, expected_exception, predicate) -> _AssertRaisesContext: | |
| # The purpose of this return statement is to work around | |
| # https://github.com/PyCQA/pylint/issues/5273; it is otherwise ignored. | |
| return self._AssertRaisesContext(None, None, None) | |
| def assertRaisesWithPredicateMatch( | |
| self, expected_exception, predicate, callable_obj: Callable[..., Any], | |
| *args, **kwargs) -> None: | |
| # The purpose of this return statement is to work around | |
| # https://github.com/PyCQA/pylint/issues/5273; it is otherwise ignored. | |
| return self._AssertRaisesContext(None, None, None) | |
| def assertRaisesWithPredicateMatch(self, expected_exception, predicate, | |
| callable_obj=None, *args, **kwargs): | |
| """Asserts that exception is thrown and predicate(exception) is true. | |
| Args: | |
| expected_exception: Exception class expected to be raised. | |
| predicate: Function of one argument that inspects the passed-in exception | |
| and returns True (success) or False (please fail the test). | |
| callable_obj: Function to be called. | |
| *args: Extra args. | |
| **kwargs: Extra keyword args. | |
| Returns: | |
| A context manager if callable_obj is None. Otherwise, None. | |
| Raises: | |
| self.failureException if callable_obj does not raise a matching exception. | |
| """ | |
| def Check(err): | |
| self.assertTrue(predicate(err), | |
| '%r does not match predicate %r' % (err, predicate)) | |
| context = self._AssertRaisesContext(expected_exception, self, Check) | |
| if callable_obj is None: | |
| return context | |
| with context: | |
| callable_obj(*args, **kwargs) | |
| def assertRaisesWithLiteralMatch( | |
| self, expected_exception, expected_exception_message | |
| ) -> _AssertRaisesContext: | |
| # The purpose of this return statement is to work around | |
| # https://github.com/PyCQA/pylint/issues/5273; it is otherwise ignored. | |
| return self._AssertRaisesContext(None, None, None) | |
| def assertRaisesWithLiteralMatch( | |
| self, expected_exception, expected_exception_message, | |
| callable_obj: Callable[..., Any], *args, **kwargs) -> None: | |
| # The purpose of this return statement is to work around | |
| # https://github.com/PyCQA/pylint/issues/5273; it is otherwise ignored. | |
| return self._AssertRaisesContext(None, None, None) | |
| def assertRaisesWithLiteralMatch(self, expected_exception, | |
| expected_exception_message, | |
| callable_obj=None, *args, **kwargs): | |
| """Asserts that the message in a raised exception equals the given string. | |
| Unlike assertRaisesRegex, this method takes a literal string, not | |
| a regular expression. | |
| with self.assertRaisesWithLiteralMatch(ExType, 'message'): | |
| DoSomething() | |
| Args: | |
| expected_exception: Exception class expected to be raised. | |
| expected_exception_message: String message expected in the raised | |
| exception. For a raise exception e, expected_exception_message must | |
| equal str(e). | |
| callable_obj: Function to be called, or None to return a context. | |
| *args: Extra args. | |
| **kwargs: Extra kwargs. | |
| Returns: | |
| A context manager if callable_obj is None. Otherwise, None. | |
| Raises: | |
| self.failureException if callable_obj does not raise a matching exception. | |
| """ | |
| def Check(err): | |
| actual_exception_message = str(err) | |
| self.assertTrue(expected_exception_message == actual_exception_message, | |
| 'Exception message does not match.\n' | |
| 'Expected: %r\n' | |
| 'Actual: %r' % (expected_exception_message, | |
| actual_exception_message)) | |
| context = self._AssertRaisesContext(expected_exception, self, Check) | |
| if callable_obj is None: | |
| return context | |
| with context: | |
| callable_obj(*args, **kwargs) | |
| def assertContainsInOrder(self, strings, target, msg=None): | |
| """Asserts that the strings provided are found in the target in order. | |
| This may be useful for checking HTML output. | |
| Args: | |
| strings: A list of strings, such as [ 'fox', 'dog' ] | |
| target: A target string in which to look for the strings, such as | |
| 'The quick brown fox jumped over the lazy dog'. | |
| msg: Optional message to report on failure. | |
| """ | |
| if isinstance(strings, (bytes, unicode if str is bytes else str)): | |
| strings = (strings,) | |
| current_index = 0 | |
| last_string = None | |
| for string in strings: | |
| index = target.find(str(string), current_index) | |
| if index == -1 and current_index == 0: | |
| self.fail("Did not find '%s' in '%s'" % | |
| (string, target), msg) | |
| elif index == -1: | |
| self.fail("Did not find '%s' after '%s' in '%s'" % | |
| (string, last_string, target), msg) | |
| last_string = string | |
| current_index = index | |
| def assertContainsSubsequence(self, container, subsequence, msg=None): | |
| """Asserts that "container" contains "subsequence" as a subsequence. | |
| Asserts that "container" contains all the elements of "subsequence", in | |
| order, but possibly with other elements interspersed. For example, [1, 2, 3] | |
| is a subsequence of [0, 0, 1, 2, 0, 3, 0] but not of [0, 0, 1, 3, 0, 2, 0]. | |
| Args: | |
| container: the list we're testing for subsequence inclusion. | |
| subsequence: the list we hope will be a subsequence of container. | |
| msg: Optional message to report on failure. | |
| """ | |
| first_nonmatching = None | |
| reversed_container = list(reversed(container)) | |
| subsequence = list(subsequence) | |
| for e in subsequence: | |
| if e not in reversed_container: | |
| first_nonmatching = e | |
| break | |
| while e != reversed_container.pop(): | |
| pass | |
| if first_nonmatching is not None: | |
| self.fail('%s not a subsequence of %s. First non-matching element: %s' % | |
| (subsequence, container, first_nonmatching), msg) | |
| def assertContainsExactSubsequence(self, container, subsequence, msg=None): | |
| """Asserts that "container" contains "subsequence" as an exact subsequence. | |
| Asserts that "container" contains all the elements of "subsequence", in | |
| order, and without other elements interspersed. For example, [1, 2, 3] is an | |
| exact subsequence of [0, 0, 1, 2, 3, 0] but not of [0, 0, 1, 2, 0, 3, 0]. | |
| Args: | |
| container: the list we're testing for subsequence inclusion. | |
| subsequence: the list we hope will be an exact subsequence of container. | |
| msg: Optional message to report on failure. | |
| """ | |
| container = list(container) | |
| subsequence = list(subsequence) | |
| longest_match = 0 | |
| for start in range(1 + len(container) - len(subsequence)): | |
| if longest_match == len(subsequence): | |
| break | |
| index = 0 | |
| while (index < len(subsequence) and | |
| subsequence[index] == container[start + index]): | |
| index += 1 | |
| longest_match = max(longest_match, index) | |
| if longest_match < len(subsequence): | |
| self.fail('%s not an exact subsequence of %s. ' | |
| 'Longest matching prefix: %s' % | |
| (subsequence, container, subsequence[:longest_match]), msg) | |
| def assertTotallyOrdered(self, *groups, **kwargs): | |
| """Asserts that total ordering has been implemented correctly. | |
| For example, say you have a class A that compares only on its attribute x. | |
| Comparators other than ``__lt__`` are omitted for brevity:: | |
| class A(object): | |
| def __init__(self, x, y): | |
| self.x = x | |
| self.y = y | |
| def __hash__(self): | |
| return hash(self.x) | |
| def __lt__(self, other): | |
| try: | |
| return self.x < other.x | |
| except AttributeError: | |
| return NotImplemented | |
| assertTotallyOrdered will check that instances can be ordered correctly. | |
| For example:: | |
| self.assertTotallyOrdered( | |
| [None], # None should come before everything else. | |
| [1], # Integers sort earlier. | |
| [A(1, 'a')], | |
| [A(2, 'b')], # 2 is after 1. | |
| [A(3, 'c'), A(3, 'd')], # The second argument is irrelevant. | |
| [A(4, 'z')], | |
| ['foo']) # Strings sort last. | |
| Args: | |
| *groups: A list of groups of elements. Each group of elements is a list | |
| of objects that are equal. The elements in each group must be less | |
| than the elements in the group after it. For example, these groups are | |
| totally ordered: ``[None]``, ``[1]``, ``[2, 2]``, ``[3]``. | |
| **kwargs: optional msg keyword argument can be passed. | |
| """ | |
| def CheckOrder(small, big): | |
| """Ensures small is ordered before big.""" | |
| self.assertFalse(small == big, | |
| self._formatMessage(msg, '%r unexpectedly equals %r' % | |
| (small, big))) | |
| self.assertTrue(small != big, | |
| self._formatMessage(msg, '%r unexpectedly equals %r' % | |
| (small, big))) | |
| self.assertLess(small, big, msg) | |
| self.assertFalse(big < small, | |
| self._formatMessage(msg, | |
| '%r unexpectedly less than %r' % | |
| (big, small))) | |
| self.assertLessEqual(small, big, msg) | |
| self.assertFalse(big <= small, self._formatMessage( | |
| '%r unexpectedly less than or equal to %r' % (big, small), msg | |
| )) | |
| self.assertGreater(big, small, msg) | |
| self.assertFalse(small > big, | |
| self._formatMessage(msg, | |
| '%r unexpectedly greater than %r' % | |
| (small, big))) | |
| self.assertGreaterEqual(big, small) | |
| self.assertFalse(small >= big, self._formatMessage( | |
| msg, | |
| '%r unexpectedly greater than or equal to %r' % (small, big))) | |
| def CheckEqual(a, b): | |
| """Ensures that a and b are equal.""" | |
| self.assertEqual(a, b, msg) | |
| self.assertFalse(a != b, | |
| self._formatMessage(msg, '%r unexpectedly unequals %r' % | |
| (a, b))) | |
| # Objects that compare equal must hash to the same value, but this only | |
| # applies if both objects are hashable. | |
| if (isinstance(a, abc.Hashable) and | |
| isinstance(b, abc.Hashable)): | |
| self.assertEqual( | |
| hash(a), hash(b), | |
| self._formatMessage( | |
| msg, 'hash %d of %r unexpectedly not equal to hash %d of %r' % | |
| (hash(a), a, hash(b), b))) | |
| self.assertFalse(a < b, | |
| self._formatMessage(msg, | |
| '%r unexpectedly less than %r' % | |
| (a, b))) | |
| self.assertFalse(b < a, | |
| self._formatMessage(msg, | |
| '%r unexpectedly less than %r' % | |
| (b, a))) | |
| self.assertLessEqual(a, b, msg) | |
| self.assertLessEqual(b, a, msg) # pylint: disable=arguments-out-of-order | |
| self.assertFalse(a > b, | |
| self._formatMessage(msg, | |
| '%r unexpectedly greater than %r' % | |
| (a, b))) | |
| self.assertFalse(b > a, | |
| self._formatMessage(msg, | |
| '%r unexpectedly greater than %r' % | |
| (b, a))) | |
| self.assertGreaterEqual(a, b, msg) | |
| self.assertGreaterEqual(b, a, msg) # pylint: disable=arguments-out-of-order | |
| msg = kwargs.get('msg') | |
| # For every combination of elements, check the order of every pair of | |
| # elements. | |
| for elements in itertools.product(*groups): | |
| elements = list(elements) | |
| for index, small in enumerate(elements[:-1]): | |
| for big in elements[index + 1:]: | |
| CheckOrder(small, big) | |
| # Check that every element in each group is equal. | |
| for group in groups: | |
| for a in group: | |
| CheckEqual(a, a) | |
| for a, b in itertools.product(group, group): | |
| CheckEqual(a, b) | |
| def assertDictEqual(self, a, b, msg=None): | |
| """Raises AssertionError if a and b are not equal dictionaries. | |
| Args: | |
| a: A dict, the expected value. | |
| b: A dict, the actual value. | |
| msg: An optional str, the associated message. | |
| Raises: | |
| AssertionError: if the dictionaries are not equal. | |
| """ | |
| self.assertIsInstance(a, dict, self._formatMessage( | |
| msg, | |
| 'First argument is not a dictionary' | |
| )) | |
| self.assertIsInstance(b, dict, self._formatMessage( | |
| msg, | |
| 'Second argument is not a dictionary' | |
| )) | |
| def Sorted(list_of_items): | |
| try: | |
| return sorted(list_of_items) # In 3.3, unordered are possible. | |
| except TypeError: | |
| return list_of_items | |
| if a == b: | |
| return | |
| a_items = Sorted(list(a.items())) | |
| b_items = Sorted(list(b.items())) | |
| unexpected = [] | |
| missing = [] | |
| different = [] | |
| safe_repr = unittest.util.safe_repr # pytype: disable=module-attr | |
| def Repr(dikt): | |
| """Deterministic repr for dict.""" | |
| # Sort the entries based on their repr, not based on their sort order, | |
| # which will be non-deterministic across executions, for many types. | |
| entries = sorted((safe_repr(k), safe_repr(v)) for k, v in dikt.items()) | |
| return '{%s}' % (', '.join('%s: %s' % pair for pair in entries)) | |
| message = ['%s != %s%s' % (Repr(a), Repr(b), ' (%s)' % msg if msg else '')] | |
| # The standard library default output confounds lexical difference with | |
| # value difference; treat them separately. | |
| for a_key, a_value in a_items: | |
| if a_key not in b: | |
| missing.append((a_key, a_value)) | |
| elif a_value != b[a_key]: | |
| different.append((a_key, a_value, b[a_key])) | |
| for b_key, b_value in b_items: | |
| if b_key not in a: | |
| unexpected.append((b_key, b_value)) | |
| if unexpected: | |
| message.append( | |
| 'Unexpected, but present entries:\n%s' % ''.join( | |
| '%s: %s\n' % (safe_repr(k), safe_repr(v)) for k, v in unexpected)) | |
| if different: | |
| message.append( | |
| 'repr() of differing entries:\n%s' % ''.join( | |
| '%s: %s != %s\n' % (safe_repr(k), safe_repr(a_value), | |
| safe_repr(b_value)) | |
| for k, a_value, b_value in different)) | |
| if missing: | |
| message.append( | |
| 'Missing entries:\n%s' % ''.join( | |
| ('%s: %s\n' % (safe_repr(k), safe_repr(v)) for k, v in missing))) | |
| raise self.failureException('\n'.join(message)) | |
| def assertUrlEqual(self, a, b, msg=None): | |
| """Asserts that urls are equal, ignoring ordering of query params.""" | |
| parsed_a = parse.urlparse(a) | |
| parsed_b = parse.urlparse(b) | |
| self.assertEqual(parsed_a.scheme, parsed_b.scheme, msg) | |
| self.assertEqual(parsed_a.netloc, parsed_b.netloc, msg) | |
| self.assertEqual(parsed_a.path, parsed_b.path, msg) | |
| self.assertEqual(parsed_a.fragment, parsed_b.fragment, msg) | |
| self.assertEqual(sorted(parsed_a.params.split(';')), | |
| sorted(parsed_b.params.split(';')), msg) | |
| self.assertDictEqual( | |
| parse.parse_qs(parsed_a.query, keep_blank_values=True), | |
| parse.parse_qs(parsed_b.query, keep_blank_values=True), msg) | |
| def assertSameStructure(self, a, b, aname='a', bname='b', msg=None): | |
| """Asserts that two values contain the same structural content. | |
| The two arguments should be data trees consisting of trees of dicts and | |
| lists. They will be deeply compared by walking into the contents of dicts | |
| and lists; other items will be compared using the == operator. | |
| If the two structures differ in content, the failure message will indicate | |
| the location within the structures where the first difference is found. | |
| This may be helpful when comparing large structures. | |
| Mixed Sequence and Set types are supported. Mixed Mapping types are | |
| supported, but the order of the keys will not be considered in the | |
| comparison. | |
| Args: | |
| a: The first structure to compare. | |
| b: The second structure to compare. | |
| aname: Variable name to use for the first structure in assertion messages. | |
| bname: Variable name to use for the second structure. | |
| msg: Additional text to include in the failure message. | |
| """ | |
| # Accumulate all the problems found so we can report all of them at once | |
| # rather than just stopping at the first | |
| problems = [] | |
| _walk_structure_for_problems(a, b, aname, bname, problems, | |
| self.assertEqual, self.failureException) | |
| # Avoid spamming the user toooo much | |
| if self.maxDiff is not None: | |
| max_problems_to_show = self.maxDiff // 80 | |
| if len(problems) > max_problems_to_show: | |
| problems = problems[0:max_problems_to_show-1] + ['...'] | |
| if problems: | |
| self.fail('; '.join(problems), msg) | |
| def assertJsonEqual(self, first, second, msg=None): | |
| """Asserts that the JSON objects defined in two strings are equal. | |
| A summary of the differences will be included in the failure message | |
| using assertSameStructure. | |
| Args: | |
| first: A string containing JSON to decode and compare to second. | |
| second: A string containing JSON to decode and compare to first. | |
| msg: Additional text to include in the failure message. | |
| """ | |
| try: | |
| first_structured = json.loads(first) | |
| except ValueError as e: | |
| raise ValueError(self._formatMessage( | |
| msg, | |
| 'could not decode first JSON value %s: %s' % (first, e))) | |
| try: | |
| second_structured = json.loads(second) | |
| except ValueError as e: | |
| raise ValueError(self._formatMessage( | |
| msg, | |
| 'could not decode second JSON value %s: %s' % (second, e))) | |
| self.assertSameStructure(first_structured, second_structured, | |
| aname='first', bname='second', msg=msg) | |
| def _getAssertEqualityFunc(self, first, second): | |
| # type: (Any, Any) -> Callable[..., None] | |
| try: | |
| return super(TestCase, self)._getAssertEqualityFunc(first, second) | |
| except AttributeError: | |
| # This is a workaround if unittest.TestCase.__init__ was never run. | |
| # It usually means that somebody created a subclass just for the | |
| # assertions and has overridden __init__. "assertTrue" is a safe | |
| # value that will not make __init__ raise a ValueError. | |
| test_method = getattr(self, '_testMethodName', 'assertTrue') | |
| super(TestCase, self).__init__(test_method) | |
| return super(TestCase, self)._getAssertEqualityFunc(first, second) | |
| def fail(self, msg=None, user_msg=None) -> NoReturn: | |
| """Fail immediately with the given standard message and user message.""" | |
| return super(TestCase, self).fail(self._formatMessage(user_msg, msg)) | |
| def _sorted_list_difference(expected, actual): | |
| # type: (List[_T], List[_T]) -> Tuple[List[_T], List[_T]] | |
| """Finds elements in only one or the other of two, sorted input lists. | |
| Returns a two-element tuple of lists. The first list contains those | |
| elements in the "expected" list but not in the "actual" list, and the | |
| second contains those elements in the "actual" list but not in the | |
| "expected" list. Duplicate elements in either input list are ignored. | |
| Args: | |
| expected: The list we expected. | |
| actual: The list we actually got. | |
| Returns: | |
| (missing, unexpected) | |
| missing: items in expected that are not in actual. | |
| unexpected: items in actual that are not in expected. | |
| """ | |
| i = j = 0 | |
| missing = [] | |
| unexpected = [] | |
| while True: | |
| try: | |
| e = expected[i] | |
| a = actual[j] | |
| if e < a: | |
| missing.append(e) | |
| i += 1 | |
| while expected[i] == e: | |
| i += 1 | |
| elif e > a: | |
| unexpected.append(a) | |
| j += 1 | |
| while actual[j] == a: | |
| j += 1 | |
| else: | |
| i += 1 | |
| try: | |
| while expected[i] == e: | |
| i += 1 | |
| finally: | |
| j += 1 | |
| while actual[j] == a: | |
| j += 1 | |
| except IndexError: | |
| missing.extend(expected[i:]) | |
| unexpected.extend(actual[j:]) | |
| break | |
| return missing, unexpected | |
| def _are_both_of_integer_type(a, b): | |
| # type: (object, object) -> bool | |
| return isinstance(a, int) and isinstance(b, int) | |
| def _are_both_of_sequence_type(a, b): | |
| # type: (object, object) -> bool | |
| return isinstance(a, abc.Sequence) and isinstance( | |
| b, abc.Sequence) and not isinstance( | |
| a, _TEXT_OR_BINARY_TYPES) and not isinstance(b, _TEXT_OR_BINARY_TYPES) | |
| def _are_both_of_set_type(a, b): | |
| # type: (object, object) -> bool | |
| return isinstance(a, abc.Set) and isinstance(b, abc.Set) | |
| def _are_both_of_mapping_type(a, b): | |
| # type: (object, object) -> bool | |
| return isinstance(a, abc.Mapping) and isinstance( | |
| b, abc.Mapping) | |
| def _walk_structure_for_problems( | |
| a, b, aname, bname, problem_list, leaf_assert_equal_func, failure_exception | |
| ): | |
| """The recursive comparison behind assertSameStructure.""" | |
| if type(a) != type(b) and not ( # pylint: disable=unidiomatic-typecheck | |
| _are_both_of_integer_type(a, b) or _are_both_of_sequence_type(a, b) or | |
| _are_both_of_set_type(a, b) or _are_both_of_mapping_type(a, b)): | |
| # We do not distinguish between int and long types as 99.99% of Python 2 | |
| # code should never care. They collapse into a single type in Python 3. | |
| problem_list.append('%s is a %r but %s is a %r' % | |
| (aname, type(a), bname, type(b))) | |
| # If they have different types there's no point continuing | |
| return | |
| if isinstance(a, abc.Set): | |
| for k in a: | |
| if k not in b: | |
| problem_list.append( | |
| '%s has %r but %s does not' % (aname, k, bname)) | |
| for k in b: | |
| if k not in a: | |
| problem_list.append('%s lacks %r but %s has it' % (aname, k, bname)) | |
| # NOTE: a or b could be a defaultdict, so we must take care that the traversal | |
| # doesn't modify the data. | |
| elif isinstance(a, abc.Mapping): | |
| for k in a: | |
| if k in b: | |
| _walk_structure_for_problems( | |
| a[k], b[k], '%s[%r]' % (aname, k), '%s[%r]' % (bname, k), | |
| problem_list, leaf_assert_equal_func, failure_exception) | |
| else: | |
| problem_list.append( | |
| "%s has [%r] with value %r but it's missing in %s" % | |
| (aname, k, a[k], bname)) | |
| for k in b: | |
| if k not in a: | |
| problem_list.append( | |
| '%s lacks [%r] but %s has it with value %r' % | |
| (aname, k, bname, b[k])) | |
| # Strings/bytes are Sequences but we'll just do those with regular != | |
| elif (isinstance(a, abc.Sequence) and | |
| not isinstance(a, _TEXT_OR_BINARY_TYPES)): | |
| minlen = min(len(a), len(b)) | |
| for i in range(minlen): | |
| _walk_structure_for_problems( | |
| a[i], b[i], '%s[%d]' % (aname, i), '%s[%d]' % (bname, i), | |
| problem_list, leaf_assert_equal_func, failure_exception) | |
| for i in range(minlen, len(a)): | |
| problem_list.append('%s has [%i] with value %r but %s does not' % | |
| (aname, i, a[i], bname)) | |
| for i in range(minlen, len(b)): | |
| problem_list.append('%s lacks [%i] but %s has it with value %r' % | |
| (aname, i, bname, b[i])) | |
| else: | |
| try: | |
| leaf_assert_equal_func(a, b) | |
| except failure_exception: | |
| problem_list.append('%s is %r but %s is %r' % (aname, a, bname, b)) | |
| def get_command_string(command): | |
| """Returns an escaped string that can be used as a shell command. | |
| Args: | |
| command: List or string representing the command to run. | |
| Returns: | |
| A string suitable for use as a shell command. | |
| """ | |
| if isinstance(command, str): | |
| return command | |
| else: | |
| if os.name == 'nt': | |
| return ' '.join(command) | |
| else: | |
| # The following is identical to Python 3's shlex.quote function. | |
| command_string = '' | |
| for word in command: | |
| # Single quote word, and replace each ' in word with '"'"' | |
| command_string += "'" + word.replace("'", "'\"'\"'") + "' " | |
| return command_string[:-1] | |
| def get_command_stderr(command, env=None, close_fds=True): | |
| """Runs the given shell command and returns a tuple. | |
| Args: | |
| command: List or string representing the command to run. | |
| env: Dictionary of environment variable settings. If None, no environment | |
| variables will be set for the child process. This is to make tests | |
| more hermetic. NOTE: this behavior is different than the standard | |
| subprocess module. | |
| close_fds: Whether or not to close all open fd's in the child after forking. | |
| On Windows, this is ignored and close_fds is always False. | |
| Returns: | |
| Tuple of (exit status, text printed to stdout and stderr by the command). | |
| """ | |
| if env is None: env = {} | |
| if os.name == 'nt': | |
| # Windows does not support setting close_fds to True while also redirecting | |
| # standard handles. | |
| close_fds = False | |
| use_shell = isinstance(command, str) | |
| process = subprocess.Popen( | |
| command, | |
| close_fds=close_fds, | |
| env=env, | |
| shell=use_shell, | |
| stderr=subprocess.STDOUT, | |
| stdout=subprocess.PIPE) | |
| output = process.communicate()[0] | |
| exit_status = process.wait() | |
| return (exit_status, output) | |
| def _quote_long_string(s): | |
| # type: (Union[Text, bytes, bytearray]) -> Text | |
| """Quotes a potentially multi-line string to make the start and end obvious. | |
| Args: | |
| s: A string. | |
| Returns: | |
| The quoted string. | |
| """ | |
| if isinstance(s, (bytes, bytearray)): | |
| try: | |
| s = s.decode('utf-8') | |
| except UnicodeDecodeError: | |
| s = str(s) | |
| return ('8<-----------\n' + | |
| s + '\n' + | |
| '----------->8\n') | |
| def print_python_version(): | |
| # type: () -> None | |
| # Having this in the test output logs by default helps debugging when all | |
| # you've got is the log and no other idea of which Python was used. | |
| sys.stderr.write('Running tests under Python {0[0]}.{0[1]}.{0[2]}: ' | |
| '{1}\n'.format( | |
| sys.version_info, | |
| sys.executable if sys.executable else 'embedded.')) | |
| def main(*args, **kwargs): | |
| # type: (Text, Any) -> None | |
| """Executes a set of Python unit tests. | |
| Usually this function is called without arguments, so the | |
| unittest.TestProgram instance will get created with the default settings, | |
| so it will run all test methods of all TestCase classes in the ``__main__`` | |
| module. | |
| Args: | |
| *args: Positional arguments passed through to | |
| ``unittest.TestProgram.__init__``. | |
| **kwargs: Keyword arguments passed through to | |
| ``unittest.TestProgram.__init__``. | |
| """ | |
| print_python_version() | |
| _run_in_app(run_tests, args, kwargs) | |
| def _is_in_app_main(): | |
| # type: () -> bool | |
| """Returns True iff app.run is active.""" | |
| f = sys._getframe().f_back # pylint: disable=protected-access | |
| while f: | |
| if f.f_code == app.run.__code__: | |
| return True | |
| f = f.f_back | |
| return False | |
| def _register_sigterm_with_faulthandler(): | |
| # type: () -> None | |
| """Have faulthandler dump stacks on SIGTERM. Useful to diagnose timeouts.""" | |
| if getattr(faulthandler, 'register', None): | |
| # faulthandler.register is not available on Windows. | |
| # faulthandler.enable() is already called by app.run. | |
| try: | |
| faulthandler.register(signal.SIGTERM, chain=True) # pytype: disable=module-attr | |
| except Exception as e: # pylint: disable=broad-except | |
| sys.stderr.write('faulthandler.register(SIGTERM) failed ' | |
| '%r; ignoring.\n' % e) | |
| def _run_in_app(function, args, kwargs): | |
| # type: (Callable[..., None], Sequence[Text], Mapping[Text, Any]) -> None | |
| """Executes a set of Python unit tests, ensuring app.run. | |
| This is a private function, users should call absltest.main(). | |
| _run_in_app calculates argv to be the command-line arguments of this program | |
| (without the flags), sets the default of FLAGS.alsologtostderr to True, | |
| then it calls function(argv, args, kwargs), making sure that `function' | |
| will get called within app.run(). _run_in_app does this by checking whether | |
| it is called by app.run(), or by calling app.run() explicitly. | |
| The reason why app.run has to be ensured is to make sure that | |
| flags are parsed and stripped properly, and other initializations done by | |
| the app module are also carried out, no matter if absltest.run() is called | |
| from within or outside app.run(). | |
| If _run_in_app is called from within app.run(), then it will reparse | |
| sys.argv and pass the result without command-line flags into the argv | |
| argument of `function'. The reason why this parsing is needed is that | |
| __main__.main() calls absltest.main() without passing its argv. So the | |
| only way _run_in_app could get to know the argv without the flags is that | |
| it reparses sys.argv. | |
| _run_in_app changes the default of FLAGS.alsologtostderr to True so that the | |
| test program's stderr will contain all the log messages unless otherwise | |
| specified on the command-line. This overrides any explicit assignment to | |
| FLAGS.alsologtostderr by the test program prior to the call to _run_in_app() | |
| (e.g. in __main__.main). | |
| Please note that _run_in_app (and the function it calls) is allowed to make | |
| changes to kwargs. | |
| Args: | |
| function: absltest.run_tests or a similar function. It will be called as | |
| function(argv, args, kwargs) where argv is a list containing the | |
| elements of sys.argv without the command-line flags. | |
| args: Positional arguments passed through to unittest.TestProgram.__init__. | |
| kwargs: Keyword arguments passed through to unittest.TestProgram.__init__. | |
| """ | |
| if _is_in_app_main(): | |
| _register_sigterm_with_faulthandler() | |
| # Change the default of alsologtostderr from False to True, so the test | |
| # programs's stderr will contain all the log messages. | |
| # If --alsologtostderr=false is specified in the command-line, or user | |
| # has called FLAGS.alsologtostderr = False before, then the value is kept | |
| # False. | |
| FLAGS.set_default('alsologtostderr', True) | |
| # Here we only want to get the `argv` without the flags. To avoid any | |
| # side effects of parsing flags, we temporarily stub out the `parse` method | |
| stored_parse_methods = {} | |
| noop_parse = lambda _: None | |
| for name in FLAGS: | |
| # Avoid any side effects of parsing flags. | |
| stored_parse_methods[name] = FLAGS[name].parse | |
| # This must be a separate loop since multiple flag names (short_name=) can | |
| # point to the same flag object. | |
| for name in FLAGS: | |
| FLAGS[name].parse = noop_parse | |
| try: | |
| argv = FLAGS(sys.argv) | |
| finally: | |
| for name in FLAGS: | |
| FLAGS[name].parse = stored_parse_methods[name] | |
| sys.stdout.flush() | |
| function(argv, args, kwargs) | |
| else: | |
| # Send logging to stderr. Use --alsologtostderr instead of --logtostderr | |
| # in case tests are reading their own logs. | |
| FLAGS.set_default('alsologtostderr', True) | |
| def main_function(argv): | |
| _register_sigterm_with_faulthandler() | |
| function(argv, args, kwargs) | |
| app.run(main=main_function) | |
| def _is_suspicious_attribute(testCaseClass, name): | |
| # type: (Type, Text) -> bool | |
| """Returns True if an attribute is a method named like a test method.""" | |
| if name.startswith('Test') and len(name) > 4 and name[4].isupper(): | |
| attr = getattr(testCaseClass, name) | |
| if inspect.isfunction(attr) or inspect.ismethod(attr): | |
| args = inspect.getfullargspec(attr) | |
| return (len(args.args) == 1 and args.args[0] == 'self' and | |
| args.varargs is None and args.varkw is None and | |
| not args.kwonlyargs) | |
| return False | |
| def skipThisClass(reason): | |
| # type: (Text) -> Callable[[_T], _T] | |
| """Skip tests in the decorated TestCase, but not any of its subclasses. | |
| This decorator indicates that this class should skip all its tests, but not | |
| any of its subclasses. Useful for if you want to share testMethod or setUp | |
| implementations between a number of concrete testcase classes. | |
| Example usage, showing how you can share some common test methods between | |
| subclasses. In this example, only ``BaseTest`` will be marked as skipped, and | |
| not RealTest or SecondRealTest:: | |
| @absltest.skipThisClass("Shared functionality") | |
| class BaseTest(absltest.TestCase): | |
| def test_simple_functionality(self): | |
| self.assertEqual(self.system_under_test.method(), 1) | |
| class RealTest(BaseTest): | |
| def setUp(self): | |
| super().setUp() | |
| self.system_under_test = MakeSystem(argument) | |
| def test_specific_behavior(self): | |
| ... | |
| class SecondRealTest(BaseTest): | |
| def setUp(self): | |
| super().setUp() | |
| self.system_under_test = MakeSystem(other_arguments) | |
| def test_other_behavior(self): | |
| ... | |
| Args: | |
| reason: The reason we have a skip in place. For instance: 'shared test | |
| methods' or 'shared assertion methods'. | |
| Returns: | |
| Decorator function that will cause a class to be skipped. | |
| """ | |
| if isinstance(reason, type): | |
| raise TypeError('Got {!r}, expected reason as string'.format(reason)) | |
| def _skip_class(test_case_class): | |
| if not issubclass(test_case_class, unittest.TestCase): | |
| raise TypeError( | |
| 'Decorating {!r}, expected TestCase subclass'.format(test_case_class)) | |
| # Only shadow the setUpClass method if it is directly defined. If it is | |
| # in the parent class we invoke it via a super() call instead of holding | |
| # a reference to it. | |
| shadowed_setupclass = test_case_class.__dict__.get('setUpClass', None) | |
| def replacement_setupclass(cls, *args, **kwargs): | |
| # Skip this class if it is the one that was decorated with @skipThisClass | |
| if cls is test_case_class: | |
| raise SkipTest(reason) | |
| if shadowed_setupclass: | |
| # Pass along `cls` so the MRO chain doesn't break. | |
| # The original method is a `classmethod` descriptor, which can't | |
| # be directly called, but `__func__` has the underlying function. | |
| return shadowed_setupclass.__func__(cls, *args, **kwargs) | |
| else: | |
| # Because there's no setUpClass() defined directly on test_case_class, | |
| # we call super() ourselves to continue execution of the inheritance | |
| # chain. | |
| return super(test_case_class, cls).setUpClass(*args, **kwargs) | |
| test_case_class.setUpClass = replacement_setupclass | |
| return test_case_class | |
| return _skip_class | |
| class TestLoader(unittest.TestLoader): | |
| """A test loader which supports common test features. | |
| Supported features include: | |
| * Banning untested methods with test-like names: methods attached to this | |
| testCase with names starting with `Test` are ignored by the test runner, | |
| and often represent mistakenly-omitted test cases. This loader will raise | |
| a TypeError when attempting to load a TestCase with such methods. | |
| * Randomization of test case execution order (optional). | |
| """ | |
| _ERROR_MSG = textwrap.dedent("""Method '%s' is named like a test case but | |
| is not one. This is often a bug. If you want it to be a test method, | |
| name it with 'test' in lowercase. If not, rename the method to not begin | |
| with 'Test'.""") | |
| def __init__(self, *args, **kwds): | |
| super(TestLoader, self).__init__(*args, **kwds) | |
| seed = _get_default_randomize_ordering_seed() | |
| if seed: | |
| self._randomize_ordering_seed = seed | |
| self._random = random.Random(self._randomize_ordering_seed) | |
| else: | |
| self._randomize_ordering_seed = None | |
| self._random = None | |
| def getTestCaseNames(self, testCaseClass): # pylint:disable=invalid-name | |
| """Validates and returns a (possibly randomized) list of test case names.""" | |
| for name in dir(testCaseClass): | |
| if _is_suspicious_attribute(testCaseClass, name): | |
| raise TypeError(TestLoader._ERROR_MSG % name) | |
| names = list(super(TestLoader, self).getTestCaseNames(testCaseClass)) | |
| if self._randomize_ordering_seed is not None: | |
| logging.info( | |
| 'Randomizing test order with seed: %d', self._randomize_ordering_seed) | |
| logging.info( | |
| 'To reproduce this order, re-run with ' | |
| '--test_randomize_ordering_seed=%d', self._randomize_ordering_seed) | |
| self._random.shuffle(names) | |
| return names | |
| def get_default_xml_output_filename(): | |
| # type: () -> Optional[Text] | |
| if os.environ.get('XML_OUTPUT_FILE'): | |
| return os.environ['XML_OUTPUT_FILE'] | |
| elif os.environ.get('RUNNING_UNDER_TEST_DAEMON'): | |
| return os.path.join(os.path.dirname(TEST_TMPDIR.value), 'test_detail.xml') | |
| elif os.environ.get('TEST_XMLOUTPUTDIR'): | |
| return os.path.join( | |
| os.environ['TEST_XMLOUTPUTDIR'], | |
| os.path.splitext(os.path.basename(sys.argv[0]))[0] + '.xml') | |
| def _setup_filtering(argv: MutableSequence[str]) -> bool: | |
| """Implements the bazel test filtering protocol. | |
| The following environment variable is used in this method: | |
| TESTBRIDGE_TEST_ONLY: string, if set, is forwarded to the unittest | |
| framework to use as a test filter. Its value is split with shlex, then: | |
| 1. On Python 3.6 and before, split values are passed as positional | |
| arguments on argv. | |
| 2. On Python 3.7+, split values are passed to unittest's `-k` flag. Tests | |
| are matched by glob patterns or substring. See | |
| https://docs.python.org/3/library/unittest.html#cmdoption-unittest-k | |
| Args: | |
| argv: the argv to mutate in-place. | |
| Returns: | |
| Whether test filtering is requested. | |
| """ | |
| test_filter = os.environ.get('TESTBRIDGE_TEST_ONLY') | |
| if argv is None or not test_filter: | |
| return False | |
| filters = shlex.split(test_filter) | |
| if sys.version_info[:2] >= (3, 7): | |
| filters = ['-k=' + test_filter for test_filter in filters] | |
| argv[1:1] = filters | |
| return True | |
| def _setup_test_runner_fail_fast(argv): | |
| # type: (MutableSequence[Text]) -> None | |
| """Implements the bazel test fail fast protocol. | |
| The following environment variable is used in this method: | |
| TESTBRIDGE_TEST_RUNNER_FAIL_FAST=<1|0> | |
| If set to 1, --failfast is passed to the unittest framework to return upon | |
| first failure. | |
| Args: | |
| argv: the argv to mutate in-place. | |
| """ | |
| if argv is None: | |
| return | |
| if os.environ.get('TESTBRIDGE_TEST_RUNNER_FAIL_FAST') != '1': | |
| return | |
| argv[1:1] = ['--failfast'] | |
| def _setup_sharding( | |
| custom_loader: Optional[unittest.TestLoader] = None, | |
| ) -> Tuple[unittest.TestLoader, Optional[int]]: | |
| """Implements the bazel sharding protocol. | |
| The following environment variables are used in this method: | |
| TEST_SHARD_STATUS_FILE: string, if set, points to a file. We write a blank | |
| file to tell the test runner that this test implements the test sharding | |
| protocol. | |
| TEST_TOTAL_SHARDS: int, if set, sharding is requested. | |
| TEST_SHARD_INDEX: int, must be set if TEST_TOTAL_SHARDS is set. Specifies | |
| the shard index for this instance of the test process. Must satisfy: | |
| 0 <= TEST_SHARD_INDEX < TEST_TOTAL_SHARDS. | |
| Args: | |
| custom_loader: A TestLoader to be made sharded. | |
| Returns: | |
| A tuple of ``(test_loader, shard_index)``. ``test_loader`` is for | |
| shard-filtering or the standard test loader depending on the sharding | |
| environment variables. ``shard_index`` is the shard index, or ``None`` when | |
| sharding is not used. | |
| """ | |
| # It may be useful to write the shard file even if the other sharding | |
| # environment variables are not set. Test runners may use this functionality | |
| # to query whether a test binary implements the test sharding protocol. | |
| if 'TEST_SHARD_STATUS_FILE' in os.environ: | |
| try: | |
| with open(os.environ['TEST_SHARD_STATUS_FILE'], 'w') as f: | |
| f.write('') | |
| except IOError: | |
| sys.stderr.write('Error opening TEST_SHARD_STATUS_FILE (%s). Exiting.' | |
| % os.environ['TEST_SHARD_STATUS_FILE']) | |
| sys.exit(1) | |
| base_loader = custom_loader or TestLoader() | |
| if 'TEST_TOTAL_SHARDS' not in os.environ: | |
| # Not using sharding, use the expected test loader. | |
| return base_loader, None | |
| total_shards = int(os.environ['TEST_TOTAL_SHARDS']) | |
| shard_index = int(os.environ['TEST_SHARD_INDEX']) | |
| if shard_index < 0 or shard_index >= total_shards: | |
| sys.stderr.write('ERROR: Bad sharding values. index=%d, total=%d\n' % | |
| (shard_index, total_shards)) | |
| sys.exit(1) | |
| # Replace the original getTestCaseNames with one that returns | |
| # the test case names for this shard. | |
| delegate_get_names = base_loader.getTestCaseNames | |
| bucket_iterator = itertools.cycle(range(total_shards)) | |
| def getShardedTestCaseNames(testCaseClass): | |
| filtered_names = [] | |
| # We need to sort the list of tests in order to determine which tests this | |
| # shard is responsible for; however, it's important to preserve the order | |
| # returned by the base loader, e.g. in the case of randomized test ordering. | |
| ordered_names = delegate_get_names(testCaseClass) | |
| for testcase in sorted(ordered_names): | |
| bucket = next(bucket_iterator) | |
| if bucket == shard_index: | |
| filtered_names.append(testcase) | |
| return [x for x in ordered_names if x in filtered_names] | |
| base_loader.getTestCaseNames = getShardedTestCaseNames | |
| return base_loader, shard_index | |
| def _run_and_get_tests_result( | |
| argv: MutableSequence[str], | |
| args: Sequence[Any], | |
| kwargs: MutableMapping[str, Any], | |
| xml_test_runner_class: Type[unittest.TextTestRunner], | |
| ) -> Tuple[unittest.TestResult, bool]: | |
| """Same as run_tests, but it doesn't exit. | |
| Args: | |
| argv: sys.argv with the command-line flags removed from the front, i.e. the | |
| argv with which :func:`app.run()<absl.app.run>` has called | |
| ``__main__.main``. It is passed to | |
| ``unittest.TestProgram.__init__(argv=)``, which does its own flag parsing. | |
| It is ignored if kwargs contains an argv entry. | |
| args: Positional arguments passed through to | |
| ``unittest.TestProgram.__init__``. | |
| kwargs: Keyword arguments passed through to | |
| ``unittest.TestProgram.__init__``. | |
| xml_test_runner_class: The type of the test runner class. | |
| Returns: | |
| A tuple of ``(test_result, fail_when_no_tests_ran)``. | |
| ``fail_when_no_tests_ran`` indicates whether the test should fail when | |
| no tests ran. | |
| """ | |
| # The entry from kwargs overrides argv. | |
| argv = kwargs.pop('argv', argv) | |
| if sys.version_info[:2] >= (3, 12): | |
| # Python 3.12 unittest changed the behavior from PASS to FAIL in | |
| # https://github.com/python/cpython/pull/102051. absltest follows this. | |
| fail_when_no_tests_ran = True | |
| else: | |
| # Historically, absltest and unittest before Python 3.12 passes if no tests | |
| # ran. | |
| fail_when_no_tests_ran = False | |
| # Set up test filtering if requested in environment. | |
| if _setup_filtering(argv): | |
| # When test filtering is requested, ideally we also want to fail when no | |
| # tests ran. However, the test filters are usually done when running bazel. | |
| # When you run multiple targets, e.g. `bazel test //my_dir/... | |
| # --test_filter=MyTest`, you don't necessarily want individual tests to fail | |
| # because no tests match in that particular target. | |
| # Due to this use case, we don't fail when test filtering is requested via | |
| # the environment variable from bazel. | |
| fail_when_no_tests_ran = False | |
| # Set up --failfast as requested in environment | |
| _setup_test_runner_fail_fast(argv) | |
| # Shard the (default or custom) loader if sharding is turned on. | |
| kwargs['testLoader'], shard_index = _setup_sharding( | |
| kwargs.get('testLoader', None) | |
| ) | |
| if shard_index is not None and shard_index > 0: | |
| # When sharding is requested, all the shards except the first one shall not | |
| # fail when no tests ran. This happens when the shard count is greater than | |
| # the test case count. | |
| fail_when_no_tests_ran = False | |
| # XML file name is based upon (sorted by priority): | |
| # --xml_output_file flag, XML_OUTPUT_FILE variable, | |
| # TEST_XMLOUTPUTDIR variable or RUNNING_UNDER_TEST_DAEMON variable. | |
| if not FLAGS.xml_output_file: | |
| FLAGS.xml_output_file = get_default_xml_output_filename() | |
| xml_output_file = FLAGS.xml_output_file | |
| xml_buffer = None | |
| if xml_output_file: | |
| xml_output_dir = os.path.dirname(xml_output_file) | |
| if xml_output_dir and not os.path.isdir(xml_output_dir): | |
| try: | |
| os.makedirs(xml_output_dir) | |
| except OSError as e: | |
| # File exists error can occur with concurrent tests | |
| if e.errno != errno.EEXIST: | |
| raise | |
| # Fail early if we can't write to the XML output file. This is so that we | |
| # don't waste people's time running tests that will just fail anyways. | |
| with _open(xml_output_file, 'w'): | |
| pass | |
| # We can reuse testRunner if it supports XML output (e. g. by inheriting | |
| # from xml_reporter.TextAndXMLTestRunner). Otherwise we need to use | |
| # xml_reporter.TextAndXMLTestRunner. | |
| if (kwargs.get('testRunner') is not None | |
| and not hasattr(kwargs['testRunner'], 'set_default_xml_stream')): | |
| sys.stderr.write('WARNING: XML_OUTPUT_FILE or --xml_output_file setting ' | |
| 'overrides testRunner=%r setting (possibly from --pdb)' | |
| % (kwargs['testRunner'])) | |
| # Passing a class object here allows TestProgram to initialize | |
| # instances based on its kwargs and/or parsed command-line args. | |
| kwargs['testRunner'] = xml_test_runner_class | |
| if kwargs.get('testRunner') is None: | |
| kwargs['testRunner'] = xml_test_runner_class | |
| # Use an in-memory buffer (not backed by the actual file) to store the XML | |
| # report, because some tools modify the file (e.g., create a placeholder | |
| # with partial information, in case the test process crashes). | |
| xml_buffer = io.StringIO() | |
| kwargs['testRunner'].set_default_xml_stream(xml_buffer) # pytype: disable=attribute-error | |
| # If we've used a seed to randomize test case ordering, we want to record it | |
| # as a top-level attribute in the `testsuites` section of the XML output. | |
| randomize_ordering_seed = getattr( | |
| kwargs['testLoader'], '_randomize_ordering_seed', None) | |
| setter = getattr(kwargs['testRunner'], 'set_testsuites_property', None) | |
| if randomize_ordering_seed and setter: | |
| setter('test_randomize_ordering_seed', randomize_ordering_seed) | |
| elif kwargs.get('testRunner') is None: | |
| kwargs['testRunner'] = _pretty_print_reporter.TextTestRunner | |
| if FLAGS.pdb_post_mortem: | |
| runner = kwargs['testRunner'] | |
| # testRunner can be a class or an instance, which must be tested for | |
| # differently. | |
| # Overriding testRunner isn't uncommon, so only enable the debugging | |
| # integration if the runner claims it does; we don't want to accidentally | |
| # clobber something on the runner. | |
| if ((isinstance(runner, type) and | |
| issubclass(runner, _pretty_print_reporter.TextTestRunner)) or | |
| isinstance(runner, _pretty_print_reporter.TextTestRunner)): | |
| runner.run_for_debugging = True | |
| # Make sure tmpdir exists. | |
| if not os.path.isdir(TEST_TMPDIR.value): | |
| try: | |
| os.makedirs(TEST_TMPDIR.value) | |
| except OSError as e: | |
| # Concurrent test might have created the directory. | |
| if e.errno != errno.EEXIST: | |
| raise | |
| # Let unittest.TestProgram.__init__ do its own argv parsing, e.g. for '-v', | |
| # on argv, which is sys.argv without the command-line flags. | |
| kwargs['argv'] = argv | |
| # Request unittest.TestProgram to not exit. The exit will be handled by | |
| # `absltest.run_tests`. | |
| kwargs['exit'] = False | |
| try: | |
| test_program = unittest.TestProgram(*args, **kwargs) | |
| return test_program.result, fail_when_no_tests_ran | |
| finally: | |
| if xml_buffer: | |
| try: | |
| with _open(xml_output_file, 'w') as f: | |
| f.write(xml_buffer.getvalue()) | |
| finally: | |
| xml_buffer.close() | |
| def run_tests( | |
| argv: MutableSequence[Text], | |
| args: Sequence[Any], | |
| kwargs: MutableMapping[Text, Any], | |
| ) -> None: | |
| """Executes a set of Python unit tests. | |
| Most users should call absltest.main() instead of run_tests. | |
| Please note that run_tests should be called from app.run. | |
| Calling absltest.main() would ensure that. | |
| Please note that run_tests is allowed to make changes to kwargs. | |
| Args: | |
| argv: sys.argv with the command-line flags removed from the front, i.e. the | |
| argv with which :func:`app.run()<absl.app.run>` has called | |
| ``__main__.main``. It is passed to | |
| ``unittest.TestProgram.__init__(argv=)``, which does its own flag parsing. | |
| It is ignored if kwargs contains an argv entry. | |
| args: Positional arguments passed through to | |
| ``unittest.TestProgram.__init__``. | |
| kwargs: Keyword arguments passed through to | |
| ``unittest.TestProgram.__init__``. | |
| """ | |
| result, fail_when_no_tests_ran = _run_and_get_tests_result( | |
| argv, args, kwargs, xml_reporter.TextAndXMLTestRunner | |
| ) | |
| if fail_when_no_tests_ran and result.testsRun == 0: | |
| # Python 3.12 unittest exits with 5 when no tests ran. The code comes from | |
| # pytest which does the same thing. | |
| sys.exit(5) | |
| sys.exit(not result.wasSuccessful()) | |
| def _rmtree_ignore_errors(path): | |
| # type: (Text) -> None | |
| if os.path.isfile(path): | |
| try: | |
| os.unlink(path) | |
| except OSError: | |
| pass | |
| else: | |
| shutil.rmtree(path, ignore_errors=True) | |
| def _get_first_part(path): | |
| # type: (Text) -> Text | |
| parts = path.split(os.sep, 1) | |
| return parts[0] | |