123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848 |
- import logging
- import multiprocessing
- import os
- import unittest.loader
- from argparse import ArgumentParser
- from contextlib import contextmanager
- from importlib import import_module
- from unittest import TestSuite, TextTestRunner, defaultTestLoader, mock
- from django.db import connections
- from django.test import SimpleTestCase
- from django.test.runner import DiscoverRunner, get_max_test_processes
- from django.test.utils import (
- NullTimeKeeper,
- TimeKeeper,
- captured_stderr,
- captured_stdout,
- )
- from django.utils.version import PY312
- @contextmanager
- def change_cwd(directory):
- current_dir = os.path.abspath(os.path.dirname(__file__))
- new_dir = os.path.join(current_dir, directory)
- old_cwd = os.getcwd()
- os.chdir(new_dir)
- try:
- yield
- finally:
- os.chdir(old_cwd)
- @contextmanager
- def change_loader_patterns(patterns):
- original_patterns = DiscoverRunner.test_loader.testNamePatterns
- DiscoverRunner.test_loader.testNamePatterns = patterns
- try:
- yield
- finally:
- DiscoverRunner.test_loader.testNamePatterns = original_patterns
- # Isolate from the real environment.
- @mock.patch.dict(os.environ, {}, clear=True)
- @mock.patch.object(multiprocessing, "cpu_count", return_value=12)
- # Python 3.8 on macOS defaults to 'spawn' mode.
- @mock.patch.object(multiprocessing, "get_start_method", return_value="fork")
- class DiscoverRunnerParallelArgumentTests(SimpleTestCase):
- def get_parser(self):
- parser = ArgumentParser()
- DiscoverRunner.add_arguments(parser)
- return parser
- def test_parallel_default(self, *mocked_objects):
- result = self.get_parser().parse_args([])
- self.assertEqual(result.parallel, 0)
- def test_parallel_flag(self, *mocked_objects):
- result = self.get_parser().parse_args(["--parallel"])
- self.assertEqual(result.parallel, "auto")
- def test_parallel_auto(self, *mocked_objects):
- result = self.get_parser().parse_args(["--parallel", "auto"])
- self.assertEqual(result.parallel, "auto")
- def test_parallel_count(self, *mocked_objects):
- result = self.get_parser().parse_args(["--parallel", "17"])
- self.assertEqual(result.parallel, 17)
- def test_parallel_invalid(self, *mocked_objects):
- with self.assertRaises(SystemExit), captured_stderr() as stderr:
- self.get_parser().parse_args(["--parallel", "unaccepted"])
- msg = "argument --parallel: 'unaccepted' is not an integer or the string 'auto'"
- self.assertIn(msg, stderr.getvalue())
- def test_get_max_test_processes(self, *mocked_objects):
- self.assertEqual(get_max_test_processes(), 12)
- @mock.patch.dict(os.environ, {"DJANGO_TEST_PROCESSES": "7"})
- def test_get_max_test_processes_env_var(self, *mocked_objects):
- self.assertEqual(get_max_test_processes(), 7)
- def test_get_max_test_processes_spawn(
- self,
- mocked_get_start_method,
- mocked_cpu_count,
- ):
- mocked_get_start_method.return_value = "spawn"
- self.assertEqual(get_max_test_processes(), 12)
- with mock.patch.dict(os.environ, {"DJANGO_TEST_PROCESSES": "7"}):
- self.assertEqual(get_max_test_processes(), 7)
- def test_get_max_test_processes_forkserver(
- self,
- mocked_get_start_method,
- mocked_cpu_count,
- ):
- mocked_get_start_method.return_value = "forkserver"
- self.assertEqual(get_max_test_processes(), 1)
- with mock.patch.dict(os.environ, {"DJANGO_TEST_PROCESSES": "7"}):
- self.assertEqual(get_max_test_processes(), 1)
- class DiscoverRunnerTests(SimpleTestCase):
- @staticmethod
- def get_test_methods_names(suite):
- return [t.__class__.__name__ + "." + t._testMethodName for t in suite._tests]
- def test_init_debug_mode(self):
- runner = DiscoverRunner()
- self.assertFalse(runner.debug_mode)
- def test_add_arguments_shuffle(self):
- parser = ArgumentParser()
- DiscoverRunner.add_arguments(parser)
- ns = parser.parse_args([])
- self.assertIs(ns.shuffle, False)
- ns = parser.parse_args(["--shuffle"])
- self.assertIsNone(ns.shuffle)
- ns = parser.parse_args(["--shuffle", "5"])
- self.assertEqual(ns.shuffle, 5)
- def test_add_arguments_debug_mode(self):
- parser = ArgumentParser()
- DiscoverRunner.add_arguments(parser)
- ns = parser.parse_args([])
- self.assertFalse(ns.debug_mode)
- ns = parser.parse_args(["--debug-mode"])
- self.assertTrue(ns.debug_mode)
- def test_setup_shuffler_no_shuffle_argument(self):
- runner = DiscoverRunner()
- self.assertIs(runner.shuffle, False)
- runner.setup_shuffler()
- self.assertIsNone(runner.shuffle_seed)
- def test_setup_shuffler_shuffle_none(self):
- runner = DiscoverRunner(shuffle=None)
- self.assertIsNone(runner.shuffle)
- with mock.patch("random.randint", return_value=1):
- with captured_stdout() as stdout:
- runner.setup_shuffler()
- self.assertEqual(stdout.getvalue(), "Using shuffle seed: 1 (generated)\n")
- self.assertEqual(runner.shuffle_seed, 1)
- def test_setup_shuffler_shuffle_int(self):
- runner = DiscoverRunner(shuffle=2)
- self.assertEqual(runner.shuffle, 2)
- with captured_stdout() as stdout:
- runner.setup_shuffler()
- expected_out = "Using shuffle seed: 2 (given)\n"
- self.assertEqual(stdout.getvalue(), expected_out)
- self.assertEqual(runner.shuffle_seed, 2)
- def test_load_tests_for_label_file_path(self):
- with change_cwd("."):
- msg = (
- "One of the test labels is a path to a file: "
- "'test_discover_runner.py', which is not supported. Use a "
- "dotted module name or path to a directory instead."
- )
- with self.assertRaisesMessage(RuntimeError, msg):
- DiscoverRunner().load_tests_for_label("test_discover_runner.py", {})
- def test_dotted_test_module(self):
- count = (
- DiscoverRunner(verbosity=0)
- .build_suite(
- ["test_runner_apps.sample.tests_sample"],
- )
- .countTestCases()
- )
- self.assertEqual(count, 4)
- def test_dotted_test_class_vanilla_unittest(self):
- count = (
- DiscoverRunner(verbosity=0)
- .build_suite(
- ["test_runner_apps.sample.tests_sample.TestVanillaUnittest"],
- )
- .countTestCases()
- )
- self.assertEqual(count, 1)
- def test_dotted_test_class_django_testcase(self):
- count = (
- DiscoverRunner(verbosity=0)
- .build_suite(
- ["test_runner_apps.sample.tests_sample.TestDjangoTestCase"],
- )
- .countTestCases()
- )
- self.assertEqual(count, 1)
- def test_dotted_test_method_django_testcase(self):
- count = (
- DiscoverRunner(verbosity=0)
- .build_suite(
- ["test_runner_apps.sample.tests_sample.TestDjangoTestCase.test_sample"],
- )
- .countTestCases()
- )
- self.assertEqual(count, 1)
- def test_pattern(self):
- count = (
- DiscoverRunner(
- pattern="*_tests.py",
- verbosity=0,
- )
- .build_suite(["test_runner_apps.sample"])
- .countTestCases()
- )
- self.assertEqual(count, 1)
- def test_name_patterns(self):
- all_test_1 = [
- "DjangoCase1.test_1",
- "DjangoCase2.test_1",
- "SimpleCase1.test_1",
- "SimpleCase2.test_1",
- "UnittestCase1.test_1",
- "UnittestCase2.test_1",
- ]
- all_test_2 = [
- "DjangoCase1.test_2",
- "DjangoCase2.test_2",
- "SimpleCase1.test_2",
- "SimpleCase2.test_2",
- "UnittestCase1.test_2",
- "UnittestCase2.test_2",
- ]
- all_tests = sorted([*all_test_1, *all_test_2, "UnittestCase2.test_3_test"])
- for pattern, expected in [
- [["test_1"], all_test_1],
- [["UnittestCase1"], ["UnittestCase1.test_1", "UnittestCase1.test_2"]],
- [["*test"], ["UnittestCase2.test_3_test"]],
- [["test*"], all_tests],
- [["test"], all_tests],
- [["test_1", "test_2"], sorted([*all_test_1, *all_test_2])],
- [["test*1"], all_test_1],
- ]:
- with self.subTest(pattern):
- suite = DiscoverRunner(
- test_name_patterns=pattern,
- verbosity=0,
- ).build_suite(["test_runner_apps.simple"])
- self.assertEqual(expected, self.get_test_methods_names(suite))
- def test_loader_patterns_not_mutated(self):
- runner = DiscoverRunner(test_name_patterns=["test_sample"], verbosity=0)
- tests = [
- ("test_runner_apps.sample.tests", 1),
- ("test_runner_apps.sample.tests.Test.test_sample", 1),
- ("test_runner_apps.sample.empty", 0),
- ("test_runner_apps.sample.tests_sample.EmptyTestCase", 0),
- ]
- for test_labels, tests_count in tests:
- with self.subTest(test_labels=test_labels):
- with change_loader_patterns(["UnittestCase1"]):
- count = runner.build_suite([test_labels]).countTestCases()
- self.assertEqual(count, tests_count)
- self.assertEqual(
- runner.test_loader.testNamePatterns, ["UnittestCase1"]
- )
- def test_loader_patterns_not_mutated_when_test_label_is_file_path(self):
- runner = DiscoverRunner(test_name_patterns=["test_sample"], verbosity=0)
- with change_cwd("."), change_loader_patterns(["UnittestCase1"]):
- with self.assertRaises(RuntimeError):
- runner.build_suite(["test_discover_runner.py"])
- self.assertEqual(runner.test_loader.testNamePatterns, ["UnittestCase1"])
- def test_file_path(self):
- with change_cwd(".."):
- count = (
- DiscoverRunner(verbosity=0)
- .build_suite(
- ["test_runner_apps/sample/"],
- )
- .countTestCases()
- )
- self.assertEqual(count, 5)
- def test_empty_label(self):
- """
- If the test label is empty, discovery should happen on the current
- working directory.
- """
- with change_cwd("."):
- suite = DiscoverRunner(verbosity=0).build_suite([])
- self.assertEqual(
- suite._tests[0].id().split(".")[0],
- os.path.basename(os.getcwd()),
- )
- def test_empty_test_case(self):
- count = (
- DiscoverRunner(verbosity=0)
- .build_suite(
- ["test_runner_apps.sample.tests_sample.EmptyTestCase"],
- )
- .countTestCases()
- )
- self.assertEqual(count, 0)
- def test_discovery_on_package(self):
- count = (
- DiscoverRunner(verbosity=0)
- .build_suite(
- ["test_runner_apps.sample.tests"],
- )
- .countTestCases()
- )
- self.assertEqual(count, 1)
- def test_ignore_adjacent(self):
- """
- When given a dotted path to a module, unittest discovery searches
- not just the module, but also the directory containing the module.
- This results in tests from adjacent modules being run when they
- should not. The discover runner avoids this behavior.
- """
- count = (
- DiscoverRunner(verbosity=0)
- .build_suite(
- ["test_runner_apps.sample.empty"],
- )
- .countTestCases()
- )
- self.assertEqual(count, 0)
- def test_testcase_ordering(self):
- with change_cwd(".."):
- suite = DiscoverRunner(verbosity=0).build_suite(
- ["test_runner_apps/sample/"]
- )
- self.assertEqual(
- suite._tests[0].__class__.__name__,
- "TestDjangoTestCase",
- msg="TestDjangoTestCase should be the first test case",
- )
- self.assertEqual(
- suite._tests[1].__class__.__name__,
- "TestZimpleTestCase",
- msg="TestZimpleTestCase should be the second test case",
- )
- # All others can follow in unspecified order, including doctests
- self.assertIn(
- "DocTestCase", [t.__class__.__name__ for t in suite._tests[2:]]
- )
- def test_duplicates_ignored(self):
- """
- Tests shouldn't be discovered twice when discovering on overlapping paths.
- """
- base_app = "forms_tests"
- sub_app = "forms_tests.field_tests"
- runner = DiscoverRunner(verbosity=0)
- with self.modify_settings(INSTALLED_APPS={"append": sub_app}):
- single = runner.build_suite([base_app]).countTestCases()
- dups = runner.build_suite([base_app, sub_app]).countTestCases()
- self.assertEqual(single, dups)
- def test_reverse(self):
- """
- Reverse should reorder tests while maintaining the grouping specified
- by ``DiscoverRunner.reorder_by``.
- """
- runner = DiscoverRunner(reverse=True, verbosity=0)
- suite = runner.build_suite(
- test_labels=("test_runner_apps.sample", "test_runner_apps.simple")
- )
- self.assertIn(
- "test_runner_apps.simple",
- next(iter(suite)).id(),
- msg="Test labels should be reversed.",
- )
- suite = runner.build_suite(test_labels=("test_runner_apps.simple",))
- suite = tuple(suite)
- self.assertIn(
- "DjangoCase", suite[0].id(), msg="Test groups should not be reversed."
- )
- self.assertIn(
- "SimpleCase", suite[4].id(), msg="Test groups order should be preserved."
- )
- self.assertIn(
- "DjangoCase2", suite[0].id(), msg="Django test cases should be reversed."
- )
- self.assertIn(
- "SimpleCase2", suite[4].id(), msg="Simple test cases should be reversed."
- )
- self.assertIn(
- "UnittestCase2",
- suite[8].id(),
- msg="Unittest test cases should be reversed.",
- )
- self.assertIn(
- "test_2", suite[0].id(), msg="Methods of Django cases should be reversed."
- )
- self.assertIn(
- "test_2", suite[4].id(), msg="Methods of simple cases should be reversed."
- )
- self.assertIn(
- "test_2", suite[9].id(), msg="Methods of unittest cases should be reversed."
- )
- def test_build_suite_failed_tests_first(self):
- # The "doesnotexist" label results in a _FailedTest instance.
- suite = DiscoverRunner(verbosity=0).build_suite(
- test_labels=["test_runner_apps.sample", "doesnotexist"],
- )
- tests = list(suite)
- self.assertIsInstance(tests[0], unittest.loader._FailedTest)
- self.assertNotIsInstance(tests[-1], unittest.loader._FailedTest)
- def test_build_suite_shuffling(self):
- # These will result in unittest.loader._FailedTest instances rather
- # than TestCase objects, but they are sufficient for testing.
- labels = ["label1", "label2", "label3", "label4"]
- cases = [
- ({}, ["label1", "label2", "label3", "label4"]),
- ({"reverse": True}, ["label4", "label3", "label2", "label1"]),
- ({"shuffle": 8}, ["label4", "label1", "label3", "label2"]),
- ({"shuffle": 8, "reverse": True}, ["label2", "label3", "label1", "label4"]),
- ]
- for kwargs, expected in cases:
- with self.subTest(kwargs=kwargs):
- # Prevent writing the seed to stdout.
- runner = DiscoverRunner(**kwargs, verbosity=0)
- tests = runner.build_suite(test_labels=labels)
- # The ids have the form "unittest.loader._FailedTest.label1".
- names = [test.id().split(".")[-1] for test in tests]
- self.assertEqual(names, expected)
- def test_overridable_get_test_runner_kwargs(self):
- self.assertIsInstance(DiscoverRunner().get_test_runner_kwargs(), dict)
- def test_overridable_test_suite(self):
- self.assertEqual(DiscoverRunner().test_suite, TestSuite)
- def test_overridable_test_runner(self):
- self.assertEqual(DiscoverRunner().test_runner, TextTestRunner)
- def test_overridable_test_loader(self):
- self.assertEqual(DiscoverRunner().test_loader, defaultTestLoader)
- def test_tags(self):
- runner = DiscoverRunner(tags=["core"], verbosity=0)
- self.assertEqual(
- runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 1
- )
- runner = DiscoverRunner(tags=["fast"], verbosity=0)
- self.assertEqual(
- runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 2
- )
- runner = DiscoverRunner(tags=["slow"], verbosity=0)
- self.assertEqual(
- runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 2
- )
- def test_exclude_tags(self):
- runner = DiscoverRunner(tags=["fast"], exclude_tags=["core"], verbosity=0)
- self.assertEqual(
- runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 1
- )
- runner = DiscoverRunner(tags=["fast"], exclude_tags=["slow"], verbosity=0)
- self.assertEqual(
- runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 0
- )
- runner = DiscoverRunner(exclude_tags=["slow"], verbosity=0)
- self.assertEqual(
- runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 0
- )
- def test_tag_inheritance(self):
- def count_tests(**kwargs):
- kwargs.setdefault("verbosity", 0)
- suite = DiscoverRunner(**kwargs).build_suite(
- ["test_runner_apps.tagged.tests_inheritance"]
- )
- return suite.countTestCases()
- self.assertEqual(count_tests(tags=["foo"]), 4)
- self.assertEqual(count_tests(tags=["bar"]), 2)
- self.assertEqual(count_tests(tags=["baz"]), 2)
- self.assertEqual(count_tests(tags=["foo"], exclude_tags=["bar"]), 2)
- self.assertEqual(count_tests(tags=["foo"], exclude_tags=["bar", "baz"]), 1)
- self.assertEqual(count_tests(exclude_tags=["foo"]), 0)
- def test_tag_fail_to_load(self):
- with self.assertRaises(SyntaxError):
- import_module("test_runner_apps.tagged.tests_syntax_error")
- runner = DiscoverRunner(tags=["syntax_error"], verbosity=0)
- # A label that doesn't exist or cannot be loaded due to syntax errors
- # is always considered matching.
- suite = runner.build_suite(["doesnotexist", "test_runner_apps.tagged"])
- self.assertEqual(
- [test.id() for test in suite],
- [
- "unittest.loader._FailedTest.doesnotexist",
- "unittest.loader._FailedTest.test_runner_apps.tagged."
- "tests_syntax_error",
- ],
- )
- def test_included_tags_displayed(self):
- runner = DiscoverRunner(tags=["foo", "bar"], verbosity=2)
- with captured_stdout() as stdout:
- runner.build_suite(["test_runner_apps.tagged.tests"])
- self.assertIn("Including test tag(s): bar, foo.\n", stdout.getvalue())
- def test_excluded_tags_displayed(self):
- runner = DiscoverRunner(exclude_tags=["foo", "bar"], verbosity=3)
- with captured_stdout() as stdout:
- runner.build_suite(["test_runner_apps.tagged.tests"])
- self.assertIn("Excluding test tag(s): bar, foo.\n", stdout.getvalue())
- def test_number_of_tests_found_displayed(self):
- runner = DiscoverRunner()
- with captured_stdout() as stdout:
- runner.build_suite(
- [
- "test_runner_apps.sample.tests_sample.TestDjangoTestCase",
- "test_runner_apps.simple",
- ]
- )
- self.assertIn("Found 14 test(s).\n", stdout.getvalue())
- def test_pdb_with_parallel(self):
- msg = "You cannot use --pdb with parallel tests; pass --parallel=1 to use it."
- with self.assertRaisesMessage(ValueError, msg):
- DiscoverRunner(pdb=True, parallel=2)
- def test_number_of_parallel_workers(self):
- """Number of processes doesn't exceed the number of TestCases."""
- runner = DiscoverRunner(parallel=5, verbosity=0)
- suite = runner.build_suite(["test_runner_apps.tagged"])
- self.assertEqual(suite.processes, len(suite.subsuites))
- def test_number_of_databases_parallel_test_suite(self):
- """
- Number of databases doesn't exceed the number of TestCases with
- parallel tests.
- """
- runner = DiscoverRunner(parallel=8, verbosity=0)
- suite = runner.build_suite(["test_runner_apps.tagged"])
- self.assertEqual(suite.processes, len(suite.subsuites))
- self.assertEqual(runner.parallel, suite.processes)
- def test_number_of_databases_no_parallel_test_suite(self):
- """
- Number of databases doesn't exceed the number of TestCases with
- non-parallel tests.
- """
- runner = DiscoverRunner(parallel=8, verbosity=0)
- suite = runner.build_suite(["test_runner_apps.simple.tests.DjangoCase1"])
- self.assertEqual(runner.parallel, 1)
- self.assertIsInstance(suite, TestSuite)
- def test_buffer_mode_test_pass(self):
- runner = DiscoverRunner(buffer=True, verbosity=0)
- with captured_stdout() as stdout, captured_stderr() as stderr:
- suite = runner.build_suite(
- [
- "test_runner_apps.buffer.tests_buffer.WriteToStdoutStderrTestCase."
- "test_pass",
- ]
- )
- runner.run_suite(suite)
- self.assertNotIn("Write to stderr.", stderr.getvalue())
- self.assertNotIn("Write to stdout.", stdout.getvalue())
- def test_buffer_mode_test_fail(self):
- runner = DiscoverRunner(buffer=True, verbosity=0)
- with captured_stdout() as stdout, captured_stderr() as stderr:
- suite = runner.build_suite(
- [
- "test_runner_apps.buffer.tests_buffer.WriteToStdoutStderrTestCase."
- "test_fail",
- ]
- )
- runner.run_suite(suite)
- self.assertIn("Write to stderr.", stderr.getvalue())
- self.assertIn("Write to stdout.", stdout.getvalue())
- def run_suite_with_runner(self, runner_class, **kwargs):
- class MyRunner(DiscoverRunner):
- def test_runner(self, *args, **kwargs):
- return runner_class()
- runner = MyRunner(**kwargs)
- # Suppress logging "Using shuffle seed" to the console.
- with captured_stdout():
- runner.setup_shuffler()
- with captured_stdout() as stdout:
- try:
- result = runner.run_suite(None)
- except RuntimeError as exc:
- result = str(exc)
- output = stdout.getvalue()
- return result, output
- def test_run_suite_logs_seed(self):
- class TestRunner:
- def run(self, suite):
- return "<fake-result>"
- expected_prefix = "Used shuffle seed"
- # Test with and without shuffling enabled.
- result, output = self.run_suite_with_runner(TestRunner)
- self.assertEqual(result, "<fake-result>")
- self.assertNotIn(expected_prefix, output)
- result, output = self.run_suite_with_runner(TestRunner, shuffle=2)
- self.assertEqual(result, "<fake-result>")
- expected_output = f"{expected_prefix}: 2 (given)\n"
- self.assertEqual(output, expected_output)
- def test_run_suite_logs_seed_exception(self):
- """
- run_suite() logs the seed when TestRunner.run() raises an exception.
- """
- class TestRunner:
- def run(self, suite):
- raise RuntimeError("my exception")
- result, output = self.run_suite_with_runner(TestRunner, shuffle=2)
- self.assertEqual(result, "my exception")
- expected_output = "Used shuffle seed: 2 (given)\n"
- self.assertEqual(output, expected_output)
- @mock.patch("faulthandler.enable")
- def test_faulthandler_enabled(self, mocked_enable):
- with mock.patch("faulthandler.is_enabled", return_value=False):
- DiscoverRunner(enable_faulthandler=True)
- mocked_enable.assert_called()
- @mock.patch("faulthandler.enable")
- def test_faulthandler_already_enabled(self, mocked_enable):
- with mock.patch("faulthandler.is_enabled", return_value=True):
- DiscoverRunner(enable_faulthandler=True)
- mocked_enable.assert_not_called()
- @mock.patch("faulthandler.enable")
- def test_faulthandler_enabled_fileno(self, mocked_enable):
- # sys.stderr that is not an actual file.
- with mock.patch(
- "faulthandler.is_enabled", return_value=False
- ), captured_stderr():
- DiscoverRunner(enable_faulthandler=True)
- mocked_enable.assert_called()
- @mock.patch("faulthandler.enable")
- def test_faulthandler_disabled(self, mocked_enable):
- with mock.patch("faulthandler.is_enabled", return_value=False):
- DiscoverRunner(enable_faulthandler=False)
- mocked_enable.assert_not_called()
- def test_timings_not_captured(self):
- runner = DiscoverRunner(timing=False)
- with captured_stderr() as stderr:
- with runner.time_keeper.timed("test"):
- pass
- runner.time_keeper.print_results()
- self.assertIsInstance(runner.time_keeper, NullTimeKeeper)
- self.assertNotIn("test", stderr.getvalue())
- def test_timings_captured(self):
- runner = DiscoverRunner(timing=True)
- with captured_stderr() as stderr:
- with runner.time_keeper.timed("test"):
- pass
- runner.time_keeper.print_results()
- self.assertIsInstance(runner.time_keeper, TimeKeeper)
- self.assertIn("test", stderr.getvalue())
- def test_log(self):
- custom_low_level = 5
- custom_high_level = 45
- msg = "logging message"
- cases = [
- (0, None, False),
- (0, custom_low_level, False),
- (0, logging.DEBUG, False),
- (0, logging.INFO, False),
- (0, logging.WARNING, False),
- (0, custom_high_level, False),
- (1, None, True),
- (1, custom_low_level, False),
- (1, logging.DEBUG, False),
- (1, logging.INFO, True),
- (1, logging.WARNING, True),
- (1, custom_high_level, True),
- (2, None, True),
- (2, custom_low_level, True),
- (2, logging.DEBUG, True),
- (2, logging.INFO, True),
- (2, logging.WARNING, True),
- (2, custom_high_level, True),
- (3, None, True),
- (3, custom_low_level, True),
- (3, logging.DEBUG, True),
- (3, logging.INFO, True),
- (3, logging.WARNING, True),
- (3, custom_high_level, True),
- ]
- for verbosity, level, output in cases:
- with self.subTest(verbosity=verbosity, level=level):
- with captured_stdout() as stdout:
- runner = DiscoverRunner(verbosity=verbosity)
- runner.log(msg, level)
- self.assertEqual(stdout.getvalue(), f"{msg}\n" if output else "")
- def test_log_logger(self):
- logger = logging.getLogger("test.logging")
- cases = [
- (None, "INFO:test.logging:log message"),
- # Test a low custom logging level.
- (5, "Level 5:test.logging:log message"),
- (logging.DEBUG, "DEBUG:test.logging:log message"),
- (logging.INFO, "INFO:test.logging:log message"),
- (logging.WARNING, "WARNING:test.logging:log message"),
- # Test a high custom logging level.
- (45, "Level 45:test.logging:log message"),
- ]
- for level, expected in cases:
- with self.subTest(level=level):
- runner = DiscoverRunner(logger=logger)
- # Pass a logging level smaller than the smallest level in cases
- # in order to capture all messages.
- with self.assertLogs("test.logging", level=1) as cm:
- runner.log("log message", level)
- self.assertEqual(cm.output, [expected])
- def test_suite_result_with_failure(self):
- cases = [
- (1, "FailureTestCase"),
- (1, "ErrorTestCase"),
- (0, "ExpectedFailureTestCase"),
- (1, "UnexpectedSuccessTestCase"),
- ]
- runner = DiscoverRunner(verbosity=0)
- for expected_failures, testcase in cases:
- with self.subTest(testcase=testcase):
- suite = runner.build_suite(
- [
- f"test_runner_apps.failures.tests_failures.{testcase}",
- ]
- )
- with captured_stderr():
- result = runner.run_suite(suite)
- failures = runner.suite_result(suite, result)
- self.assertEqual(failures, expected_failures)
- @unittest.skipUnless(PY312, "unittest --durations option requires Python 3.12")
- def test_durations(self):
- with captured_stderr() as stderr, captured_stdout():
- runner = DiscoverRunner(durations=10)
- suite = runner.build_suite(["test_runner_apps.simple.tests.SimpleCase1"])
- runner.run_suite(suite)
- self.assertIn("Slowest test durations", stderr.getvalue())
- @unittest.skipUnless(PY312, "unittest --durations option requires Python 3.12")
- def test_durations_debug_sql(self):
- with captured_stderr() as stderr, captured_stdout():
- runner = DiscoverRunner(durations=10, debug_sql=True)
- suite = runner.build_suite(["test_runner_apps.simple.SimpleCase1"])
- runner.run_suite(suite)
- self.assertIn("Slowest test durations", stderr.getvalue())
- class DiscoverRunnerGetDatabasesTests(SimpleTestCase):
- runner = DiscoverRunner(verbosity=2)
- skip_msg = "Skipping setup of unused database(s): "
- def get_databases(self, test_labels):
- with captured_stdout() as stdout:
- suite = self.runner.build_suite(test_labels)
- databases = self.runner.get_databases(suite)
- return databases, stdout.getvalue()
- def assertSkippedDatabases(self, test_labels, expected_databases):
- databases, output = self.get_databases(test_labels)
- self.assertEqual(databases, expected_databases)
- skipped_databases = set(connections) - set(expected_databases)
- if skipped_databases:
- self.assertIn(self.skip_msg + ", ".join(sorted(skipped_databases)), output)
- else:
- self.assertNotIn(self.skip_msg, output)
- def test_mixed(self):
- databases, output = self.get_databases(["test_runner_apps.databases.tests"])
- self.assertEqual(databases, {"default": True, "other": False})
- self.assertNotIn(self.skip_msg, output)
- def test_all(self):
- databases, output = self.get_databases(
- ["test_runner_apps.databases.tests.AllDatabasesTests"]
- )
- self.assertEqual(databases, {alias: False for alias in connections})
- self.assertNotIn(self.skip_msg, output)
- def test_default_and_other(self):
- self.assertSkippedDatabases(
- [
- "test_runner_apps.databases.tests.DefaultDatabaseTests",
- "test_runner_apps.databases.tests.OtherDatabaseTests",
- ],
- {"default": False, "other": False},
- )
- def test_default_only(self):
- self.assertSkippedDatabases(
- [
- "test_runner_apps.databases.tests.DefaultDatabaseTests",
- ],
- {"default": False},
- )
- def test_other_only(self):
- self.assertSkippedDatabases(
- ["test_runner_apps.databases.tests.OtherDatabaseTests"], {"other": False}
- )
- def test_no_databases_required(self):
- self.assertSkippedDatabases(
- ["test_runner_apps.databases.tests.NoDatabaseTests"], {}
- )
- def test_serialize(self):
- databases, _ = self.get_databases(
- ["test_runner_apps.databases.tests.DefaultDatabaseSerializedTests"]
- )
- self.assertEqual(databases, {"default": True})
|