mirror of
				https://github.com/zulip/zulip.git
				synced 2025-11-04 05:53:43 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			598 lines
		
	
	
		
			24 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			598 lines
		
	
	
		
			24 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
from functools import partial
 | 
						|
import random
 | 
						|
import sys
 | 
						|
 | 
						|
from typing import Any, Callable, Dict, List, Optional, Set, Tuple, \
 | 
						|
    Type, TypeVar, Union
 | 
						|
from unittest import loader, runner
 | 
						|
from unittest.result import TestResult
 | 
						|
 | 
						|
from django.conf import settings
 | 
						|
from django.db import connections, ProgrammingError
 | 
						|
from django.urls.resolvers import URLPattern
 | 
						|
from django.test import TestCase
 | 
						|
from django.test import runner as django_runner
 | 
						|
from django.test.runner import DiscoverRunner
 | 
						|
from django.test.signals import template_rendered
 | 
						|
 | 
						|
from zerver.lib import test_helpers
 | 
						|
from zerver.lib.cache import bounce_key_prefix_for_testing
 | 
						|
from zerver.lib.rate_limiter import bounce_redis_key_prefix_for_testing
 | 
						|
from zerver.lib.sqlalchemy_utils import get_sqlalchemy_connection
 | 
						|
from zerver.lib.test_helpers import (
 | 
						|
    write_instrumentation_reports,
 | 
						|
    append_instrumentation_data
 | 
						|
)
 | 
						|
 | 
						|
import os
 | 
						|
import time
 | 
						|
import unittest
 | 
						|
import shutil
 | 
						|
 | 
						|
from multiprocessing.sharedctypes import Synchronized
 | 
						|
 | 
						|
from scripts.lib.zulip_tools import get_dev_uuid_var_path, TEMPLATE_DATABASE_DIR, \
 | 
						|
    get_or_create_dev_uuid_var_path
 | 
						|
 | 
						|
# We need to pick an ID for this test-backend invocation, and store it
 | 
						|
# in this global so it can be used in init_worker; this is used to
 | 
						|
# ensure the database IDs we select are unique for each `test-backend`
 | 
						|
# run.  This probably should use a locking mechanism rather than the
 | 
						|
# below hack, which fails 1/10000000 of the time.
 | 
						|
random_id_range_start = str(random.randint(1, 10000000))
 | 
						|
 | 
						|
def get_database_id(worker_id: Optional[int]=None) -> str:
 | 
						|
    if worker_id:
 | 
						|
        return "{}_{}".format(random_id_range_start, worker_id)
 | 
						|
    return random_id_range_start
 | 
						|
 | 
						|
# The root directory for this run of the test suite.
 | 
						|
TEST_RUN_DIR = get_or_create_dev_uuid_var_path(
 | 
						|
    os.path.join('test-backend', 'run_{}'.format(get_database_id())))
 | 
						|
 | 
						|
_worker_id = 0  # Used to identify the worker process.
 | 
						|
 | 
						|
ReturnT = TypeVar('ReturnT')  # Constrain return type to match
 | 
						|
 | 
						|
def slow(slowness_reason: str) -> Callable[[Callable[..., ReturnT]], Callable[..., ReturnT]]:
 | 
						|
    '''
 | 
						|
    This is a decorate that annotates a test as being "known
 | 
						|
    to be slow."  The decorator will set expected_run_time and slowness_reason
 | 
						|
    as attributes of the function.  Other code can use this annotation
 | 
						|
    as needed, e.g. to exclude these tests in "fast" mode.
 | 
						|
    '''
 | 
						|
    def decorator(f: Callable[..., ReturnT]) -> Callable[..., ReturnT]:
 | 
						|
        setattr(f, 'slowness_reason', slowness_reason)
 | 
						|
        return f
 | 
						|
 | 
						|
    return decorator
 | 
						|
 | 
						|
def is_known_slow_test(test_method: Callable[..., ReturnT]) -> bool:
 | 
						|
    return hasattr(test_method, 'slowness_reason')
 | 
						|
 | 
						|
def full_test_name(test: TestCase) -> str:
 | 
						|
    test_module = test.__module__
 | 
						|
    test_class = test.__class__.__name__
 | 
						|
    test_method = test._testMethodName
 | 
						|
    return '%s.%s.%s' % (test_module, test_class, test_method)
 | 
						|
 | 
						|
def get_test_method(test: TestCase) -> Callable[[], None]:
 | 
						|
    return getattr(test, test._testMethodName)
 | 
						|
 | 
						|
# Each tuple is delay, test_name, slowness_reason
 | 
						|
TEST_TIMINGS: List[Tuple[float, str, str]] = []
 | 
						|
 | 
						|
 | 
						|
def report_slow_tests() -> None:
 | 
						|
    timings = sorted(TEST_TIMINGS, reverse=True)
 | 
						|
    print('SLOWNESS REPORT')
 | 
						|
    print(' delay test')
 | 
						|
    print(' ----  ----')
 | 
						|
    for delay, test_name, slowness_reason in timings[:15]:
 | 
						|
        if not slowness_reason:
 | 
						|
            slowness_reason = 'UNKNOWN WHY SLOW, please investigate'
 | 
						|
        print(' %0.3f %s\n       %s\n' % (delay, test_name, slowness_reason))
 | 
						|
 | 
						|
    print('...')
 | 
						|
    for delay, test_name, slowness_reason in timings[100:]:
 | 
						|
        if slowness_reason:
 | 
						|
            print(' %.3f %s is not that slow' % (delay, test_name))
 | 
						|
            print('      consider removing @slow decorator')
 | 
						|
            print('      This may no longer be true: %s' % (slowness_reason,))
 | 
						|
 | 
						|
def enforce_timely_test_completion(test_method: Callable[..., ReturnT], test_name: str,
 | 
						|
                                   delay: float, result: unittest.TestResult) -> None:
 | 
						|
    if hasattr(test_method, 'slowness_reason'):
 | 
						|
        max_delay = 2.0  # seconds
 | 
						|
    else:
 | 
						|
        max_delay = 0.4  # seconds
 | 
						|
 | 
						|
    assert isinstance(result, TextTestResult) or isinstance(result, RemoteTestResult)
 | 
						|
 | 
						|
    if delay > max_delay:
 | 
						|
        msg = '** Test is TOO slow: %s (%.3f s)\n' % (test_name, delay)
 | 
						|
        result.addInfo(test_method, msg)
 | 
						|
 | 
						|
def fast_tests_only() -> bool:
 | 
						|
    return "FAST_TESTS_ONLY" in os.environ
 | 
						|
 | 
						|
def run_test(test: TestCase, result: TestResult) -> bool:
 | 
						|
    failed = False
 | 
						|
    test_method = get_test_method(test)
 | 
						|
 | 
						|
    if fast_tests_only() and is_known_slow_test(test_method):
 | 
						|
        return failed
 | 
						|
 | 
						|
    test_name = full_test_name(test)
 | 
						|
 | 
						|
    bounce_key_prefix_for_testing(test_name)
 | 
						|
    bounce_redis_key_prefix_for_testing(test_name)
 | 
						|
 | 
						|
    try:
 | 
						|
        test._pre_setup()
 | 
						|
    except Exception:
 | 
						|
        result.addError(test, sys.exc_info())
 | 
						|
        return True
 | 
						|
 | 
						|
    start_time = time.time()
 | 
						|
 | 
						|
    test(result)  # unittest will handle skipping, error, failure and success.
 | 
						|
 | 
						|
    delay = time.time() - start_time
 | 
						|
    enforce_timely_test_completion(test_method, test_name, delay, result)
 | 
						|
    slowness_reason = getattr(test_method, 'slowness_reason', '')
 | 
						|
    TEST_TIMINGS.append((delay, test_name, slowness_reason))
 | 
						|
 | 
						|
    test._post_teardown()
 | 
						|
    return failed
 | 
						|
 | 
						|
class TextTestResult(runner.TextTestResult):
 | 
						|
    """
 | 
						|
    This class has unpythonic function names because base class follows
 | 
						|
    this style.
 | 
						|
    """
 | 
						|
 | 
						|
    def __init__(self, *args: Any, **kwargs: Any) -> None:
 | 
						|
        super().__init__(*args, **kwargs)
 | 
						|
        self.failed_tests: List[str] = []
 | 
						|
 | 
						|
    def addInfo(self, test: TestCase, msg: str) -> None:
 | 
						|
        self.stream.write(msg)  # type: ignore[attr-defined] # https://github.com/python/typeshed/issues/3139
 | 
						|
        self.stream.flush()  # type: ignore[attr-defined] # https://github.com/python/typeshed/issues/3139
 | 
						|
 | 
						|
    def addInstrumentation(self, test: TestCase, data: Dict[str, Any]) -> None:
 | 
						|
        append_instrumentation_data(data)
 | 
						|
 | 
						|
    def startTest(self, test: TestCase) -> None:
 | 
						|
        TestResult.startTest(self, test)
 | 
						|
        self.stream.writeln("Running {}".format(full_test_name(test)))  # type: ignore[attr-defined] # https://github.com/python/typeshed/issues/3139
 | 
						|
        self.stream.flush()  # type: ignore[attr-defined] # https://github.com/python/typeshed/issues/3139
 | 
						|
 | 
						|
    def addSuccess(self, *args: Any, **kwargs: Any) -> None:
 | 
						|
        TestResult.addSuccess(self, *args, **kwargs)
 | 
						|
 | 
						|
    def addError(self, *args: Any, **kwargs: Any) -> None:
 | 
						|
        TestResult.addError(self, *args, **kwargs)
 | 
						|
        test_name = full_test_name(args[0])
 | 
						|
        self.failed_tests.append(test_name)
 | 
						|
 | 
						|
    def addFailure(self, *args: Any, **kwargs: Any) -> None:
 | 
						|
        TestResult.addFailure(self, *args, **kwargs)
 | 
						|
        test_name = full_test_name(args[0])
 | 
						|
        self.failed_tests.append(test_name)
 | 
						|
 | 
						|
    def addSkip(self, test: TestCase, reason: str) -> None:
 | 
						|
        TestResult.addSkip(self, test, reason)
 | 
						|
        self.stream.writeln("** Skipping {}: {}".format(  # type: ignore[attr-defined] # https://github.com/python/typeshed/issues/3139
 | 
						|
            full_test_name(test),
 | 
						|
            reason))
 | 
						|
        self.stream.flush()  # type: ignore[attr-defined] # https://github.com/python/typeshed/issues/3139
 | 
						|
 | 
						|
class RemoteTestResult(django_runner.RemoteTestResult):
 | 
						|
    """
 | 
						|
    The class follows the unpythonic style of function names of the
 | 
						|
    base class.
 | 
						|
    """
 | 
						|
 | 
						|
    def addInfo(self, test: TestCase, msg: str) -> None:
 | 
						|
        self.events.append(('addInfo', self.test_index, msg))
 | 
						|
 | 
						|
    def addInstrumentation(self, test: TestCase, data: Dict[str, Any]) -> None:
 | 
						|
        # Some elements of data['info'] cannot be serialized.
 | 
						|
        if 'info' in data:
 | 
						|
            del data['info']
 | 
						|
 | 
						|
        self.events.append(('addInstrumentation', self.test_index, data))
 | 
						|
 | 
						|
def process_instrumented_calls(func: Callable[[Dict[str, Any]], None]) -> None:
 | 
						|
    for call in test_helpers.INSTRUMENTED_CALLS:
 | 
						|
        func(call)
 | 
						|
 | 
						|
SerializedSubsuite = Tuple[Type['TestSuite'], List[str]]
 | 
						|
SubsuiteArgs = Tuple[Type['RemoteTestRunner'], int, SerializedSubsuite, bool]
 | 
						|
 | 
						|
def run_subsuite(args: SubsuiteArgs) -> Tuple[int, Any]:
 | 
						|
    # Reset the accumulated INSTRUMENTED_CALLS before running this subsuite.
 | 
						|
    test_helpers.INSTRUMENTED_CALLS = []
 | 
						|
    # The first argument is the test runner class but we don't need it
 | 
						|
    # because we run our own version of the runner class.
 | 
						|
    _, subsuite_index, subsuite, failfast = args
 | 
						|
    runner = RemoteTestRunner(failfast=failfast)
 | 
						|
    result = runner.run(deserialize_suite(subsuite))
 | 
						|
    # Now we send instrumentation related events. This data will be
 | 
						|
    # appended to the data structure in the main thread. For Mypy,
 | 
						|
    # type of Partial is different from Callable. All the methods of
 | 
						|
    # TestResult are passed TestCase as the first argument but
 | 
						|
    # addInstrumentation does not need it.
 | 
						|
    process_instrumented_calls(partial(result.addInstrumentation, None))
 | 
						|
    return subsuite_index, result.events
 | 
						|
 | 
						|
def destroy_test_databases(worker_id: Optional[int]=None) -> None:
 | 
						|
    for alias in connections:
 | 
						|
        connection = connections[alias]
 | 
						|
        try:
 | 
						|
            # In the parallel mode, the test databases are created
 | 
						|
            # through the N=self.parallel child processes, and in the
 | 
						|
            # parent process (which calls `destroy_test_databases`),
 | 
						|
            # `settings_dict` remains unchanged, with the original
 | 
						|
            # template database name (zulip_test_template).  So to
 | 
						|
            # delete the database zulip_test_template_<number>, we
 | 
						|
            # need to pass `number` to `destroy_test_db`.
 | 
						|
            #
 | 
						|
            # When we run in serial mode (self.parallel=1), we don't
 | 
						|
            # fork and thus both creation and destruction occur in the
 | 
						|
            # same process, which means `settings_dict` has been
 | 
						|
            # updated to have `zulip_test_template_<number>` as its
 | 
						|
            # database name by the creation code.  As a result, to
 | 
						|
            # delete that database, we need to not pass a number
 | 
						|
            # argument to destroy_test_db.
 | 
						|
            if worker_id is not None:
 | 
						|
                """Modified from the Django original to """
 | 
						|
                database_id = get_database_id(worker_id)
 | 
						|
                connection.creation.destroy_test_db(suffix=database_id)
 | 
						|
            else:
 | 
						|
                connection.creation.destroy_test_db()
 | 
						|
        except ProgrammingError:
 | 
						|
            # DB doesn't exist. No need to do anything.
 | 
						|
            pass
 | 
						|
 | 
						|
def create_test_databases(worker_id: int) -> None:
 | 
						|
    database_id = get_database_id(worker_id)
 | 
						|
    for alias in connections:
 | 
						|
        connection = connections[alias]
 | 
						|
        connection.creation.clone_test_db(
 | 
						|
            suffix=database_id,
 | 
						|
            keepdb=True,
 | 
						|
        )
 | 
						|
 | 
						|
        settings_dict = connection.creation.get_test_db_clone_settings(database_id)
 | 
						|
        # connection.settings_dict must be updated in place for changes to be
 | 
						|
        # reflected in django.db.connections. If the following line assigned
 | 
						|
        # connection.settings_dict = settings_dict, new threads would connect
 | 
						|
        # to the default database instead of the appropriate clone.
 | 
						|
        connection.settings_dict.update(settings_dict)
 | 
						|
        connection.close()
 | 
						|
 | 
						|
def init_worker(counter: Synchronized) -> None:
 | 
						|
    """
 | 
						|
    This function runs only under parallel mode. It initializes the
 | 
						|
    individual processes which are also called workers.
 | 
						|
    """
 | 
						|
    global _worker_id
 | 
						|
 | 
						|
    with counter.get_lock():
 | 
						|
        counter.value += 1
 | 
						|
        _worker_id = counter.value
 | 
						|
 | 
						|
    """
 | 
						|
    You can now use _worker_id.
 | 
						|
    """
 | 
						|
 | 
						|
    # Clear the cache
 | 
						|
    from zerver.lib.cache import get_cache_backend
 | 
						|
    cache = get_cache_backend(None)
 | 
						|
    cache.clear()
 | 
						|
 | 
						|
    # Close all connections
 | 
						|
    connections.close_all()
 | 
						|
 | 
						|
    destroy_test_databases(_worker_id)
 | 
						|
    create_test_databases(_worker_id)
 | 
						|
    initialize_worker_path(_worker_id)
 | 
						|
 | 
						|
    def is_upload_avatar_url(url: URLPattern) -> bool:
 | 
						|
        if url.pattern.regex.pattern == r'^user_avatars/(?P<path>.*)$':
 | 
						|
            return True
 | 
						|
        return False
 | 
						|
 | 
						|
    # We manually update the upload directory path in the url regex.
 | 
						|
    from zproject import dev_urls
 | 
						|
    found = False
 | 
						|
    for url in dev_urls.urls:
 | 
						|
        if is_upload_avatar_url(url):
 | 
						|
            found = True
 | 
						|
            new_root = os.path.join(settings.LOCAL_UPLOADS_DIR, "avatars")
 | 
						|
            url.default_args['document_root'] = new_root
 | 
						|
 | 
						|
    if not found:
 | 
						|
        print("*** Upload directory not found.")
 | 
						|
 | 
						|
class TestSuite(unittest.TestSuite):
 | 
						|
    def run(self, result: TestResult, debug: Optional[bool]=False) -> TestResult:
 | 
						|
        """
 | 
						|
        This function mostly contains the code from
 | 
						|
        unittest.TestSuite.run. The need to override this function
 | 
						|
        occurred because we use run_test to run the testcase.
 | 
						|
        """
 | 
						|
        topLevel = False
 | 
						|
        if getattr(result, '_testRunEntered', False) is False:
 | 
						|
            result._testRunEntered = topLevel = True  # type: ignore[attr-defined]
 | 
						|
 | 
						|
        for test in self:
 | 
						|
            # but this is correct. Taken from unittest.
 | 
						|
            if result.shouldStop:
 | 
						|
                break
 | 
						|
 | 
						|
            if isinstance(test, TestSuite):
 | 
						|
                test.run(result, debug=debug)
 | 
						|
            else:
 | 
						|
                self._tearDownPreviousClass(test, result)  # type: ignore[attr-defined]
 | 
						|
                self._handleModuleFixture(test, result)  # type: ignore[attr-defined]
 | 
						|
                self._handleClassSetUp(test, result)  # type: ignore[attr-defined]
 | 
						|
                result._previousTestClass = test.__class__  # type: ignore[attr-defined]
 | 
						|
                if (getattr(test.__class__, '_classSetupFailed', False) or
 | 
						|
                        getattr(result, '_moduleSetUpFailed', False)):
 | 
						|
                    continue
 | 
						|
 | 
						|
                failed = run_test(test, result)
 | 
						|
                if failed or result.shouldStop:
 | 
						|
                    result.shouldStop = True
 | 
						|
                    break
 | 
						|
 | 
						|
        if topLevel:
 | 
						|
            self._tearDownPreviousClass(None, result)  # type: ignore[attr-defined]
 | 
						|
            self._handleModuleTearDown(result)  # type: ignore[attr-defined]
 | 
						|
            result._testRunEntered = False  # type: ignore[attr-defined]
 | 
						|
        return result
 | 
						|
 | 
						|
class TestLoader(loader.TestLoader):
 | 
						|
    suiteClass = TestSuite
 | 
						|
 | 
						|
class ParallelTestSuite(django_runner.ParallelTestSuite):
 | 
						|
    run_subsuite = run_subsuite
 | 
						|
    init_worker = init_worker
 | 
						|
 | 
						|
    def __init__(self, suite: TestSuite, processes: int, failfast: bool) -> None:
 | 
						|
        super().__init__(suite, processes, failfast)
 | 
						|
        # We can't specify a consistent type for self.subsuites, since
 | 
						|
        # the whole idea here is to monkey-patch that so we can use
 | 
						|
        # most of django_runner.ParallelTestSuite with our own suite
 | 
						|
        # definitions.
 | 
						|
        self.subsuites = SubSuiteList(self.subsuites)  # type: ignore[has-type] # Type of self.subsuites changes.
 | 
						|
 | 
						|
def check_import_error(test_name: str) -> None:
 | 
						|
    try:
 | 
						|
        # Directly using __import__ is not recommended, but here it gives
 | 
						|
        # clearer traceback as compared to importlib.import_module.
 | 
						|
        __import__(test_name)
 | 
						|
    except ImportError as exc:
 | 
						|
        raise exc from exc  # Disable exception chaining in Python 3.
 | 
						|
 | 
						|
 | 
						|
def initialize_worker_path(worker_id: int) -> None:
 | 
						|
    # Allow each test worker process to write to a unique directory
 | 
						|
    # within `TEST_RUN_DIR`.
 | 
						|
    worker_path = os.path.join(TEST_RUN_DIR, 'worker_{}'.format(_worker_id))
 | 
						|
    os.makedirs(worker_path, exist_ok=True)
 | 
						|
    settings.TEST_WORKER_DIR = worker_path
 | 
						|
 | 
						|
    # Every process should upload to a separate directory so that
 | 
						|
    # race conditions can be avoided.
 | 
						|
    settings.LOCAL_UPLOADS_DIR = get_or_create_dev_uuid_var_path(
 | 
						|
        os.path.join("test-backend",
 | 
						|
                     os.path.basename(TEST_RUN_DIR),
 | 
						|
                     os.path.basename(worker_path),
 | 
						|
                     "test_uploads"))
 | 
						|
 | 
						|
class Runner(DiscoverRunner):
 | 
						|
    test_suite = TestSuite
 | 
						|
    test_loader = TestLoader()
 | 
						|
    parallel_test_suite = ParallelTestSuite
 | 
						|
 | 
						|
    def __init__(self, *args: Any, **kwargs: Any) -> None:
 | 
						|
        DiscoverRunner.__init__(self, *args, **kwargs)
 | 
						|
 | 
						|
        # `templates_rendered` holds templates which were rendered
 | 
						|
        # in proper logical tests.
 | 
						|
        self.templates_rendered: Set[str] = set()
 | 
						|
        # `shallow_tested_templates` holds templates which were rendered
 | 
						|
        # in `zerver.tests.test_templates`.
 | 
						|
        self.shallow_tested_templates: Set[str] = set()
 | 
						|
        template_rendered.connect(self.on_template_rendered)
 | 
						|
 | 
						|
    def get_resultclass(self) -> Type[TestResult]:
 | 
						|
        return TextTestResult
 | 
						|
 | 
						|
    def on_template_rendered(self, sender: Any, context: Dict[str, Any], **kwargs: Any) -> None:
 | 
						|
        if hasattr(sender, 'template'):
 | 
						|
            template_name = sender.template.name
 | 
						|
            if template_name not in self.templates_rendered:
 | 
						|
                if context.get('shallow_tested') and template_name not in self.templates_rendered:
 | 
						|
                    self.shallow_tested_templates.add(template_name)
 | 
						|
                else:
 | 
						|
                    self.templates_rendered.add(template_name)
 | 
						|
                    self.shallow_tested_templates.discard(template_name)
 | 
						|
 | 
						|
    def get_shallow_tested_templates(self) -> Set[str]:
 | 
						|
        return self.shallow_tested_templates
 | 
						|
 | 
						|
    def setup_test_environment(self, *args: Any, **kwargs: Any) -> Any:
 | 
						|
        settings.DATABASES['default']['NAME'] = settings.BACKEND_DATABASE_TEMPLATE
 | 
						|
        # We create/destroy the test databases in run_tests to avoid
 | 
						|
        # duplicate work when running in parallel mode.
 | 
						|
 | 
						|
        # Write the template database ids to a file that we can
 | 
						|
        # reference for cleaning them up if they leak.
 | 
						|
        filepath = os.path.join(get_dev_uuid_var_path(),
 | 
						|
                                TEMPLATE_DATABASE_DIR,
 | 
						|
                                get_database_id())
 | 
						|
        os.makedirs(os.path.dirname(filepath), exist_ok=True)
 | 
						|
        with open(filepath, "w") as f:
 | 
						|
            if self.parallel > 1:
 | 
						|
                for index in range(self.parallel):
 | 
						|
                    f.write(get_database_id(index + 1) + "\n")
 | 
						|
            else:
 | 
						|
                f.write(get_database_id() + "\n")
 | 
						|
 | 
						|
        # Check if we are in serial mode to avoid unnecessarily making a directory.
 | 
						|
        # We add "worker_0" in the path for consistency with parallel mode.
 | 
						|
        if self.parallel == 1:
 | 
						|
            initialize_worker_path(0)
 | 
						|
 | 
						|
        return super().setup_test_environment(*args, **kwargs)
 | 
						|
 | 
						|
    def teardown_test_environment(self, *args: Any, **kwargs: Any) -> Any:
 | 
						|
        # The test environment setup clones the zulip_test_template
 | 
						|
        # database, creating databases with names:
 | 
						|
        #     'zulip_test_template_N_<worker_id>',
 | 
						|
        # where N is `random_id_range_start`, and `worker_id` is a
 | 
						|
        # value between <1, self.parallel>.
 | 
						|
        #
 | 
						|
        # We need to delete those databases to avoid leaking disk
 | 
						|
        # (Django is smart and calls this on SIGINT too).
 | 
						|
        if self.parallel > 1:
 | 
						|
            for index in range(self.parallel):
 | 
						|
                destroy_test_databases(index + 1)
 | 
						|
        else:
 | 
						|
            destroy_test_databases()
 | 
						|
 | 
						|
        # Clean up our record of which databases this process created.
 | 
						|
        filepath = os.path.join(get_dev_uuid_var_path(),
 | 
						|
                                TEMPLATE_DATABASE_DIR,
 | 
						|
                                get_database_id())
 | 
						|
        os.remove(filepath)
 | 
						|
 | 
						|
        # Clean up our test runs root directory.
 | 
						|
        try:
 | 
						|
            shutil.rmtree(TEST_RUN_DIR)
 | 
						|
        except OSError:
 | 
						|
            print("Unable to clean up the test run's directory.")
 | 
						|
        return super().teardown_test_environment(*args, **kwargs)
 | 
						|
 | 
						|
    def test_imports(self, test_labels: List[str], suite: Union[TestSuite, ParallelTestSuite]) -> None:
 | 
						|
        prefix_old = 'unittest.loader.ModuleImportFailure.'  # Python <= 3.4
 | 
						|
        prefix_new = 'unittest.loader._FailedTest.'  # Python > 3.4
 | 
						|
        error_prefixes = [prefix_old, prefix_new]
 | 
						|
        for test_name in get_test_names(suite):
 | 
						|
            for prefix in error_prefixes:
 | 
						|
                if test_name.startswith(prefix):
 | 
						|
                    test_name = test_name[len(prefix):]
 | 
						|
                    for label in test_labels:
 | 
						|
                        # This code block is for Python 3.5 when test label is
 | 
						|
                        # directly provided, for example:
 | 
						|
                        # ./tools/test-backend zerver.tests.test_alert_words.py
 | 
						|
                        #
 | 
						|
                        # In this case, the test name is of this form:
 | 
						|
                        # 'unittest.loader._FailedTest.test_alert_words'
 | 
						|
                        #
 | 
						|
                        # Whereas check_import_error requires test names of
 | 
						|
                        # this form:
 | 
						|
                        # 'unittest.loader._FailedTest.zerver.tests.test_alert_words'.
 | 
						|
                        if test_name in label:
 | 
						|
                            test_name = label
 | 
						|
                            break
 | 
						|
                    check_import_error(test_name)
 | 
						|
 | 
						|
    def run_tests(self, test_labels: List[str],
 | 
						|
                  extra_tests: Optional[List[TestCase]]=None,
 | 
						|
                  full_suite: bool=False,
 | 
						|
                  include_webhooks: bool=False,
 | 
						|
                  **kwargs: Any) -> Tuple[bool, List[str]]:
 | 
						|
        self.setup_test_environment()
 | 
						|
        try:
 | 
						|
            suite = self.build_suite(test_labels, extra_tests)
 | 
						|
        except AttributeError:
 | 
						|
            # We are likely to get here only when running tests in serial
 | 
						|
            # mode on Python 3.4 or lower.
 | 
						|
            # test_labels are always normalized to include the correct prefix.
 | 
						|
            # If we run the command with ./tools/test-backend test_alert_words,
 | 
						|
            # test_labels will be equal to ['zerver.tests.test_alert_words'].
 | 
						|
            for test_label in test_labels:
 | 
						|
                check_import_error(test_label)
 | 
						|
 | 
						|
            # I think we won't reach this line under normal circumstances, but
 | 
						|
            # for some unforeseen scenario in which the AttributeError was not
 | 
						|
            # caused by an import error, let's re-raise the exception for
 | 
						|
            # debugging purposes.
 | 
						|
            raise
 | 
						|
 | 
						|
        self.test_imports(test_labels, suite)
 | 
						|
        if self.parallel == 1:
 | 
						|
            # We are running in serial mode so create the databases here.
 | 
						|
            # For parallel mode, the databases are created in init_worker.
 | 
						|
            # We don't want to create and destroy DB in setup_test_environment
 | 
						|
            # because it will be called for both serial and parallel modes.
 | 
						|
            # However, at this point we know in which mode we would be running
 | 
						|
            # since that decision has already been made in build_suite().
 | 
						|
            #
 | 
						|
            # We pass a _worker_id, which in this code path is always 0
 | 
						|
            destroy_test_databases(_worker_id)
 | 
						|
            create_test_databases(_worker_id)
 | 
						|
 | 
						|
        # We have to do the next line to avoid flaky scenarios where we
 | 
						|
        # run a single test and getting an SA connection causes data from
 | 
						|
        # a Django connection to be rolled back mid-test.
 | 
						|
        get_sqlalchemy_connection()
 | 
						|
        result = self.run_suite(suite)
 | 
						|
        self.teardown_test_environment()
 | 
						|
        failed = self.suite_result(suite, result)
 | 
						|
        if not failed:
 | 
						|
            write_instrumentation_reports(full_suite=full_suite, include_webhooks=include_webhooks)
 | 
						|
        return failed, result.failed_tests
 | 
						|
 | 
						|
def get_test_names(suite: Union[TestSuite, ParallelTestSuite]) -> List[str]:
 | 
						|
    if isinstance(suite, ParallelTestSuite):
 | 
						|
        # suite is ParallelTestSuite. It will have a subsuites parameter of
 | 
						|
        # type SubSuiteList. Each element of a SubsuiteList is a tuple whose
 | 
						|
        # first element is the type of TestSuite and the second element is a
 | 
						|
        # list of test names in that test suite. See serialize_suite() for the
 | 
						|
        # implementation details.
 | 
						|
        return [name for subsuite in suite.subsuites for name in subsuite[1]]
 | 
						|
    else:
 | 
						|
        return [full_test_name(t) for t in get_tests_from_suite(suite)]
 | 
						|
 | 
						|
def get_tests_from_suite(suite: unittest.TestSuite) -> TestCase:
 | 
						|
    for test in suite:
 | 
						|
        if isinstance(test, TestSuite):
 | 
						|
            yield from get_tests_from_suite(test)
 | 
						|
        else:
 | 
						|
            yield test
 | 
						|
 | 
						|
def serialize_suite(suite: TestSuite) -> Tuple[Type[TestSuite], List[str]]:
 | 
						|
    return type(suite), get_test_names(suite)
 | 
						|
 | 
						|
def deserialize_suite(args: Tuple[Type[TestSuite], List[str]]) -> TestSuite:
 | 
						|
    suite_class, test_names = args
 | 
						|
    suite = suite_class()
 | 
						|
    tests = TestLoader().loadTestsFromNames(test_names)
 | 
						|
    for test in get_tests_from_suite(tests):
 | 
						|
        suite.addTest(test)
 | 
						|
    return suite
 | 
						|
 | 
						|
class RemoteTestRunner(django_runner.RemoteTestRunner):
 | 
						|
    resultclass = RemoteTestResult
 | 
						|
 | 
						|
class SubSuiteList(List[Tuple[Type[TestSuite], List[str]]]):
 | 
						|
    """
 | 
						|
    This class allows us to avoid changing the main logic of
 | 
						|
    ParallelTestSuite and still make it serializable.
 | 
						|
    """
 | 
						|
 | 
						|
    def __init__(self, suites: List[TestSuite]) -> None:
 | 
						|
        serialized_suites = [serialize_suite(s) for s in suites]
 | 
						|
        super().__init__(serialized_suites)
 | 
						|
 | 
						|
    def __getitem__(self, index: Any) -> Any:
 | 
						|
        suite = super().__getitem__(index)
 | 
						|
        return deserialize_suite(suite)
 |