Files
zulip/tools/test-queue-worker-reload
Anders Kaseorg 365fe0b3d5 python: Sort imports with isort.
Fixes #2665.

Regenerated by tabbott with `lint --fix` after a rebase and change in
parameters.

Note from tabbott: In a few cases, this converts technical debt in the
form of unsorted imports into different technical debt in the form of
our largest files having very long, ugly import sequences at the
start.  I expect this change will increase pressure for us to split
those files, which isn't a bad thing.

Signed-off-by: Anders Kaseorg <anders@zulip.com>
2020-06-11 16:45:32 -07:00

91 lines
2.8 KiB
Python
Executable File

#!/usr/bin/env python3
import os
import signal
import subprocess
import sys
import time
import types
# check for the venv
from lib import sanity_check
sanity_check.check_venv(__file__)
# TODO: Convert this to use scripts/lib/queue_workers.py
TOOLS_DIR = os.path.dirname(os.path.abspath(__file__))
successful_worker_launch = '[process_queue] 16 queue worker threads were launched\n'
def check_worker_launch(run_dev: "subprocess.Popen[str]") -> bool:
failed = False
i = 0
def on_timer(signum: int, frame: types.FrameType) -> None:
nonlocal failed, i
sys.stdout.write('.')
sys.stdout.flush()
i += 1
if i == 200:
failed = True
run_dev.send_signal(signal.SIGINT)
signal.setitimer(signal.ITIMER_REAL, 0, 0)
log_output = []
print("Polling run-dev", end='')
# Attempt to poll the log file for 60 sec. to see if all worker threads are launched.
old_handler = signal.signal(signal.SIGALRM, on_timer)
signal.setitimer(signal.ITIMER_REAL, 0.3, 0.3)
assert run_dev.stdout is not None
for line in run_dev.stdout:
log_output.append(line)
if line.endswith(successful_worker_launch):
break
else:
failed = True
signal.setitimer(signal.ITIMER_REAL, 0, 0)
signal.signal(signal.SIGALRM, old_handler)
sys.stdout.write('\n')
if not failed:
print('Worker threads launched successfully')
else:
print('Error in server startup. Dumping logs')
print(''.join(log_output))
return failed
if __name__ == '__main__':
print('\nStarting Development Server')
args = [f"{TOOLS_DIR}/run-dev.py"]
run_dev = subprocess.Popen(
args,
bufsize=1, # line buffered
stdout=subprocess.PIPE, stderr=subprocess.STDOUT,
universal_newlines=True)
failed = check_worker_launch(run_dev)
if failed:
run_dev.send_signal(signal.SIGINT)
run_dev.wait()
sys.exit(1)
# In dev. environment, queues are run through Django's autoreload code. The
# autoreload code of Django works by looping over the files associated with
# all the loaded modules. This loop is run after every 1 second. If the
# file is found for the first time by the loop, it is assumed that the
# file is new and is not modified between the time it is loaded and is
# checked by the loop. This assumption is the source of a race condition.
# We can either implement a more sensitive version of the loop or we can
# just allow enough time to the Django loop to touch every file at least
# once.
time.sleep(1.3)
print("Attempting to modify a file")
os.utime('zerver/lib/actions.py')
failed = check_worker_launch(run_dev)
run_dev.send_signal(signal.SIGINT)
run_dev.wait()
if failed:
sys.exit(1)