Files
zulip/tools/test-queue-worker-reload
Mateusz Mandera dd40649e04 queue_processors: Remove the slow_queries queue.
While this functionality to post slow queries to a Zulip stream was
very useful in the early days of Zulip, when there were only a few
hundred accounts, it's long since been useless since (1) the total
request volume on larger Zulip servers run by Zulip developers, and
(2) other server operators don't want real-time notifications of slow
backend queries.  The right structure for this is just a log file.

We get rid of the queue and replace it with a "zulip.slow_queries"
logger, which will still log to /var/log/zulip/slow_queries.log for
ease of access to this information and propagate to the other logging
handlers.  Reducing the amount of queues is good for lowering zulip's
memory footprint and restart performance, since we run at least one
dedicated queue worker process for each one in most configurations.
2020-05-11 00:45:13 -07:00

92 lines
2.8 KiB
Python
Executable File

#!/usr/bin/env python3
import os
import sys
import time
import signal
import subprocess
import types
# check for the venv
from lib import sanity_check
sanity_check.check_venv(__file__)
# TODO: Convert this to use scripts/lib/queue_workers.py
TOOLS_DIR = os.path.dirname(os.path.abspath(__file__))
successful_worker_launch = '[process_queue] 16 queue worker threads were launched\n'
def check_worker_launch(run_dev: "subprocess.Popen[str]") -> bool:
failed = False
i = 0
def on_timer(signum: int, frame: types.FrameType) -> None:
nonlocal failed, i
sys.stdout.write('.')
sys.stdout.flush()
i += 1
if i == 200:
failed = True
run_dev.send_signal(signal.SIGINT)
signal.setitimer(signal.ITIMER_REAL, 0, 0)
log_output = []
print("Polling run-dev", end='')
# Attempt to poll the log file for 60 sec. to see if all worker threads are launched.
old_handler = signal.signal(signal.SIGALRM, on_timer)
signal.setitimer(signal.ITIMER_REAL, 0.3, 0.3)
assert run_dev.stdout is not None
for line in run_dev.stdout:
log_output.append(line)
if line.endswith(successful_worker_launch):
break
else:
failed = True
signal.setitimer(signal.ITIMER_REAL, 0, 0)
signal.signal(signal.SIGALRM, old_handler)
sys.stdout.write('\n')
if not failed:
print('Worker threads launched successfully')
else:
print('Error in server startup. Dumping logs')
print(''.join(log_output))
return failed
if __name__ == '__main__':
print('\nStarting Development Server')
args = ["{}/run-dev.py".format(TOOLS_DIR)]
run_dev = subprocess.Popen(
args,
bufsize=1, # line buffered
stdout=subprocess.PIPE, stderr=subprocess.STDOUT,
universal_newlines=True)
failed = check_worker_launch(run_dev)
if failed:
run_dev.send_signal(signal.SIGINT)
run_dev.wait()
sys.exit(1)
# In dev. environment, queues are run through Django's autoreload code. The
# autoreload code of Django works by looping over the files associated with
# all the loaded modules. This loop is run after every 1 second. If the
# file is found for the first time by the loop, it is assumed that the
# file is new and is not modified between the time it is loaded and is
# checked by the loop. This assumption is the source of a race condition.
# We can either implement a more sensitive version of the loop or we can
# just allow enough time to the Django loop to touch every file at least
# once.
time.sleep(1.3)
print("Attempting to modify a file")
os.utime('zerver/lib/actions.py')
failed = check_worker_launch(run_dev)
run_dev.send_signal(signal.SIGINT)
run_dev.wait()
if failed:
sys.exit(1)