mirror of
https://github.com/zulip/zulip.git
synced 2025-10-24 00:23:49 +00:00
Ever since we started bundling the app with webpack, there’s been less and less overlap between our ‘static’ directory (files belonging to the frontend app) and Django’s interpretation of the ‘static’ directory (files served directly to the web). Split the app out to its own ‘web’ directory outside of ‘static’, and remove all the custom collectstatic --ignore rules. This makes it much clearer what’s actually being served to the web, and what’s being bundled by webpack. It also shrinks the release tarball by 3%. Signed-off-by: Anders Kaseorg <anders@zulip.com>
394 lines
12 KiB
Python
Executable File
394 lines
12 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
import argparse
|
|
import glob
|
|
import os
|
|
import shutil
|
|
import sys
|
|
from typing import List
|
|
|
|
ZULIP_PATH = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
|
|
|
sys.path.append(ZULIP_PATH)
|
|
import pygments
|
|
|
|
from scripts.lib import clean_unused_caches
|
|
from scripts.lib.zulip_tools import (
|
|
ENDC,
|
|
OKBLUE,
|
|
get_dev_uuid_var_path,
|
|
get_tzdata_zi,
|
|
is_digest_obsolete,
|
|
run,
|
|
run_as_root,
|
|
write_new_digest,
|
|
)
|
|
from tools.setup.generate_zulip_bots_static_files import generate_zulip_bots_static_files
|
|
from version import PROVISION_VERSION
|
|
|
|
VENV_PATH = "/srv/zulip-py3-venv"
|
|
UUID_VAR_PATH = get_dev_uuid_var_path()
|
|
|
|
with get_tzdata_zi() as f:
|
|
line = f.readline()
|
|
assert line.startswith("# version ")
|
|
timezones_version = line[len("# version ") :]
|
|
|
|
|
|
def create_var_directories() -> None:
|
|
# create var/coverage, var/log, etc.
|
|
var_dir = os.path.join(ZULIP_PATH, "var")
|
|
sub_dirs = [
|
|
"coverage",
|
|
"log",
|
|
"node-coverage",
|
|
"test_uploads",
|
|
"uploads",
|
|
"xunit-test-results",
|
|
]
|
|
for sub_dir in sub_dirs:
|
|
path = os.path.join(var_dir, sub_dir)
|
|
os.makedirs(path, exist_ok=True)
|
|
|
|
|
|
def build_pygments_data_paths() -> List[str]:
|
|
paths = [
|
|
"tools/setup/build_pygments_data",
|
|
"tools/setup/lang.json",
|
|
]
|
|
return paths
|
|
|
|
|
|
def build_timezones_data_paths() -> List[str]:
|
|
paths = [
|
|
"tools/setup/build_timezone_values",
|
|
]
|
|
return paths
|
|
|
|
|
|
def compilemessages_paths() -> List[str]:
|
|
paths = ["zerver/management/commands/compilemessages.py"]
|
|
paths += glob.glob("locale/*/LC_MESSAGES/*.po")
|
|
paths += glob.glob("locale/*/translations.json")
|
|
return paths
|
|
|
|
|
|
def inline_email_css_paths() -> List[str]:
|
|
paths = [
|
|
"scripts/setup/inline_email_css.py",
|
|
"templates/zerver/emails/email.css",
|
|
]
|
|
paths += glob.glob("templates/zerver/emails/*.source.html")
|
|
return paths
|
|
|
|
|
|
def configure_rabbitmq_paths() -> List[str]:
|
|
paths = [
|
|
"scripts/setup/configure-rabbitmq",
|
|
]
|
|
return paths
|
|
|
|
|
|
def setup_shell_profile(shell_profile: str) -> None:
|
|
shell_profile_path = os.path.expanduser(shell_profile)
|
|
|
|
def write_command(command: str) -> None:
|
|
if os.path.exists(shell_profile_path):
|
|
with open(shell_profile_path) as shell_profile_file:
|
|
lines = [line.strip() for line in shell_profile_file.readlines()]
|
|
if command not in lines:
|
|
with open(shell_profile_path, "a+") as shell_profile_file:
|
|
shell_profile_file.writelines(command + "\n")
|
|
else:
|
|
with open(shell_profile_path, "w") as shell_profile_file:
|
|
shell_profile_file.writelines(command + "\n")
|
|
|
|
source_activate_command = "source " + os.path.join(VENV_PATH, "bin", "activate")
|
|
write_command(source_activate_command)
|
|
if os.path.exists("/srv/zulip"):
|
|
write_command("cd /srv/zulip")
|
|
|
|
|
|
def setup_bash_profile() -> None:
|
|
"""Select a bash profile file to add setup code to."""
|
|
|
|
BASH_PROFILES = [
|
|
os.path.expanduser(p) for p in ("~/.bash_profile", "~/.bash_login", "~/.profile")
|
|
]
|
|
|
|
def clear_old_profile() -> None:
|
|
# An earlier version of this script would output a fresh .bash_profile
|
|
# even though a .profile existed in the image used. As a convenience to
|
|
# existing developers (and, perhaps, future developers git-bisecting the
|
|
# provisioning scripts), check for this situation, and blow away the
|
|
# created .bash_profile if one is found.
|
|
|
|
BASH_PROFILE = BASH_PROFILES[0]
|
|
DOT_PROFILE = BASH_PROFILES[2]
|
|
OLD_PROFILE_TEXT = "source /srv/zulip-py3-venv/bin/activate\ncd /srv/zulip\n"
|
|
|
|
if os.path.exists(DOT_PROFILE):
|
|
try:
|
|
with open(BASH_PROFILE) as f:
|
|
profile_contents = f.read()
|
|
if profile_contents == OLD_PROFILE_TEXT:
|
|
os.unlink(BASH_PROFILE)
|
|
except FileNotFoundError:
|
|
pass
|
|
|
|
clear_old_profile()
|
|
|
|
for candidate_profile in BASH_PROFILES:
|
|
if os.path.exists(candidate_profile):
|
|
setup_shell_profile(candidate_profile)
|
|
break
|
|
else:
|
|
# no existing bash profile found; claim .bash_profile
|
|
setup_shell_profile(BASH_PROFILES[0])
|
|
|
|
|
|
def need_to_run_build_pygments_data() -> bool:
|
|
if not os.path.exists("web/generated/pygments_data.json"):
|
|
return True
|
|
|
|
return is_digest_obsolete(
|
|
"build_pygments_data_hash",
|
|
build_pygments_data_paths(),
|
|
[pygments.__version__],
|
|
)
|
|
|
|
|
|
def need_to_run_build_timezone_data() -> bool:
|
|
if not os.path.exists("web/generated/timezones.json"):
|
|
return True
|
|
|
|
return is_digest_obsolete(
|
|
"build_timezones_data_hash",
|
|
build_timezones_data_paths(),
|
|
[timezones_version],
|
|
)
|
|
|
|
|
|
def need_to_run_compilemessages() -> bool:
|
|
if not os.path.exists("locale/language_name_map.json"):
|
|
# User may have cleaned their Git checkout.
|
|
print("Need to run compilemessages due to missing language_name_map.json")
|
|
return True
|
|
|
|
return is_digest_obsolete(
|
|
"last_compilemessages_hash",
|
|
compilemessages_paths(),
|
|
)
|
|
|
|
|
|
def need_to_run_inline_email_css() -> bool:
|
|
if not os.path.exists("templates/zerver/emails/compiled/"):
|
|
return True
|
|
|
|
return is_digest_obsolete(
|
|
"last_email_source_files_hash",
|
|
inline_email_css_paths(),
|
|
)
|
|
|
|
|
|
def need_to_run_configure_rabbitmq(settings_list: List[str]) -> bool:
|
|
obsolete = is_digest_obsolete(
|
|
"last_configure_rabbitmq_hash",
|
|
configure_rabbitmq_paths(),
|
|
settings_list,
|
|
)
|
|
|
|
if obsolete:
|
|
return True
|
|
|
|
try:
|
|
from zerver.lib.queue import SimpleQueueClient
|
|
|
|
SimpleQueueClient()
|
|
return False
|
|
except Exception:
|
|
return True
|
|
|
|
|
|
def main(options: argparse.Namespace) -> int:
|
|
setup_bash_profile()
|
|
setup_shell_profile("~/.zprofile")
|
|
|
|
# This needs to happen before anything that imports zproject.settings.
|
|
run(["scripts/setup/generate_secrets.py", "--development"])
|
|
|
|
create_var_directories()
|
|
|
|
# The `build_emoji` script requires `emoji-datasource` package
|
|
# which we install via npm; thus this step is after installing npm
|
|
# packages.
|
|
run(["tools/setup/emoji/build_emoji"])
|
|
|
|
# copy over static files from the zulip_bots package
|
|
generate_zulip_bots_static_files()
|
|
|
|
if options.is_force or need_to_run_build_pygments_data():
|
|
run(["tools/setup/build_pygments_data"])
|
|
write_new_digest(
|
|
"build_pygments_data_hash",
|
|
build_pygments_data_paths(),
|
|
[pygments.__version__],
|
|
)
|
|
else:
|
|
print("No need to run `tools/setup/build_pygments_data`.")
|
|
|
|
if options.is_force or need_to_run_build_timezone_data():
|
|
run(["tools/setup/build_timezone_values"])
|
|
write_new_digest(
|
|
"build_timezones_data_hash",
|
|
build_timezones_data_paths(),
|
|
[timezones_version],
|
|
)
|
|
else:
|
|
print("No need to run `tools/setup/build_timezone_values`.")
|
|
|
|
if options.is_force or need_to_run_inline_email_css():
|
|
run(["scripts/setup/inline_email_css.py"])
|
|
write_new_digest(
|
|
"last_email_source_files_hash",
|
|
inline_email_css_paths(),
|
|
)
|
|
else:
|
|
print("No need to run `scripts/setup/inline_email_css.py`.")
|
|
|
|
if not options.is_build_release_tarball_only:
|
|
# The following block is skipped when we just need the development
|
|
# environment to build a release tarball.
|
|
|
|
# Need to set up Django before using template_status
|
|
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "zproject.settings")
|
|
import django
|
|
|
|
django.setup()
|
|
|
|
from django.conf import settings
|
|
|
|
from zerver.lib.test_fixtures import (
|
|
DEV_DATABASE,
|
|
TEST_DATABASE,
|
|
destroy_leaked_test_databases,
|
|
)
|
|
|
|
assert settings.RABBITMQ_PASSWORD is not None
|
|
if options.is_force or need_to_run_configure_rabbitmq([settings.RABBITMQ_PASSWORD]):
|
|
run_as_root(["scripts/setup/configure-rabbitmq"])
|
|
write_new_digest(
|
|
"last_configure_rabbitmq_hash",
|
|
configure_rabbitmq_paths(),
|
|
[settings.RABBITMQ_PASSWORD],
|
|
)
|
|
else:
|
|
print("No need to run `scripts/setup/configure-rabbitmq.")
|
|
|
|
dev_template_db_status = DEV_DATABASE.template_status()
|
|
if options.is_force or dev_template_db_status == "needs_rebuild":
|
|
run(["tools/setup/postgresql-init-dev-db"])
|
|
if options.skip_dev_db_build:
|
|
# We don't need to build the manual development
|
|
# database on continuous integration for running tests, so we can
|
|
# just leave it as a template db and save a minute.
|
|
#
|
|
# Important: We don't write a digest as that would
|
|
# incorrectly claim that we ran migrations.
|
|
pass
|
|
else:
|
|
run(["tools/rebuild-dev-database"])
|
|
DEV_DATABASE.write_new_db_digest()
|
|
elif dev_template_db_status == "run_migrations":
|
|
DEV_DATABASE.run_db_migrations()
|
|
elif dev_template_db_status == "current":
|
|
print("No need to regenerate the dev DB.")
|
|
|
|
test_template_db_status = TEST_DATABASE.template_status()
|
|
if options.is_force or test_template_db_status == "needs_rebuild":
|
|
run(["tools/setup/postgresql-init-test-db"])
|
|
run(["tools/rebuild-test-database"])
|
|
TEST_DATABASE.write_new_db_digest()
|
|
elif test_template_db_status == "run_migrations":
|
|
TEST_DATABASE.run_db_migrations()
|
|
elif test_template_db_status == "current":
|
|
print("No need to regenerate the test DB.")
|
|
|
|
if options.is_force or need_to_run_compilemessages():
|
|
run(["./manage.py", "compilemessages"])
|
|
write_new_digest(
|
|
"last_compilemessages_hash",
|
|
compilemessages_paths(),
|
|
)
|
|
else:
|
|
print("No need to run `manage.py compilemessages`.")
|
|
|
|
destroyed = destroy_leaked_test_databases()
|
|
if destroyed:
|
|
print(f"Dropped {destroyed} stale test databases!")
|
|
|
|
clean_unused_caches.main(
|
|
argparse.Namespace(
|
|
threshold_days=6,
|
|
# The defaults here should match parse_cache_script_args in zulip_tools.py
|
|
dry_run=False,
|
|
verbose=False,
|
|
no_headings=True,
|
|
)
|
|
)
|
|
|
|
# Keeping this cache file around can cause eslint to throw
|
|
# random TypeErrors when new/updated dependencies are added
|
|
if os.path.isfile(".eslintcache"):
|
|
# Remove this block when
|
|
# https://github.com/eslint/eslint/issues/11639 is fixed
|
|
# upstream.
|
|
os.remove(".eslintcache")
|
|
|
|
# Clean up the root of the `var/` directory for various
|
|
# testing-related files that we have migrated to
|
|
# `var/<uuid>/test-backend`.
|
|
print("Cleaning var/ directory files...")
|
|
var_paths = glob.glob("var/test*")
|
|
var_paths.append("var/bot_avatar")
|
|
for path in var_paths:
|
|
try:
|
|
if os.path.isdir(path):
|
|
shutil.rmtree(path)
|
|
else:
|
|
os.remove(path)
|
|
except FileNotFoundError:
|
|
pass
|
|
|
|
version_file = os.path.join(UUID_VAR_PATH, "provision_version")
|
|
print(f"writing to {version_file}\n")
|
|
with open(version_file, "w") as f:
|
|
f.write(".".join(map(str, PROVISION_VERSION)) + "\n")
|
|
|
|
print()
|
|
print(OKBLUE + "Zulip development environment setup succeeded!" + ENDC)
|
|
return 0
|
|
|
|
|
|
if __name__ == "__main__":
|
|
parser = argparse.ArgumentParser()
|
|
parser.add_argument(
|
|
"--force",
|
|
action="store_true",
|
|
dest="is_force",
|
|
help="Ignore all provisioning optimizations.",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--build-release-tarball-only",
|
|
action="store_true",
|
|
dest="is_build_release_tarball_only",
|
|
help="Provision for test suite with production settings.",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--skip-dev-db-build", action="store_true", help="Don't run migrations on dev database."
|
|
)
|
|
|
|
options = parser.parse_args()
|
|
sys.exit(main(options))
|