mirror of
https://github.com/zulip/zulip.git
synced 2025-11-04 22:13:26 +00:00
Compare commits
17 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
a625ca49ec | ||
|
|
96bd1c38dc | ||
|
|
9748780192 | ||
|
|
bc3f096918 | ||
|
|
af4aac6836 | ||
|
|
e5f7000a23 | ||
|
|
00bf7b25b5 | ||
|
|
2c6bfe136a | ||
|
|
db51a1c547 | ||
|
|
3f76745235 | ||
|
|
b59b5cac35 | ||
|
|
5dd330e769 | ||
|
|
140e598a89 | ||
|
|
8159c03205 | ||
|
|
0d12dfd06f | ||
|
|
6888826d5b | ||
|
|
f0add4638c |
83
README.md
83
README.md
@@ -1,3 +1,49 @@
|
|||||||
|
Zulip
|
||||||
|
=====
|
||||||
|
|
||||||
|
Zulip is a powerful, open source group chat application. Written in
|
||||||
|
Python and using the Django framework, Zulip supports both private
|
||||||
|
messaging and group chats via conversation streams.
|
||||||
|
|
||||||
|
Zulip also supports fast search, drag-and-drop file uploads, image
|
||||||
|
previews, group private messages, audible notifications,
|
||||||
|
missed-message emails, desktop apps, and much more.
|
||||||
|
|
||||||
|
Further information on the Zulip project and its features can be found
|
||||||
|
at https://www.zulip.org
|
||||||
|
|
||||||
|
Contributing to Zulip
|
||||||
|
=====================
|
||||||
|
|
||||||
|
Zulip welcomes all forms of contributions!
|
||||||
|
|
||||||
|
Before a pull request can be merged, you need to to sign the [Dropbox
|
||||||
|
Contributor License Agreement](https://opensource.dropbox.com/cla/).
|
||||||
|
|
||||||
|
Please run the tests (tools/test-all) before submitting your pull
|
||||||
|
request and read our [commit message style
|
||||||
|
guidelines](http://zulip.readthedocs.org/en/latest/code-style.html#commit-messages).
|
||||||
|
|
||||||
|
Zulip has a growing collection of developer documentation including
|
||||||
|
detailed documentation on coding style available on [Read The
|
||||||
|
Docs](https://zulip.readthedocs.org/).
|
||||||
|
|
||||||
|
Zulip also has a [development discussion mailing list](https://groups.google.com/forum/#!forum/zulip-devel)
|
||||||
|
|
||||||
|
Feel free to send any questions or suggestions of areas where you'd
|
||||||
|
love to see more documentation to the list!
|
||||||
|
|
||||||
|
We recommend sending proposals for large features or refactorings to
|
||||||
|
the zulip-devel list for discussion and advice before getting too deep
|
||||||
|
into implementation.
|
||||||
|
|
||||||
|
Please report any security issues you discover to support@zulip.com.
|
||||||
|
|
||||||
|
Running Zulip in production
|
||||||
|
===========================
|
||||||
|
|
||||||
|
This is documented in https://zulip.org/server.html and README.prod.md.
|
||||||
|
|
||||||
Installing the Zulip Development environment
|
Installing the Zulip Development environment
|
||||||
============================================
|
============================================
|
||||||
|
|
||||||
@@ -10,7 +56,7 @@ This is the recommended approach, and is tested on OS X 10.10 as well as Ubuntu
|
|||||||
<http://download.virtualbox.org/virtualbox/4.3.30/VirtualBox-4.3.30-101610-OSX.dmg>
|
<http://download.virtualbox.org/virtualbox/4.3.30/VirtualBox-4.3.30-101610-OSX.dmg>
|
||||||
and install it.
|
and install it.
|
||||||
* If your host is Ubuntu 14.04:
|
* If your host is Ubuntu 14.04:
|
||||||
sudo apt-get install vagrant lxc lxc-templates cgroup-lite redir && vagrant plugin install vagrant-lxc
|
`sudo apt-get install vagrant lxc lxc-templates cgroup-lite redir && vagrant plugin install vagrant-lxc`
|
||||||
|
|
||||||
Once that's done, simply change to your zulip directory and run
|
Once that's done, simply change to your zulip directory and run
|
||||||
`vagrant up` in your terminal. That will install the development
|
`vagrant up` in your terminal. That will install the development
|
||||||
@@ -115,8 +161,8 @@ individual tests, e.g.:
|
|||||||
./tools/test-js-with-casper 10-navigation.js
|
./tools/test-js-with-casper 10-navigation.js
|
||||||
```
|
```
|
||||||
|
|
||||||
Possible issues
|
Possible testing issues
|
||||||
===============
|
=======================
|
||||||
|
|
||||||
The Casper tests are flaky on the Virtualbox environment (probably due
|
The Casper tests are flaky on the Virtualbox environment (probably due
|
||||||
to some performance-sensitive races). Until this issue is debugged,
|
to some performance-sensitive races). Until this issue is debugged,
|
||||||
@@ -132,37 +178,6 @@ When running the test suite, if you get an error like this:
|
|||||||
|
|
||||||
… then you need to install tsearch-extras, described above. Afterwards, re-run the `init*-db` and the `do-destroy-rebuild*-database` scripts.
|
… then you need to install tsearch-extras, described above. Afterwards, re-run the `init*-db` and the `do-destroy-rebuild*-database` scripts.
|
||||||
|
|
||||||
Contributing to Zulip
|
|
||||||
=====================
|
|
||||||
|
|
||||||
Zulip welcomes all forms of contributions!
|
|
||||||
|
|
||||||
Before a pull request can be merged, you need to to sign the [Dropbox
|
|
||||||
Contributor License Agreement](https://opensource.dropbox.com/cla/).
|
|
||||||
|
|
||||||
Please run the tests (tools/test-all) before submitting your pull
|
|
||||||
request.
|
|
||||||
|
|
||||||
Zulip has a growing collection of developer documentation including
|
|
||||||
detailed documentation on coding style available on [Read The
|
|
||||||
Docs](https://zulip.readthedocs.org/).
|
|
||||||
|
|
||||||
Zulip also has a [development discussion mailing list](https://groups.google.com/forum/#!forum/zulip-devel)
|
|
||||||
|
|
||||||
Feel free to send any questions or suggestions of areas where you'd
|
|
||||||
love to see more documentation to the list!
|
|
||||||
|
|
||||||
We recommend sending proposals for large features or refactorings to
|
|
||||||
the zulip-devel list for discussion and advice before getting too deep
|
|
||||||
into implementation.
|
|
||||||
|
|
||||||
Please report any security issues you discover to support@zulip.com.
|
|
||||||
|
|
||||||
Running Zulip in production
|
|
||||||
===========================
|
|
||||||
|
|
||||||
This is documented in https://zulip.org/server.html and README.prod.md.
|
|
||||||
|
|
||||||
License
|
License
|
||||||
=======
|
=======
|
||||||
|
|
||||||
|
|||||||
@@ -11,5 +11,5 @@ ZULIP_DIR=/home/zulip/deployments/current
|
|||||||
STATE_DIR=/var/lib/nagios_state
|
STATE_DIR=/var/lib/nagios_state
|
||||||
STATE_FILE=$STATE_DIR/check-rabbitmq-consumers-$queue
|
STATE_FILE=$STATE_DIR/check-rabbitmq-consumers-$queue
|
||||||
|
|
||||||
$ZULIP_DIR/bots/check-rabbitmq-consumers --queue=$queue &> ${STATE_FILE}-tmp;
|
"$ZULIP_DIR/bots/check-rabbitmq-consumers" "--queue=$queue" &> "${STATE_FILE}-tmp";
|
||||||
mv ${STATE_FILE}-tmp $STATE_FILE
|
mv "${STATE_FILE}-tmp" "$STATE_FILE"
|
||||||
|
|||||||
@@ -54,6 +54,6 @@ while backoff.keep_going():
|
|||||||
print ""
|
print ""
|
||||||
print ""
|
print ""
|
||||||
print "ERROR: The Jabber mirroring bot is unable to continue mirroring Jabber."
|
print "ERROR: The Jabber mirroring bot is unable to continue mirroring Jabber."
|
||||||
print "Please contact support@zulip.com if you need assistence."
|
print "Please contact support@zulip.com if you need assistance."
|
||||||
print ""
|
print ""
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ server {
|
|||||||
listen 443;
|
listen 443;
|
||||||
|
|
||||||
ssl on;
|
ssl on;
|
||||||
ssl_certificate /etc/ssl/certs/zulip-combined-chain.crt;
|
ssl_certificate /etc/ssl/certs/zulip.combined-chain.crt;
|
||||||
ssl_certificate_key /etc/ssl/private/zulip.key;
|
ssl_certificate_key /etc/ssl/private/zulip.key;
|
||||||
|
|
||||||
location /user_uploads {
|
location /user_uploads {
|
||||||
|
|||||||
@@ -1,22 +0,0 @@
|
|||||||
[main]
|
|
||||||
server = puppet.zulip.com
|
|
||||||
environment = production
|
|
||||||
confdir = /etc/puppet
|
|
||||||
logdir=/var/log/puppet
|
|
||||||
vardir=/var/lib/puppet
|
|
||||||
ssldir=/var/lib/puppet/ssl
|
|
||||||
rundir=/var/run/puppet
|
|
||||||
factpath=$vardir/lib/facter
|
|
||||||
templatedir=$confdir/templates
|
|
||||||
prerun_command=/etc/puppet/etckeeper-commit-pre
|
|
||||||
postrun_command=/etc/puppet/etckeeper-commit-post
|
|
||||||
modulepath = /root/zulip/puppet:/etc/puppet/modules:/usr/share/puppet/modules
|
|
||||||
|
|
||||||
[master]
|
|
||||||
environment = production
|
|
||||||
manifest = $confdir/environments/$environment/manifests/site.pp
|
|
||||||
modulepath = $confdir/environments/$environment/modules
|
|
||||||
[agent]
|
|
||||||
report = true
|
|
||||||
show_diff = true
|
|
||||||
environment = production
|
|
||||||
@@ -1,6 +1,6 @@
|
|||||||
# Redis configuration file example
|
# Redis configuration file example
|
||||||
|
|
||||||
# Note on units: when memory size is needed, it is possible to specifiy
|
# Note on units: when memory size is needed, it is possible to specify
|
||||||
# it in the usual form of 1k 5GB 4M and so forth:
|
# it in the usual form of 1k 5GB 4M and so forth:
|
||||||
#
|
#
|
||||||
# 1k => 1000 bytes
|
# 1k => 1000 bytes
|
||||||
@@ -12,6 +12,26 @@
|
|||||||
#
|
#
|
||||||
# units are case insensitive so 1GB 1Gb 1gB are all the same.
|
# units are case insensitive so 1GB 1Gb 1gB are all the same.
|
||||||
|
|
||||||
|
################################## INCLUDES ###################################
|
||||||
|
|
||||||
|
# Include one or more other config files here. This is useful if you
|
||||||
|
# have a standard template that goes to all Redis server but also need
|
||||||
|
# to customize a few per-server settings. Include files can include
|
||||||
|
# other files, so use this wisely.
|
||||||
|
#
|
||||||
|
# Notice option "include" won't be rewritten by command "CONFIG REWRITE"
|
||||||
|
# from admin or Redis Sentinel. Since Redis always uses the last processed
|
||||||
|
# line as value of a configuration directive, you'd better put includes
|
||||||
|
# at the beginning of this file to avoid overwriting config change at runtime.
|
||||||
|
#
|
||||||
|
# If instead you are interested in using includes to override configuration
|
||||||
|
# options, it is better to use include as the last line.
|
||||||
|
#
|
||||||
|
# include /path/to/local.conf
|
||||||
|
# include /path/to/other.conf
|
||||||
|
|
||||||
|
################################ GENERAL #####################################
|
||||||
|
|
||||||
# By default Redis does not run as a daemon. Use 'yes' if you need it.
|
# By default Redis does not run as a daemon. Use 'yes' if you need it.
|
||||||
# Note that Redis will write a pid file in /var/run/redis.pid when daemonized.
|
# Note that Redis will write a pid file in /var/run/redis.pid when daemonized.
|
||||||
daemonize yes
|
daemonize yes
|
||||||
@@ -24,9 +44,14 @@ pidfile /var/run/redis/redis-server.pid
|
|||||||
# If port 0 is specified Redis will not listen on a TCP socket.
|
# If port 0 is specified Redis will not listen on a TCP socket.
|
||||||
port 6379
|
port 6379
|
||||||
|
|
||||||
# If you want you can bind a single interface, if the bind option is not
|
# By default Redis listens for connections from all the network interfaces
|
||||||
# specified all the interfaces will listen for incoming connections.
|
# available on the server. It is possible to listen to just one or multiple
|
||||||
|
# interfaces using the "bind" configuration directive, followed by one or
|
||||||
|
# more IP addresses.
|
||||||
#
|
#
|
||||||
|
# Examples:
|
||||||
|
#
|
||||||
|
# bind 192.168.1.100 10.0.0.1
|
||||||
bind 127.0.0.1
|
bind 127.0.0.1
|
||||||
|
|
||||||
# Specify the path for the unix socket that will be used to listen for
|
# Specify the path for the unix socket that will be used to listen for
|
||||||
@@ -39,15 +64,31 @@ bind 127.0.0.1
|
|||||||
# Close the connection after a client is idle for N seconds (0 to disable)
|
# Close the connection after a client is idle for N seconds (0 to disable)
|
||||||
timeout 0
|
timeout 0
|
||||||
|
|
||||||
# Set server verbosity to 'debug'
|
# TCP keepalive.
|
||||||
# it can be one of:
|
#
|
||||||
|
# If non-zero, use SO_KEEPALIVE to send TCP ACKs to clients in absence
|
||||||
|
# of communication. This is useful for two reasons:
|
||||||
|
#
|
||||||
|
# 1) Detect dead peers.
|
||||||
|
# 2) Take the connection alive from the point of view of network
|
||||||
|
# equipment in the middle.
|
||||||
|
#
|
||||||
|
# On Linux, the specified value (in seconds) is the period used to send ACKs.
|
||||||
|
# Note that to close the connection the double of the time is needed.
|
||||||
|
# On other kernels the period depends on the kernel configuration.
|
||||||
|
#
|
||||||
|
# A reasonable value for this option is 60 seconds.
|
||||||
|
tcp-keepalive 0
|
||||||
|
|
||||||
|
# Specify the server verbosity level.
|
||||||
|
# This can be one of:
|
||||||
# debug (a lot of information, useful for development/testing)
|
# debug (a lot of information, useful for development/testing)
|
||||||
# verbose (many rarely useful info, but not a mess like the debug level)
|
# verbose (many rarely useful info, but not a mess like the debug level)
|
||||||
# notice (moderately verbose, what you want in production probably)
|
# notice (moderately verbose, what you want in production probably)
|
||||||
# warning (only very important / critical messages are logged)
|
# warning (only very important / critical messages are logged)
|
||||||
loglevel notice
|
loglevel notice
|
||||||
|
|
||||||
# Specify the log file name. Also 'stdout' can be used to force
|
# Specify the log file name. Also the empty string can be used to force
|
||||||
# Redis to log on the standard output. Note that if you use standard
|
# Redis to log on the standard output. Note that if you use standard
|
||||||
# output for logging but daemonize, logs will be sent to /dev/null
|
# output for logging but daemonize, logs will be sent to /dev/null
|
||||||
logfile /var/log/redis/redis-server.log
|
logfile /var/log/redis/redis-server.log
|
||||||
@@ -59,7 +100,7 @@ logfile /var/log/redis/redis-server.log
|
|||||||
# Specify the syslog identity.
|
# Specify the syslog identity.
|
||||||
# syslog-ident redis
|
# syslog-ident redis
|
||||||
|
|
||||||
# Specify the syslog facility. Must be USER or between LOCAL0-LOCAL7.
|
# Specify the syslog facility. Must be USER or between LOCAL0-LOCAL7.
|
||||||
# syslog-facility local0
|
# syslog-facility local0
|
||||||
|
|
||||||
# Set the number of databases. The default database is DB 0, you can select
|
# Set the number of databases. The default database is DB 0, you can select
|
||||||
@@ -67,7 +108,7 @@ logfile /var/log/redis/redis-server.log
|
|||||||
# dbid is a number between 0 and 'databases'-1
|
# dbid is a number between 0 and 'databases'-1
|
||||||
databases 16
|
databases 16
|
||||||
|
|
||||||
################################ SNAPSHOTTING #################################
|
################################ SNAPSHOTTING ################################
|
||||||
#
|
#
|
||||||
# Save the DB on disk:
|
# Save the DB on disk:
|
||||||
#
|
#
|
||||||
@@ -82,10 +123,31 @@ databases 16
|
|||||||
# after 60 sec if at least 10000 keys changed
|
# after 60 sec if at least 10000 keys changed
|
||||||
#
|
#
|
||||||
# Note: you can disable saving at all commenting all the "save" lines.
|
# Note: you can disable saving at all commenting all the "save" lines.
|
||||||
|
#
|
||||||
|
# It is also possible to remove all the previously configured save
|
||||||
|
# points by adding a save directive with a single empty string argument
|
||||||
|
# like in the following example:
|
||||||
|
#
|
||||||
|
# save ""
|
||||||
|
|
||||||
# save 900 1
|
save 900 1
|
||||||
# save 300 10
|
save 300 10
|
||||||
# save 60 10000
|
save 60 10000
|
||||||
|
|
||||||
|
# By default Redis will stop accepting writes if RDB snapshots are enabled
|
||||||
|
# (at least one save point) and the latest background save failed.
|
||||||
|
# This will make the user aware (in a hard way) that data is not persisting
|
||||||
|
# on disk properly, otherwise chances are that no one will notice and some
|
||||||
|
# disaster will happen.
|
||||||
|
#
|
||||||
|
# If the background saving process will start working again Redis will
|
||||||
|
# automatically allow writes again.
|
||||||
|
#
|
||||||
|
# However if you have setup your proper monitoring of the Redis server
|
||||||
|
# and persistence, you may want to disable this feature so that Redis will
|
||||||
|
# continue to work as usual even if there are problems with disk,
|
||||||
|
# permissions, and so forth.
|
||||||
|
stop-writes-on-bgsave-error yes
|
||||||
|
|
||||||
# Compress string objects using LZF when dump .rdb databases?
|
# Compress string objects using LZF when dump .rdb databases?
|
||||||
# For default that's set to 'yes' as it's almost always a win.
|
# For default that's set to 'yes' as it's almost always a win.
|
||||||
@@ -93,6 +155,15 @@ databases 16
|
|||||||
# the dataset will likely be bigger if you have compressible values or keys.
|
# the dataset will likely be bigger if you have compressible values or keys.
|
||||||
rdbcompression yes
|
rdbcompression yes
|
||||||
|
|
||||||
|
# Since version 5 of RDB a CRC64 checksum is placed at the end of the file.
|
||||||
|
# This makes the format more resistant to corruption but there is a performance
|
||||||
|
# hit to pay (around 10%) when saving and loading RDB files, so you can disable it
|
||||||
|
# for maximum performances.
|
||||||
|
#
|
||||||
|
# RDB files created with checksum disabled have a checksum of zero that will
|
||||||
|
# tell the loading code to skip the check.
|
||||||
|
rdbchecksum yes
|
||||||
|
|
||||||
# The filename where to dump the DB
|
# The filename where to dump the DB
|
||||||
dbfilename dump.rdb
|
dbfilename dump.rdb
|
||||||
|
|
||||||
@@ -101,7 +172,7 @@ dbfilename dump.rdb
|
|||||||
# The DB will be written inside this directory, with the filename specified
|
# The DB will be written inside this directory, with the filename specified
|
||||||
# above using the 'dbfilename' configuration directive.
|
# above using the 'dbfilename' configuration directive.
|
||||||
#
|
#
|
||||||
# Also the Append Only File will be created inside this directory.
|
# The Append Only File will also be created inside this directory.
|
||||||
#
|
#
|
||||||
# Note that you must specify a directory here, not a file name.
|
# Note that you must specify a directory here, not a file name.
|
||||||
dir /var/lib/redis
|
dir /var/lib/redis
|
||||||
@@ -122,27 +193,46 @@ dir /var/lib/redis
|
|||||||
#
|
#
|
||||||
# masterauth <master-password>
|
# masterauth <master-password>
|
||||||
|
|
||||||
# When a slave lost the connection with the master, or when the replication
|
# When a slave loses its connection with the master, or when the replication
|
||||||
# is still in progress, the slave can act in two different ways:
|
# is still in progress, the slave can act in two different ways:
|
||||||
#
|
#
|
||||||
# 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will
|
# 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will
|
||||||
# still reply to client requests, possibly with out of data data, or the
|
# still reply to client requests, possibly with out of date data, or the
|
||||||
# data set may just be empty if this is the first synchronization.
|
# data set may just be empty if this is the first synchronization.
|
||||||
#
|
#
|
||||||
# 2) if slave-serve-stale data is set to 'no' the slave will reply with
|
# 2) if slave-serve-stale-data is set to 'no' the slave will reply with
|
||||||
# an error "SYNC with master in progress" to all the kind of commands
|
# an error "SYNC with master in progress" to all the kind of commands
|
||||||
# but to INFO and SLAVEOF.
|
# but to INFO and SLAVEOF.
|
||||||
#
|
#
|
||||||
slave-serve-stale-data yes
|
slave-serve-stale-data yes
|
||||||
|
|
||||||
|
# You can configure a slave instance to accept writes or not. Writing against
|
||||||
|
# a slave instance may be useful to store some ephemeral data (because data
|
||||||
|
# written on a slave will be easily deleted after resync with the master) but
|
||||||
|
# may also cause problems if clients are writing to it because of a
|
||||||
|
# misconfiguration.
|
||||||
|
#
|
||||||
|
# Since Redis 2.6 by default slaves are read-only.
|
||||||
|
#
|
||||||
|
# Note: read only slaves are not designed to be exposed to untrusted clients
|
||||||
|
# on the internet. It's just a protection layer against misuse of the instance.
|
||||||
|
# Still a read only slave exports by default all the administrative commands
|
||||||
|
# such as CONFIG, DEBUG, and so forth. To a limited extent you can improve
|
||||||
|
# security of read only slaves using 'rename-command' to shadow all the
|
||||||
|
# administrative / dangerous commands.
|
||||||
|
slave-read-only yes
|
||||||
|
|
||||||
# Slaves send PINGs to server in a predefined interval. It's possible to change
|
# Slaves send PINGs to server in a predefined interval. It's possible to change
|
||||||
# this interval with the repl_ping_slave_period option. The default value is 10
|
# this interval with the repl_ping_slave_period option. The default value is 10
|
||||||
# seconds.
|
# seconds.
|
||||||
#
|
#
|
||||||
# repl-ping-slave-period 10
|
# repl-ping-slave-period 10
|
||||||
|
|
||||||
# The following option sets a timeout for both Bulk transfer I/O timeout and
|
# The following option sets the replication timeout for:
|
||||||
# master data or ping response timeout. The default value is 60 seconds.
|
#
|
||||||
|
# 1) Bulk transfer I/O during SYNC, from the point of view of slave.
|
||||||
|
# 2) Master timeout from the point of view of slaves (data, pings).
|
||||||
|
# 3) Slave timeout from the point of view of masters (REPLCONF ACK pings).
|
||||||
#
|
#
|
||||||
# It is important to make sure that this value is greater than the value
|
# It is important to make sure that this value is greater than the value
|
||||||
# specified for repl-ping-slave-period otherwise a timeout will be detected
|
# specified for repl-ping-slave-period otherwise a timeout will be detected
|
||||||
@@ -150,6 +240,80 @@ slave-serve-stale-data yes
|
|||||||
#
|
#
|
||||||
# repl-timeout 60
|
# repl-timeout 60
|
||||||
|
|
||||||
|
# Disable TCP_NODELAY on the slave socket after SYNC?
|
||||||
|
#
|
||||||
|
# If you select "yes" Redis will use a smaller number of TCP packets and
|
||||||
|
# less bandwidth to send data to slaves. But this can add a delay for
|
||||||
|
# the data to appear on the slave side, up to 40 milliseconds with
|
||||||
|
# Linux kernels using a default configuration.
|
||||||
|
#
|
||||||
|
# If you select "no" the delay for data to appear on the slave side will
|
||||||
|
# be reduced but more bandwidth will be used for replication.
|
||||||
|
#
|
||||||
|
# By default we optimize for low latency, but in very high traffic conditions
|
||||||
|
# or when the master and slaves are many hops away, turning this to "yes" may
|
||||||
|
# be a good idea.
|
||||||
|
repl-disable-tcp-nodelay no
|
||||||
|
|
||||||
|
# Set the replication backlog size. The backlog is a buffer that accumulates
|
||||||
|
# slave data when slaves are disconnected for some time, so that when a slave
|
||||||
|
# wants to reconnect again, often a full resync is not needed, but a partial
|
||||||
|
# resync is enough, just passing the portion of data the slave missed while
|
||||||
|
# disconnected.
|
||||||
|
#
|
||||||
|
# The biggest the replication backlog, the longer the time the slave can be
|
||||||
|
# disconnected and later be able to perform a partial resynchronization.
|
||||||
|
#
|
||||||
|
# The backlog is only allocated once there is at least a slave connected.
|
||||||
|
#
|
||||||
|
# repl-backlog-size 1mb
|
||||||
|
|
||||||
|
# After a master has no longer connected slaves for some time, the backlog
|
||||||
|
# will be freed. The following option configures the amount of seconds that
|
||||||
|
# need to elapse, starting from the time the last slave disconnected, for
|
||||||
|
# the backlog buffer to be freed.
|
||||||
|
#
|
||||||
|
# A value of 0 means to never release the backlog.
|
||||||
|
#
|
||||||
|
# repl-backlog-ttl 3600
|
||||||
|
|
||||||
|
# The slave priority is an integer number published by Redis in the INFO output.
|
||||||
|
# It is used by Redis Sentinel in order to select a slave to promote into a
|
||||||
|
# master if the master is no longer working correctly.
|
||||||
|
#
|
||||||
|
# A slave with a low priority number is considered better for promotion, so
|
||||||
|
# for instance if there are three slaves with priority 10, 100, 25 Sentinel will
|
||||||
|
# pick the one with priority 10, that is the lowest.
|
||||||
|
#
|
||||||
|
# However a special priority of 0 marks the slave as not able to perform the
|
||||||
|
# role of master, so a slave with priority of 0 will never be selected by
|
||||||
|
# Redis Sentinel for promotion.
|
||||||
|
#
|
||||||
|
# By default the priority is 100.
|
||||||
|
slave-priority 100
|
||||||
|
|
||||||
|
# It is possible for a master to stop accepting writes if there are less than
|
||||||
|
# N slaves connected, having a lag less or equal than M seconds.
|
||||||
|
#
|
||||||
|
# The N slaves need to be in "online" state.
|
||||||
|
#
|
||||||
|
# The lag in seconds, that must be <= the specified value, is calculated from
|
||||||
|
# the last ping received from the slave, that is usually sent every second.
|
||||||
|
#
|
||||||
|
# This option does not GUARANTEES that N replicas will accept the write, but
|
||||||
|
# will limit the window of exposure for lost writes in case not enough slaves
|
||||||
|
# are available, to the specified number of seconds.
|
||||||
|
#
|
||||||
|
# For example to require at least 3 slaves with a lag <= 10 seconds use:
|
||||||
|
#
|
||||||
|
# min-slaves-to-write 3
|
||||||
|
# min-slaves-max-lag 10
|
||||||
|
#
|
||||||
|
# Setting one or the other to 0 disables the feature.
|
||||||
|
#
|
||||||
|
# By default min-slaves-to-write is set to 0 (feature disabled) and
|
||||||
|
# min-slaves-max-lag is set to 10.
|
||||||
|
|
||||||
################################## SECURITY ###################################
|
################################## SECURITY ###################################
|
||||||
|
|
||||||
# Require clients to issue AUTH <PASSWORD> before processing any other
|
# Require clients to issue AUTH <PASSWORD> before processing any other
|
||||||
@@ -167,33 +331,39 @@ slave-serve-stale-data yes
|
|||||||
|
|
||||||
# Command renaming.
|
# Command renaming.
|
||||||
#
|
#
|
||||||
# It is possilbe to change the name of dangerous commands in a shared
|
# It is possible to change the name of dangerous commands in a shared
|
||||||
# environment. For instance the CONFIG command may be renamed into something
|
# environment. For instance the CONFIG command may be renamed into something
|
||||||
# of hard to guess so that it will be still available for internal-use
|
# hard to guess so that it will still be available for internal-use tools
|
||||||
# tools but not available for general clients.
|
# but not available for general clients.
|
||||||
#
|
#
|
||||||
# Example:
|
# Example:
|
||||||
#
|
#
|
||||||
# rename-command CONFIG b840fc02d524045429941cc15f59e41cb7be6c52
|
# rename-command CONFIG b840fc02d524045429941cc15f59e41cb7be6c52
|
||||||
#
|
#
|
||||||
# It is also possilbe to completely kill a command renaming it into
|
# It is also possible to completely kill a command by renaming it into
|
||||||
# an empty string:
|
# an empty string:
|
||||||
#
|
#
|
||||||
# rename-command CONFIG ""
|
# rename-command CONFIG ""
|
||||||
|
#
|
||||||
|
# Please note that changing the name of commands that are logged into the
|
||||||
|
# AOF file or transmitted to slaves may cause problems.
|
||||||
|
|
||||||
################################### LIMITS ####################################
|
################################### LIMITS ####################################
|
||||||
|
|
||||||
# Set the max number of connected clients at the same time. By default there
|
# Set the max number of connected clients at the same time. By default
|
||||||
# is no limit, and it's up to the number of file descriptors the Redis process
|
# this limit is set to 10000 clients, however if the Redis server is not
|
||||||
# is able to open. The special value '0' means no limits.
|
# able to configure the process file limit to allow for the specified limit
|
||||||
|
# the max number of allowed clients is set to the current file limit
|
||||||
|
# minus 32 (as Redis reserves a few file descriptors for internal uses).
|
||||||
|
#
|
||||||
# Once the limit is reached Redis will close all the new connections sending
|
# Once the limit is reached Redis will close all the new connections sending
|
||||||
# an error 'max number of clients reached'.
|
# an error 'max number of clients reached'.
|
||||||
#
|
#
|
||||||
# maxclients 128
|
# maxclients 10000
|
||||||
|
|
||||||
# Don't use more memory than the specified amount of bytes.
|
# Don't use more memory than the specified amount of bytes.
|
||||||
# When the memory limit is reached Redis will try to remove keys
|
# When the memory limit is reached Redis will try to remove keys
|
||||||
# accordingly to the eviction policy selected (see maxmemmory-policy).
|
# according to the eviction policy selected (see maxmemory-policy).
|
||||||
#
|
#
|
||||||
# If Redis can't remove keys according to the policy, or if the policy is
|
# If Redis can't remove keys according to the policy, or if the policy is
|
||||||
# set to 'noeviction', Redis will start to reply with errors to commands
|
# set to 'noeviction', Redis will start to reply with errors to commands
|
||||||
@@ -201,7 +371,7 @@ slave-serve-stale-data yes
|
|||||||
# to reply to read-only commands like GET.
|
# to reply to read-only commands like GET.
|
||||||
#
|
#
|
||||||
# This option is usually useful when using Redis as an LRU cache, or to set
|
# This option is usually useful when using Redis as an LRU cache, or to set
|
||||||
# an hard memory limit for an instance (using the 'noeviction' policy).
|
# a hard memory limit for an instance (using the 'noeviction' policy).
|
||||||
#
|
#
|
||||||
# WARNING: If you have slaves attached to an instance with maxmemory on,
|
# WARNING: If you have slaves attached to an instance with maxmemory on,
|
||||||
# the size of the output buffers needed to feed the slaves are subtracted
|
# the size of the output buffers needed to feed the slaves are subtracted
|
||||||
@@ -217,16 +387,16 @@ slave-serve-stale-data yes
|
|||||||
# maxmemory <bytes>
|
# maxmemory <bytes>
|
||||||
|
|
||||||
# MAXMEMORY POLICY: how Redis will select what to remove when maxmemory
|
# MAXMEMORY POLICY: how Redis will select what to remove when maxmemory
|
||||||
# is reached? You can select among five behavior:
|
# is reached. You can select among five behaviors:
|
||||||
#
|
#
|
||||||
# volatile-lru -> remove the key with an expire set using an LRU algorithm
|
# volatile-lru -> remove the key with an expire set using an LRU algorithm
|
||||||
# allkeys-lru -> remove any key accordingly to the LRU algorithm
|
# allkeys-lru -> remove any key accordingly to the LRU algorithm
|
||||||
# volatile-random -> remove a random key with an expire set
|
# volatile-random -> remove a random key with an expire set
|
||||||
# allkeys->random -> remove a random key, any key
|
# allkeys-random -> remove a random key, any key
|
||||||
# volatile-ttl -> remove the key with the nearest expire time (minor TTL)
|
# volatile-ttl -> remove the key with the nearest expire time (minor TTL)
|
||||||
# noeviction -> don't expire at all, just return an error on write operations
|
# noeviction -> don't expire at all, just return an error on write operations
|
||||||
#
|
#
|
||||||
# Note: with all the kind of policies, Redis will return an error on write
|
# Note: with any of the above policies, Redis will return an error on write
|
||||||
# operations, when there are not suitable keys for eviction.
|
# operations, when there are not suitable keys for eviction.
|
||||||
#
|
#
|
||||||
# At the date of writing this commands are: set setnx setex append
|
# At the date of writing this commands are: set setnx setex append
|
||||||
@@ -249,26 +419,29 @@ slave-serve-stale-data yes
|
|||||||
|
|
||||||
############################## APPEND ONLY MODE ###############################
|
############################## APPEND ONLY MODE ###############################
|
||||||
|
|
||||||
# By default Redis asynchronously dumps the dataset on disk. If you can live
|
# By default Redis asynchronously dumps the dataset on disk. This mode is
|
||||||
# with the idea that the latest records will be lost if something like a crash
|
# good enough in many applications, but an issue with the Redis process or
|
||||||
# happens this is the preferred way to run Redis. If instead you care a lot
|
# a power outage may result into a few minutes of writes lost (depending on
|
||||||
# about your data and don't want to that a single record can get lost you should
|
# the configured save points).
|
||||||
# enable the append only mode: when this mode is enabled Redis will append
|
|
||||||
# every write operation received in the file appendonly.aof. This file will
|
|
||||||
# be read on startup in order to rebuild the full dataset in memory.
|
|
||||||
#
|
#
|
||||||
# Note that you can have both the async dumps and the append only file if you
|
# The Append Only File is an alternative persistence mode that provides
|
||||||
# like (you have to comment the "save" statements above to disable the dumps).
|
# much better durability. For instance using the default data fsync policy
|
||||||
# Still if append only mode is enabled Redis will load the data from the
|
# (see later in the config file) Redis can lose just one second of writes in a
|
||||||
# log file at startup ignoring the dump.rdb file.
|
# dramatic event like a server power outage, or a single write if something
|
||||||
|
# wrong with the Redis process itself happens, but the operating system is
|
||||||
|
# still running correctly.
|
||||||
#
|
#
|
||||||
# IMPORTANT: Check the BGREWRITEAOF to check how to rewrite the append
|
# AOF and RDB persistence can be enabled at the same time without problems.
|
||||||
# log file in background when it gets too big.
|
# If the AOF is enabled on startup Redis will load the AOF, that is the file
|
||||||
|
# with the better durability guarantees.
|
||||||
|
#
|
||||||
|
# Please check http://redis.io/topics/persistence for more information.
|
||||||
|
|
||||||
appendonly no
|
appendonly no
|
||||||
|
|
||||||
# The name of the append only file (default: "appendonly.aof")
|
# The name of the append only file (default: "appendonly.aof")
|
||||||
# appendfilename appendonly.aof
|
|
||||||
|
appendfilename "appendonly.aof"
|
||||||
|
|
||||||
# The fsync() call tells the Operating System to actually write data on disk
|
# The fsync() call tells the Operating System to actually write data on disk
|
||||||
# instead to wait for more data in the output buffer. Some OS will really flush
|
# instead to wait for more data in the output buffer. Some OS will really flush
|
||||||
@@ -278,16 +451,19 @@ appendonly no
|
|||||||
#
|
#
|
||||||
# no: don't fsync, just let the OS flush the data when it wants. Faster.
|
# no: don't fsync, just let the OS flush the data when it wants. Faster.
|
||||||
# always: fsync after every write to the append only log . Slow, Safest.
|
# always: fsync after every write to the append only log . Slow, Safest.
|
||||||
# everysec: fsync only if one second passed since the last fsync. Compromise.
|
# everysec: fsync only one time every second. Compromise.
|
||||||
#
|
#
|
||||||
# The default is "everysec" that's usually the right compromise between
|
# The default is "everysec", as that's usually the right compromise between
|
||||||
# speed and data safety. It's up to you to understand if you can relax this to
|
# speed and data safety. It's up to you to understand if you can relax this to
|
||||||
# "no" that will will let the operating system flush the output buffer when
|
# "no" that will let the operating system flush the output buffer when
|
||||||
# it wants, for better performances (but if you can live with the idea of
|
# it wants, for better performances (but if you can live with the idea of
|
||||||
# some data loss consider the default persistence mode that's snapshotting),
|
# some data loss consider the default persistence mode that's snapshotting),
|
||||||
# or on the contrary, use "always" that's very slow but a bit safer than
|
# or on the contrary, use "always" that's very slow but a bit safer than
|
||||||
# everysec.
|
# everysec.
|
||||||
#
|
#
|
||||||
|
# More details please check the following article:
|
||||||
|
# http://antirez.com/post/redis-persistence-demystified.html
|
||||||
|
#
|
||||||
# If unsure, use "everysec".
|
# If unsure, use "everysec".
|
||||||
|
|
||||||
# appendfsync always
|
# appendfsync always
|
||||||
@@ -305,21 +481,22 @@ appendfsync everysec
|
|||||||
# that will prevent fsync() from being called in the main process while a
|
# that will prevent fsync() from being called in the main process while a
|
||||||
# BGSAVE or BGREWRITEAOF is in progress.
|
# BGSAVE or BGREWRITEAOF is in progress.
|
||||||
#
|
#
|
||||||
# This means that while another child is saving the durability of Redis is
|
# This means that while another child is saving, the durability of Redis is
|
||||||
# the same as "appendfsync none", that in pratical terms means that it is
|
# the same as "appendfsync none". In practical terms, this means that it is
|
||||||
# possible to lost up to 30 seconds of log in the worst scenario (with the
|
# possible to lose up to 30 seconds of log in the worst scenario (with the
|
||||||
# default Linux settings).
|
# default Linux settings).
|
||||||
#
|
#
|
||||||
# If you have latency problems turn this to "yes". Otherwise leave it as
|
# If you have latency problems turn this to "yes". Otherwise leave it as
|
||||||
# "no" that is the safest pick from the point of view of durability.
|
# "no" that is the safest pick from the point of view of durability.
|
||||||
|
|
||||||
no-appendfsync-on-rewrite no
|
no-appendfsync-on-rewrite no
|
||||||
|
|
||||||
# Automatic rewrite of the append only file.
|
# Automatic rewrite of the append only file.
|
||||||
# Redis is able to automatically rewrite the log file implicitly calling
|
# Redis is able to automatically rewrite the log file implicitly calling
|
||||||
# BGREWRITEAOF when the AOF log size will growth by the specified percentage.
|
# BGREWRITEAOF when the AOF log size grows by the specified percentage.
|
||||||
#
|
#
|
||||||
# This is how it works: Redis remembers the size of the AOF file after the
|
# This is how it works: Redis remembers the size of the AOF file after the
|
||||||
# latest rewrite (or if no rewrite happened since the restart, the size of
|
# latest rewrite (if no rewrite has happened since the restart, the size of
|
||||||
# the AOF at startup is used).
|
# the AOF at startup is used).
|
||||||
#
|
#
|
||||||
# This base size is compared to the current size. If the current size is
|
# This base size is compared to the current size. If the current size is
|
||||||
@@ -328,12 +505,30 @@ no-appendfsync-on-rewrite no
|
|||||||
# is useful to avoid rewriting the AOF file even if the percentage increase
|
# is useful to avoid rewriting the AOF file even if the percentage increase
|
||||||
# is reached but it is still pretty small.
|
# is reached but it is still pretty small.
|
||||||
#
|
#
|
||||||
# Specify a precentage of zero in order to disable the automatic AOF
|
# Specify a percentage of zero in order to disable the automatic AOF
|
||||||
# rewrite feature.
|
# rewrite feature.
|
||||||
|
|
||||||
auto-aof-rewrite-percentage 100
|
auto-aof-rewrite-percentage 100
|
||||||
auto-aof-rewrite-min-size 64mb
|
auto-aof-rewrite-min-size 64mb
|
||||||
|
|
||||||
|
################################ LUA SCRIPTING ###############################
|
||||||
|
|
||||||
|
# Max execution time of a Lua script in milliseconds.
|
||||||
|
#
|
||||||
|
# If the maximum execution time is reached Redis will log that a script is
|
||||||
|
# still in execution after the maximum allowed time and will start to
|
||||||
|
# reply to queries with an error.
|
||||||
|
#
|
||||||
|
# When a long running script exceed the maximum execution time only the
|
||||||
|
# SCRIPT KILL and SHUTDOWN NOSAVE commands are available. The first can be
|
||||||
|
# used to stop a script that did not yet called write commands. The second
|
||||||
|
# is the only way to shut down the server in the case a write commands was
|
||||||
|
# already issue by the script but the user don't want to wait for the natural
|
||||||
|
# termination of the script.
|
||||||
|
#
|
||||||
|
# Set it to 0 or a negative value for unlimited execution without warnings.
|
||||||
|
lua-time-limit 5000
|
||||||
|
|
||||||
################################## SLOW LOG ###################################
|
################################## SLOW LOG ###################################
|
||||||
|
|
||||||
# The Redis Slow Log is a system to log queries that exceeded a specified
|
# The Redis Slow Log is a system to log queries that exceeded a specified
|
||||||
@@ -358,88 +553,59 @@ slowlog-log-slower-than 10000
|
|||||||
# You can reclaim memory used by the slow log with SLOWLOG RESET.
|
# You can reclaim memory used by the slow log with SLOWLOG RESET.
|
||||||
slowlog-max-len 128
|
slowlog-max-len 128
|
||||||
|
|
||||||
################################ VIRTUAL MEMORY ###############################
|
############################# Event notification ##############################
|
||||||
|
|
||||||
### WARNING! Virtual Memory is deprecated in Redis 2.4
|
# Redis can notify Pub/Sub clients about events happening in the key space.
|
||||||
### The use of Virtual Memory is strongly discouraged.
|
# This feature is documented at http://redis.io/topics/keyspace-events
|
||||||
|
|
||||||
# Virtual Memory allows Redis to work with datasets bigger than the actual
|
|
||||||
# amount of RAM needed to hold the whole dataset in memory.
|
|
||||||
# In order to do so very used keys are taken in memory while the other keys
|
|
||||||
# are swapped into a swap file, similarly to what operating systems do
|
|
||||||
# with memory pages.
|
|
||||||
#
|
#
|
||||||
# To enable VM just set 'vm-enabled' to yes, and set the following three
|
# For instance if keyspace events notification is enabled, and a client
|
||||||
# VM parameters accordingly to your needs.
|
# performs a DEL operation on key "foo" stored in the Database 0, two
|
||||||
|
# messages will be published via Pub/Sub:
|
||||||
vm-enabled no
|
|
||||||
# vm-enabled yes
|
|
||||||
|
|
||||||
# This is the path of the Redis swap file. As you can guess, swap files
|
|
||||||
# can't be shared by different Redis instances, so make sure to use a swap
|
|
||||||
# file for every redis process you are running. Redis will complain if the
|
|
||||||
# swap file is already in use.
|
|
||||||
#
|
#
|
||||||
# The best kind of storage for the Redis swap file (that's accessed at random)
|
# PUBLISH __keyspace@0__:foo del
|
||||||
# is a Solid State Disk (SSD).
|
# PUBLISH __keyevent@0__:del foo
|
||||||
#
|
#
|
||||||
# *** WARNING *** if you are using a shared hosting the default of putting
|
# It is possible to select the events that Redis will notify among a set
|
||||||
# the swap file under /tmp is not secure. Create a dir with access granted
|
# of classes. Every class is identified by a single character:
|
||||||
# only to Redis user and configure Redis to create the swap file there.
|
|
||||||
vm-swap-file /var/lib/redis/redis.swap
|
|
||||||
|
|
||||||
# vm-max-memory configures the VM to use at max the specified amount of
|
|
||||||
# RAM. Everything that deos not fit will be swapped on disk *if* possible, that
|
|
||||||
# is, if there is still enough contiguous space in the swap file.
|
|
||||||
#
|
#
|
||||||
# With vm-max-memory 0 the system will swap everything it can. Not a good
|
# K Keyspace events, published with __keyspace@<db>__ prefix.
|
||||||
# default, just specify the max amount of RAM you can in bytes, but it's
|
# E Keyevent events, published with __keyevent@<db>__ prefix.
|
||||||
# better to leave some margin. For instance specify an amount of RAM
|
# g Generic commands (non-type specific) like DEL, EXPIRE, RENAME, ...
|
||||||
# that's more or less between 60 and 80% of your free RAM.
|
# $ String commands
|
||||||
vm-max-memory 0
|
# l List commands
|
||||||
|
# s Set commands
|
||||||
# Redis swap files is split into pages. An object can be saved using multiple
|
# h Hash commands
|
||||||
# contiguous pages, but pages can't be shared between different objects.
|
# z Sorted set commands
|
||||||
# So if your page is too big, small objects swapped out on disk will waste
|
# x Expired events (events generated every time a key expires)
|
||||||
# a lot of space. If you page is too small, there is less space in the swap
|
# e Evicted events (events generated when a key is evicted for maxmemory)
|
||||||
# file (assuming you configured the same number of total swap file pages).
|
# A Alias for g$lshzxe, so that the "AKE" string means all the events.
|
||||||
#
|
#
|
||||||
# If you use a lot of small objects, use a page size of 64 or 32 bytes.
|
# The "notify-keyspace-events" takes as argument a string that is composed
|
||||||
# If you use a lot of big objects, use a bigger page size.
|
# by zero or multiple characters. The empty string means that notifications
|
||||||
# If unsure, use the default :)
|
# are disabled at all.
|
||||||
vm-page-size 32
|
|
||||||
|
|
||||||
# Number of total memory pages in the swap file.
|
|
||||||
# Given that the page table (a bitmap of free/used pages) is taken in memory,
|
|
||||||
# every 8 pages on disk will consume 1 byte of RAM.
|
|
||||||
#
|
#
|
||||||
# The total swap size is vm-page-size * vm-pages
|
# Example: to enable list and generic events, from the point of view of the
|
||||||
|
# event name, use:
|
||||||
#
|
#
|
||||||
# With the default of 32-bytes memory pages and 134217728 pages Redis will
|
# notify-keyspace-events Elg
|
||||||
# use a 4 GB swap file, that will use 16 MB of RAM for the page table.
|
|
||||||
#
|
#
|
||||||
# It's better to use the smallest acceptable value for your application,
|
# Example 2: to get the stream of the expired keys subscribing to channel
|
||||||
# but the default is large in order to work in most conditions.
|
# name __keyevent@0__:expired use:
|
||||||
vm-pages 134217728
|
|
||||||
|
|
||||||
# Max number of VM I/O threads running at the same time.
|
|
||||||
# This threads are used to read/write data from/to swap file, since they
|
|
||||||
# also encode and decode objects from disk to memory or the reverse, a bigger
|
|
||||||
# number of threads can help with big objects even if they can't help with
|
|
||||||
# I/O itself as the physical device may not be able to couple with many
|
|
||||||
# reads/writes operations at the same time.
|
|
||||||
#
|
#
|
||||||
# The special value of 0 turn off threaded I/O and enables the blocking
|
# notify-keyspace-events Ex
|
||||||
# Virtual Memory implementation.
|
#
|
||||||
vm-max-threads 4
|
# By default all notifications are disabled because most users don't need
|
||||||
|
# this feature and the feature has some overhead. Note that if you don't
|
||||||
|
# specify at least one of K or E, no events will be delivered.
|
||||||
|
notify-keyspace-events ""
|
||||||
|
|
||||||
############################### ADVANCED CONFIG ###############################
|
############################### ADVANCED CONFIG ###############################
|
||||||
# Hashes are encoded in a special way (much more memory efficient) when they
|
|
||||||
# have at max a given numer of elements, and the biggest element does not
|
# Hashes are encoded using a memory efficient data structure when they have a
|
||||||
# exceed a given threshold. You can configure this limits with the following
|
# small number of entries, and the biggest entry does not exceed a given
|
||||||
# configuration directives.
|
# threshold. These thresholds can be configured using the following directives.
|
||||||
hash-max-zipmap-entries 512
|
hash-max-ziplist-entries 512
|
||||||
hash-max-zipmap-value 64
|
hash-max-ziplist-value 64
|
||||||
|
|
||||||
# Similarly to hashes, small lists are also encoded in a special way in order
|
# Similarly to hashes, small lists are also encoded in a special way in order
|
||||||
# to save a lot of space. The special representation is only used when
|
# to save a lot of space. The special representation is only used when
|
||||||
@@ -462,9 +628,9 @@ zset-max-ziplist-value 64
|
|||||||
|
|
||||||
# Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in
|
# Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in
|
||||||
# order to help rehashing the main Redis hash table (the one mapping top-level
|
# order to help rehashing the main Redis hash table (the one mapping top-level
|
||||||
# keys to values). The hash table implementation redis uses (see dict.c)
|
# keys to values). The hash table implementation Redis uses (see dict.c)
|
||||||
# performs a lazy rehashing: the more operation you run into an hash table
|
# performs a lazy rehashing: the more operation you run into a hash table
|
||||||
# that is rhashing, the more rehashing "steps" are performed, so if the
|
# that is rehashing, the more rehashing "steps" are performed, so if the
|
||||||
# server is idle the rehashing is never complete and some more memory is used
|
# server is idle the rehashing is never complete and some more memory is used
|
||||||
# by the hash table.
|
# by the hash table.
|
||||||
#
|
#
|
||||||
@@ -480,12 +646,65 @@ zset-max-ziplist-value 64
|
|||||||
# want to free memory asap when possible.
|
# want to free memory asap when possible.
|
||||||
activerehashing yes
|
activerehashing yes
|
||||||
|
|
||||||
################################## INCLUDES ###################################
|
# The client output buffer limits can be used to force disconnection of clients
|
||||||
|
# that are not reading data from the server fast enough for some reason (a
|
||||||
# Include one or more other config files here. This is useful if you
|
# common reason is that a Pub/Sub client can't consume messages as fast as the
|
||||||
# have a standard template that goes to all redis server but also need
|
# publisher can produce them).
|
||||||
# to customize a few per-server settings. Include files can include
|
|
||||||
# other files, so use this wisely.
|
|
||||||
#
|
#
|
||||||
# include /path/to/local.conf
|
# The limit can be set differently for the three different classes of clients:
|
||||||
# include /path/to/other.conf
|
#
|
||||||
|
# normal -> normal clients
|
||||||
|
# slave -> slave clients and MONITOR clients
|
||||||
|
# pubsub -> clients subscribed to at least one pubsub channel or pattern
|
||||||
|
#
|
||||||
|
# The syntax of every client-output-buffer-limit directive is the following:
|
||||||
|
#
|
||||||
|
# client-output-buffer-limit <class> <hard limit> <soft limit> <soft seconds>
|
||||||
|
#
|
||||||
|
# A client is immediately disconnected once the hard limit is reached, or if
|
||||||
|
# the soft limit is reached and remains reached for the specified number of
|
||||||
|
# seconds (continuously).
|
||||||
|
# So for instance if the hard limit is 32 megabytes and the soft limit is
|
||||||
|
# 16 megabytes / 10 seconds, the client will get disconnected immediately
|
||||||
|
# if the size of the output buffers reach 32 megabytes, but will also get
|
||||||
|
# disconnected if the client reaches 16 megabytes and continuously overcomes
|
||||||
|
# the limit for 10 seconds.
|
||||||
|
#
|
||||||
|
# By default normal clients are not limited because they don't receive data
|
||||||
|
# without asking (in a push way), but just after a request, so only
|
||||||
|
# asynchronous clients may create a scenario where data is requested faster
|
||||||
|
# than it can read.
|
||||||
|
#
|
||||||
|
# Instead there is a default limit for pubsub and slave clients, since
|
||||||
|
# subscribers and slaves receive data in a push fashion.
|
||||||
|
#
|
||||||
|
# Both the hard or the soft limit can be disabled by setting them to zero.
|
||||||
|
client-output-buffer-limit normal 0 0 0
|
||||||
|
client-output-buffer-limit slave 256mb 64mb 60
|
||||||
|
client-output-buffer-limit pubsub 32mb 8mb 60
|
||||||
|
|
||||||
|
# Redis calls an internal function to perform many background tasks, like
|
||||||
|
# closing connections of clients in timeout, purging expired keys that are
|
||||||
|
# never requested, and so forth.
|
||||||
|
#
|
||||||
|
# Not all tasks are performed with the same frequency, but Redis checks for
|
||||||
|
# tasks to perform accordingly to the specified "hz" value.
|
||||||
|
#
|
||||||
|
# By default "hz" is set to 10. Raising the value will use more CPU when
|
||||||
|
# Redis is idle, but at the same time will make Redis more responsive when
|
||||||
|
# there are many keys expiring at the same time, and timeouts may be
|
||||||
|
# handled with more precision.
|
||||||
|
#
|
||||||
|
# The range is between 1 and 500, however a value over 100 is usually not
|
||||||
|
# a good idea. Most users should use the default of 10 and raise this up to
|
||||||
|
# 100 only in environments where very low latency is required.
|
||||||
|
hz 10
|
||||||
|
|
||||||
|
# When a child rewrites the AOF file, if the following option is enabled
|
||||||
|
# the file will be fsync-ed every 32 MB of data generated. This is useful
|
||||||
|
# in order to commit the file to the disk more incrementally and avoid
|
||||||
|
# big latency spikes.
|
||||||
|
aof-rewrite-incremental-fsync yes
|
||||||
|
|
||||||
|
# Zulip-specific configuration: disable saving to disk.
|
||||||
|
save ""
|
||||||
|
|||||||
@@ -40,14 +40,6 @@ class zulip::base {
|
|||||||
group => 'zulip',
|
group => 'zulip',
|
||||||
}
|
}
|
||||||
|
|
||||||
file { '/etc/puppet/puppet.conf':
|
|
||||||
ensure => file,
|
|
||||||
mode => 640,
|
|
||||||
owner => "root",
|
|
||||||
group => "root",
|
|
||||||
source => 'puppet:///modules/zulip/puppet.conf',
|
|
||||||
}
|
|
||||||
|
|
||||||
file { '/etc/security/limits.conf':
|
file { '/etc/security/limits.conf':
|
||||||
ensure => file,
|
ensure => file,
|
||||||
mode => 640,
|
mode => 640,
|
||||||
@@ -56,6 +48,13 @@ class zulip::base {
|
|||||||
source => 'puppet:///modules/zulip/limits.conf',
|
source => 'puppet:///modules/zulip/limits.conf',
|
||||||
}
|
}
|
||||||
|
|
||||||
|
# This directory is written to by cron jobs for reading by Nagios
|
||||||
|
file { '/var/lib/nagios_state/':
|
||||||
|
ensure => directory,
|
||||||
|
group => 'zulip',
|
||||||
|
mode => 774,
|
||||||
|
}
|
||||||
|
|
||||||
file { '/var/log/zulip':
|
file { '/var/log/zulip':
|
||||||
ensure => 'directory',
|
ensure => 'directory',
|
||||||
owner => 'zulip',
|
owner => 'zulip',
|
||||||
|
|||||||
@@ -4,4 +4,4 @@ if [ "$(hostname)" = "staging.zulip.net" ]; then
|
|||||||
else
|
else
|
||||||
site="https://api.zulip.com"
|
site="https://api.zulip.com"
|
||||||
fi
|
fi
|
||||||
/home/zulip/deployments/current/bots/check_send_receive.py --munin $1 --site="$site"
|
/home/zulip/deployments/current/bots/check_send_receive.py --munin "$1" --site="$site"
|
||||||
|
|||||||
@@ -9,11 +9,11 @@ cd /home/zulip/deployments/current
|
|||||||
|
|
||||||
BACKLOG=$(./manage.py print_email_delivery_backlog)
|
BACKLOG=$(./manage.py print_email_delivery_backlog)
|
||||||
|
|
||||||
if [ $BACKLOG -gt 0 -a $BACKLOG -lt 10 ]
|
if [ "$BACKLOG" -gt 0 ] && [ "$BACKLOG" -lt 10 ]
|
||||||
then
|
then
|
||||||
echo "backlog of $BACKLOG"
|
echo "backlog of $BACKLOG"
|
||||||
exit 1
|
exit 1
|
||||||
elif [ $BACKLOG -ge 10 ]
|
elif [ "$BACKLOG" -ge 10 ]
|
||||||
then
|
then
|
||||||
echo "backlog of $BACKLOG"
|
echo "backlog of $BACKLOG"
|
||||||
exit 2
|
exit 2
|
||||||
|
|||||||
@@ -13,14 +13,14 @@ if [ "$STATUS" == "RUNNING" ]
|
|||||||
then
|
then
|
||||||
echo "Running"
|
echo "Running"
|
||||||
exit 0
|
exit 0
|
||||||
elif [ $(echo "$STATUS" | egrep '(STOPPED)|(STARTING)|(BACKOFF)|(STOPPING)|(EXITED)|(FATAL)|(UNKNOWN)$') ]
|
elif [ "$(echo "$STATUS" | egrep '(STOPPED)|(STARTING)|(BACKOFF)|(STOPPING)|(EXITED)|(FATAL)|(UNKNOWN)$')" ]
|
||||||
then
|
then
|
||||||
# not "RUNNING", but a recognized supervisor status
|
# not "RUNNING", but a recognized supervisor status
|
||||||
echo $STATUS
|
echo "$STATUS"
|
||||||
exit 1
|
exit 1
|
||||||
else
|
else
|
||||||
# we don't recognize the second column in this SUPERVISOR_STATUS.
|
# we don't recognize the second column in this SUPERVISOR_STATUS.
|
||||||
# This may be indicative of a supervisor configuration problem
|
# This may be indicative of a supervisor configuration problem
|
||||||
echo $SUPERVISOR_STATUS
|
echo "$SUPERVISOR_STATUS"
|
||||||
exit 3
|
exit 3
|
||||||
fi
|
fi
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
# Checks for any Zulip queue workers that are leaking memory and thus have a high vsize
|
# Checks for any Zulip queue workers that are leaking memory and thus have a high vsize
|
||||||
datafile=$(mktemp)
|
datafile=$(mktemp)
|
||||||
ps -o vsize,size,pid,user,command --sort -vsize $(pgrep -f '^python /home/zulip/deployments/current/manage.py process_queue') > "$datafile"
|
ps -o vsize,size,pid,user,command --sort -vsize "$(pgrep -f '^python /home/zulip/deployments/current/manage.py process_queue')" > "$datafile"
|
||||||
cat "$datafile"
|
cat "$datafile"
|
||||||
top_worker=$(cat "$datafile" | head -n2 | tail -n1)
|
top_worker=$(cat "$datafile" | head -n2 | tail -n1)
|
||||||
top_worker_memory_usage=$(echo "$top_worker" | cut -f1 -d" ")
|
top_worker_memory_usage=$(echo "$top_worker" | cut -f1 -d" ")
|
||||||
|
|||||||
@@ -5,9 +5,9 @@ fi
|
|||||||
|
|
||||||
export AWS_ACCESS_KEY_ID=xxxxxxxxxxxxxxxxxxxx
|
export AWS_ACCESS_KEY_ID=xxxxxxxxxxxxxxxxxxxx
|
||||||
export AWS_SECRET_ACCESS_KEY=xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
|
export AWS_SECRET_ACCESS_KEY=xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
|
||||||
s3_backup_bucket=$(crudini --get $ZULIP_CONF database s3_backup_bucket 2>&1)
|
s3_backup_bucket=$(crudini --get "$ZULIP_CONF" database s3_backup_bucket 2>&1)
|
||||||
if [ $? -ne 0 ]; then
|
if [ $? -ne 0 ]; then
|
||||||
echo "Could not determine which s3 bucket to use:" $s3_backup_bucket
|
echo "Could not determine which s3 bucket to use:" "$s3_backup_bucket"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
export WALE_S3_PREFIX=s3://$s3_backup_bucket
|
export WALE_S3_PREFIX=s3://$s3_backup_bucket
|
||||||
|
|||||||
@@ -102,13 +102,6 @@ class zulip_internal::base {
|
|||||||
group => "nagios",
|
group => "nagios",
|
||||||
mode => 600,
|
mode => 600,
|
||||||
}
|
}
|
||||||
file { '/var/lib/nagios_state/':
|
|
||||||
ensure => directory,
|
|
||||||
require => User['nagios'],
|
|
||||||
owner => "nagios",
|
|
||||||
group => "nagios",
|
|
||||||
mode => 777,
|
|
||||||
}
|
|
||||||
file { '/var/lib/nagios/.ssh':
|
file { '/var/lib/nagios/.ssh':
|
||||||
ensure => directory,
|
ensure => directory,
|
||||||
require => File['/var/lib/nagios/'],
|
require => File['/var/lib/nagios/'],
|
||||||
|
|||||||
@@ -12,13 +12,7 @@ EOF
|
|||||||
|
|
||||||
apt-get update
|
apt-get update
|
||||||
apt-get -y dist-upgrade
|
apt-get -y dist-upgrade
|
||||||
apt-get install -y puppet git
|
apt-get install -y puppet git python
|
||||||
cp -a /root/zulip/puppet/zulip/files/puppet.conf /etc/puppet/
|
|
||||||
|
|
||||||
# Hack to get required python-django-guardian while the PPA build of django-guardian isn't working :(
|
|
||||||
apt-get -y install python-django
|
|
||||||
wget https://zulip.com/dist/packages/python-django-guardian_1.3-1~zulip4_all.deb
|
|
||||||
dpkg -i ./python-django-guardian_1.3-1~zulip4_all.deb
|
|
||||||
|
|
||||||
mkdir -p /etc/zulip
|
mkdir -p /etc/zulip
|
||||||
echo -e "[machine]\npuppet_classes = zulip::voyager\ndeploy_type = voyager" > /etc/zulip/zulip.conf
|
echo -e "[machine]\npuppet_classes = zulip::voyager\ndeploy_type = voyager" > /etc/zulip/zulip.conf
|
||||||
|
|||||||
@@ -3,9 +3,9 @@
|
|||||||
# Delete the "guest" default user and replace it with a Zulip user
|
# Delete the "guest" default user and replace it with a Zulip user
|
||||||
# with a real password
|
# with a real password
|
||||||
|
|
||||||
RMQPW=$($(dirname $0)/../../bin/get-django-setting RABBITMQ_PASSWORD)
|
RMQPW=$("$(dirname "$0")/../../bin/get-django-setting" RABBITMQ_PASSWORD)
|
||||||
sudo rabbitmqctl delete_user zulip || true
|
sudo rabbitmqctl delete_user zulip || true
|
||||||
sudo rabbitmqctl delete_user guest || true
|
sudo rabbitmqctl delete_user guest || true
|
||||||
sudo rabbitmqctl add_user zulip $RMQPW
|
sudo rabbitmqctl add_user zulip "$RMQPW"
|
||||||
sudo rabbitmqctl set_user_tags zulip administrator
|
sudo rabbitmqctl set_user_tags zulip administrator
|
||||||
sudo rabbitmqctl set_permissions -p / zulip '.*' '.*' '.*'
|
sudo rabbitmqctl set_permissions -p / zulip '.*' '.*' '.*'
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#!/bin/bash -xe
|
#!/bin/bash -xe
|
||||||
|
|
||||||
# Change to root directory of the checkout that we're running from
|
# Change to root directory of the checkout that we're running from
|
||||||
cd $(dirname $0)/../..
|
cd "$(dirname "$0")/../.."
|
||||||
|
|
||||||
python manage.py checkconfig
|
python manage.py checkconfig
|
||||||
|
|
||||||
|
|||||||
@@ -1,3 +1,3 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
mkdir -p /var/log/zulip
|
mkdir -p /var/log/zulip
|
||||||
$(dirname $(dirname $0))/lib/install "$@" 2>&1 | tee -a /var/log/zulip/install.log
|
"$(dirname "$(dirname "$0")")/lib/install" "$@" 2>&1 | tee -a /var/log/zulip/install.log
|
||||||
|
|||||||
@@ -1,2 +1,2 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
$(dirname $0)/lib/upgrade-zulip "$@" 2>&1 | tee -a /var/log/zulip/upgrade.log
|
"$(dirname "$0")/lib/upgrade-zulip" "$@" 2>&1 | tee -a /var/log/zulip/upgrade.log
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ include apt
|
|||||||
for pclass in re.split(r'\s*,\s*', config.get('machine', 'puppet_classes')):
|
for pclass in re.split(r'\s*,\s*', config.get('machine', 'puppet_classes')):
|
||||||
puppet_config += "include %s\n" % (pclass,)
|
puppet_config += "include %s\n" % (pclass,)
|
||||||
|
|
||||||
puppet_cmd = ["puppet", "apply", "-e", puppet_config]
|
puppet_cmd = ["puppet", "apply", "--modulepath=/root/zulip/puppet", "-e", puppet_config]
|
||||||
puppet_cmd += extra_args
|
puppet_cmd += extra_args
|
||||||
|
|
||||||
if force:
|
if force:
|
||||||
|
|||||||
@@ -26,7 +26,7 @@
|
|||||||
code changes, issue tickets, build system results, and much more. If you don't see the system you would like to integrate with it, or run into any
|
code changes, issue tickets, build system results, and much more. If you don't see the system you would like to integrate with it, or run into any
|
||||||
trouble, don't hesitate to <a href="mailto:support@zulip.com?subject=Integration%20question">email us</a>.</p>
|
trouble, don't hesitate to <a href="mailto:support@zulip.com?subject=Integration%20question">email us</a>.</p>
|
||||||
|
|
||||||
<p>Many of these integrations require creating a Zulip bot. You can do so on your <a href="https://zulip.com/#settings">Zulip settings page</a>. Be sure to note its username and API key.</p>
|
<p>Many of these integrations require creating a Zulip bot. You can do so on your <a href="/#settings">Zulip settings page</a>. Be sure to note its username and API key.</p>
|
||||||
|
|
||||||
<div id="integration-instruction-block" class="integration-instruction-block">
|
<div id="integration-instruction-block" class="integration-instruction-block">
|
||||||
<a href="#" id="integration-list-link"><i class="icon-vector-circle-arrow-left"></i> Back to list</a>
|
<a href="#" id="integration-list-link"><i class="icon-vector-circle-arrow-left"></i> Back to list</a>
|
||||||
@@ -852,7 +852,7 @@
|
|||||||
<pre><code>npm install --save hubot-zulip</code></pre>
|
<pre><code>npm install --save hubot-zulip</code></pre>
|
||||||
</li>
|
</li>
|
||||||
|
|
||||||
<li><p>On your <a href="https://zulip.com/#settings">Zulip settings page</a>, create a bot account. Note its username, API key and full name; you will use them on the next step.</p></li>
|
<li><p>On your <a href="/#settings">Zulip settings page</a>, create a bot account. Note its username, API key and full name; you will use them on the next step.</p></li>
|
||||||
|
|
||||||
<li>To run Hubot locally, use:
|
<li>To run Hubot locally, use:
|
||||||
<pre><code>HUBOT_ZULIP_BOT=hubot-bot@example.com HUBOT_ZULIP_API_KEY=your_key bin/hubot --adapter zulip --name "myhubot"</code></pre>
|
<pre><code>HUBOT_ZULIP_BOT=hubot-bot@example.com HUBOT_ZULIP_API_KEY=your_key bin/hubot --adapter zulip --name "myhubot"</code></pre>
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ dist=$1
|
|||||||
file=$2
|
file=$2
|
||||||
|
|
||||||
if [ -z "$dist" ] || [ -z "$file" ]; then
|
if [ -z "$dist" ] || [ -z "$file" ]; then
|
||||||
echo "$(echo $0 | rev | cut -d "/" -f 1-1 | rev) -- build Debian packages on a Zulip buildslave"
|
echo "$(echo "$0" | rev | cut -d "/" -f 1-1 | rev) -- build Debian packages on a Zulip buildslave"
|
||||||
echo
|
echo
|
||||||
echo "USAGE: $0 dist path/to/package.dsc"
|
echo "USAGE: $0 dist path/to/package.dsc"
|
||||||
exit 1
|
exit 1
|
||||||
@@ -21,10 +21,10 @@ fi
|
|||||||
set -xe
|
set -xe
|
||||||
|
|
||||||
ret=0
|
ret=0
|
||||||
path=$(ssh -q -l $BUILDD_USERNAME $BUILDD_HOST -- mktemp -d $BUILDD_BASE_PATH/$USER.`date -u +%F.%R`.XXXXXXX)/
|
path=$(ssh -q -l "$BUILDD_USERNAME" "$BUILDD_HOST" -- "mktemp -d '$BUILDD_BASE_PATH/$USER.`date -u +%F.%R`.XXXXXXX'")/
|
||||||
|
|
||||||
dcmd rsync -vz --copy-links $file $BUILDD_USERNAME@$BUILDD_HOST:$path/
|
dcmd rsync -vz --copy-links "$file" "$BUILDD_USERNAME@$BUILDD_HOST:$path/"
|
||||||
file=$(basename $file)
|
file=$(basename "$file")
|
||||||
|
|
||||||
# -A specifies to build arch-all packages (non-arch dependent) in addition to
|
# -A specifies to build arch-all packages (non-arch dependent) in addition to
|
||||||
# binary packages
|
# binary packages
|
||||||
@@ -38,9 +38,9 @@ file=$(basename $file)
|
|||||||
#
|
#
|
||||||
# We always build for amd64. There is no 32-bit.
|
# We always build for amd64. There is no 32-bit.
|
||||||
|
|
||||||
ssh -t -l $BUILDD_USERNAME $BUILDD_HOST -- "cd $path && sbuild -A -s --force-orig-source --dist=$dist --arch=amd64 $file"
|
ssh -t -l "$BUILDD_USERNAME" "$BUILDD_HOST" -- "cd '$path' && sbuild -A -s --force-orig-source '--dist=$dist' --arch=amd64 '$file'"
|
||||||
|
|
||||||
rsync -Lvz --copy-links $BUILDD_USERNAME@$BUILDD_HOST:$path/* .
|
rsync -Lvz --copy-links "$BUILDD_USERNAME@$BUILDD_HOST:$path/*" .
|
||||||
ssh -l $BUILDD_USERNAME $BUILDD_HOST rm -r $path
|
ssh -l "$BUILDD_USERNAME" "$BUILDD_HOST" rm -r "$path"
|
||||||
|
|
||||||
exit $ret
|
exit "$ret"
|
||||||
|
|||||||
@@ -11,8 +11,8 @@ prefix="zulip-server-$version"
|
|||||||
|
|
||||||
if [ "$(uname)" = "Darwin" ]; then
|
if [ "$(uname)" = "Darwin" ]; then
|
||||||
TMPDIR=/tmp/voyager-build
|
TMPDIR=/tmp/voyager-build
|
||||||
rm -Rf $TMPDIR
|
rm -Rf "$TMPDIR"
|
||||||
mkdir -p $TMPDIR
|
mkdir -p "$TMPDIR"
|
||||||
else
|
else
|
||||||
TMPDIR=$(mktemp -d)
|
TMPDIR=$(mktemp -d)
|
||||||
fi
|
fi
|
||||||
@@ -27,10 +27,10 @@ TMP_CHECKOUT=$TMPDIR/$prefix/
|
|||||||
TARBALL=$TMPDIR/$prefix.tar
|
TARBALL=$TMPDIR/$prefix.tar
|
||||||
|
|
||||||
# .gitattributes lists the files that are not exported
|
# .gitattributes lists the files that are not exported
|
||||||
git archive -o $TARBALL --prefix=$prefix/ HEAD
|
git archive -o "$TARBALL" "--prefix=$prefix/" HEAD
|
||||||
|
|
||||||
|
|
||||||
if tar -tf $TARBALL | grep -q -e zilencer -e zproject/local_settings.py -e puppet/zulip_internal; then
|
if tar -tf "$TARBALL" | grep -q -e zilencer -e zproject/local_settings.py -e puppet/zulip_internal; then
|
||||||
echo "Excluded files remain in tarball!";
|
echo "Excluded files remain in tarball!";
|
||||||
echo "Versions of git 1.8.1.1 - 1.8.1.6 have broken .gitattributes syntax";
|
echo "Versions of git 1.8.1.1 - 1.8.1.6 have broken .gitattributes syntax";
|
||||||
exit 1;
|
exit 1;
|
||||||
@@ -39,8 +39,8 @@ else
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
# Check out a temporary full copy of the index to generate static files
|
# Check out a temporary full copy of the index to generate static files
|
||||||
git checkout-index -f -a --prefix $TMP_CHECKOUT
|
git checkout-index -f -a --prefix "$TMP_CHECKOUT"
|
||||||
cd $TMP_CHECKOUT
|
cd "$TMP_CHECKOUT"
|
||||||
|
|
||||||
# Use default settings so there is no chance of leaking secrets
|
# Use default settings so there is no chance of leaking secrets
|
||||||
cp zproject/local_settings_template.py zproject/local_settings.py
|
cp zproject/local_settings_template.py zproject/local_settings.py
|
||||||
@@ -70,8 +70,8 @@ echo; echo "\033[33mRunning update-prod-static; check ${TMP_CHECKOUT}update-prod
|
|||||||
set -x
|
set -x
|
||||||
|
|
||||||
./tools/update-prod-static
|
./tools/update-prod-static
|
||||||
echo $GITID > build_id
|
echo "$GITID" > build_id
|
||||||
echo $version > version
|
echo "$version" > version
|
||||||
mv update-prod-static.log ..
|
mv update-prod-static.log ..
|
||||||
|
|
||||||
rm -f zproject/dev-secrets.conf
|
rm -f zproject/dev-secrets.conf
|
||||||
@@ -79,11 +79,11 @@ rm -f zproject/dev-secrets.conf
|
|||||||
# We don't need duplicate copies of emoji with hashed paths, and they would break bugdown
|
# We don't need duplicate copies of emoji with hashed paths, and they would break bugdown
|
||||||
find prod-static/serve/third/gemoji/images/emoji/ -regex '.*\.[0-9a-f]+\.png' -delete
|
find prod-static/serve/third/gemoji/images/emoji/ -regex '.*\.[0-9a-f]+\.png' -delete
|
||||||
|
|
||||||
cd $TMPDIR
|
cd "$TMPDIR"
|
||||||
|
|
||||||
tar --append -f $TARBALL $prefix/prod-static $prefix/build_id $prefix/version
|
tar --append -f "$TARBALL" "$prefix/prod-static" "$prefix/build_id" "$prefix/version"
|
||||||
|
|
||||||
rm -rf $prefix
|
rm -rf "$prefix"
|
||||||
|
|
||||||
gzip $TARBALL
|
gzip "$TARBALL"
|
||||||
echo "Generated $TARBALL.gz"
|
echo "Generated $TARBALL.gz"
|
||||||
|
|||||||
@@ -4,7 +4,7 @@
|
|||||||
# and also any branches in origin which are ancestors of
|
# and also any branches in origin which are ancestors of
|
||||||
# origin/master and are named like $USER-*.
|
# origin/master and are named like $USER-*.
|
||||||
|
|
||||||
push_args=''
|
push_args=()
|
||||||
|
|
||||||
function is_merged {
|
function is_merged {
|
||||||
! git rev-list -n 1 origin/master.."$1" | grep -q .
|
! git rev-list -n 1 origin/master.."$1" | grep -q .
|
||||||
@@ -31,7 +31,7 @@ function clean_ref {
|
|||||||
echo -n "Deleting remote branch $remote_name"
|
echo -n "Deleting remote branch $remote_name"
|
||||||
echo " (was $(git rev-parse --short "$ref"))"
|
echo " (was $(git rev-parse --short "$ref"))"
|
||||||
# NB: this won't handle spaces in ref names
|
# NB: this won't handle spaces in ref names
|
||||||
push_args="$push_args :$remote_name"
|
push_args=("${push_args[@]}" ":$remote_name")
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
@@ -44,8 +44,8 @@ fi
|
|||||||
|
|
||||||
git fetch --prune origin
|
git fetch --prune origin
|
||||||
|
|
||||||
eval $(git for-each-ref --shell --format='clean_ref %(refname);')
|
eval "$(git for-each-ref --shell --format='clean_ref %(refname);')"
|
||||||
|
|
||||||
if [ -n "$push_args" ]; then
|
if [ "${#push_args}" -ne 0 ]; then
|
||||||
git push origin $push_args
|
git push origin "${push_args[@]}"
|
||||||
fi
|
fi
|
||||||
|
|||||||
@@ -2,7 +2,7 @@
|
|||||||
|
|
||||||
function error_out {
|
function error_out {
|
||||||
echo -en '\e[0;31m'
|
echo -en '\e[0;31m'
|
||||||
echo $1
|
echo "$1"
|
||||||
echo -en '\e[0m'
|
echo -en '\e[0m'
|
||||||
exit 1
|
exit 1
|
||||||
}
|
}
|
||||||
@@ -12,16 +12,16 @@ status=$(git status --porcelain | grep -v '^??')
|
|||||||
|
|
||||||
old_ref=$(git rev-list --max-count=1 HEAD)
|
old_ref=$(git rev-list --max-count=1 HEAD)
|
||||||
branch=$1
|
branch=$1
|
||||||
branch_ref=$(git rev-list --max-count=1 $branch)
|
branch_ref=$(git rev-list --max-count=1 "$branch")
|
||||||
|
|
||||||
[ $? -ne 0 ] && error_out "Unknown branch: $branch"
|
[ $? -ne 0 ] && error_out "Unknown branch: $branch"
|
||||||
|
|
||||||
if [ "$old_ref" == "$branch_ref" ]; then
|
if [ "$old_ref" == "$branch_ref" ]; then
|
||||||
new_ref=master
|
new_ref=master
|
||||||
else
|
else
|
||||||
ref_name=$(git describe --all --exact $old_ref)
|
ref_name=$(git describe --all --exact "$old_ref")
|
||||||
if [ $? -eq 0 ]; then
|
if [ $? -eq 0 ]; then
|
||||||
new_ref=$(echo $ref_name | perl -pe 's{^(heads|remotes)/}{}')
|
new_ref=$(echo "$ref_name" | perl -pe 's{^(heads|remotes)/}{}')
|
||||||
else
|
else
|
||||||
new_ref=$old_ref
|
new_ref=$old_ref
|
||||||
fi
|
fi
|
||||||
@@ -31,13 +31,13 @@ fi
|
|||||||
|
|
||||||
git fetch -p
|
git fetch -p
|
||||||
|
|
||||||
git rebase origin/master $branch
|
git rebase origin/master "$branch"
|
||||||
[ $? -ne 0 ] && error_out "Rebase onto origin/master failed"
|
[ $? -ne 0 ] && error_out "Rebase onto origin/master failed"
|
||||||
|
|
||||||
git push . HEAD:master
|
git push . HEAD:master
|
||||||
git push origin master
|
git push origin master
|
||||||
[ $? -ne 0 ] && error_out "Push of master to origin/master failed"
|
[ $? -ne 0 ] && error_out "Push of master to origin/master failed"
|
||||||
|
|
||||||
git checkout $new_ref
|
git checkout "$new_ref"
|
||||||
git branch -D $branch
|
git branch -D "$branch"
|
||||||
git push origin :$branch
|
git push origin ":$branch"
|
||||||
|
|||||||
@@ -15,23 +15,23 @@ exit 1
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
if [[ $# < 1 || $# >2 ]]; then
|
if [[ $# -lt 1 || $# -gt 2 ]]; then
|
||||||
usage
|
usage
|
||||||
fi
|
fi
|
||||||
|
|
||||||
function get_status {
|
function get_status {
|
||||||
if (ssh zulip@$1.zulip.net '[ -d "'"$lockdir"'" ]'); then
|
if (ssh "zulip@$1.zulip.net" '[ -d "'"$lockdir"'" ]'); then
|
||||||
printf "%-10s %b" "$1" "is currently \e[31mlocked\e[0m\n"
|
printf "%-10s %b" "$1" "is currently \e[31mlocked\e[0m\n"
|
||||||
else
|
else
|
||||||
printf "%-10s %b" "$1" "is currently \e[32munlocked\e[0m\n"
|
printf "%-10s %b" "$1" "is currently \e[32munlocked\e[0m\n"
|
||||||
fi
|
fi
|
||||||
}
|
}
|
||||||
|
|
||||||
if [[ $1 == "lock" ]]; then
|
if [[ "$1" == "lock" ]]; then
|
||||||
verb="mkdir"
|
verb="mkdir"
|
||||||
elif [[ $1 == "unlock" ]]; then
|
elif [[ "$1" == "unlock" ]]; then
|
||||||
verb="rmdir"
|
verb="rmdir"
|
||||||
elif [[ $# == 1 && $1 == "status" ]]; then
|
elif [[ $# == 1 && "$1" == "status" ]]; then
|
||||||
get_status "staging"
|
get_status "staging"
|
||||||
get_status "prod0"
|
get_status "prod0"
|
||||||
exit
|
exit
|
||||||
@@ -40,11 +40,11 @@ else
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
|
|
||||||
if [[ $2 == "staging" ]]; then
|
if [[ "$2" == "staging" ]]; then
|
||||||
ssh zulip@staging.zulip.net "$verb $lockdir"
|
ssh zulip@staging.zulip.net "$verb $lockdir"
|
||||||
get_status "staging"
|
get_status "staging"
|
||||||
exit
|
exit
|
||||||
elif [[ $2 == "prod" ]]; then
|
elif [[ "$2" == "prod" ]]; then
|
||||||
ssh zulip@prod0.zulip.net "$verb $lockdir"
|
ssh zulip@prod0.zulip.net "$verb $lockdir"
|
||||||
get_status "prod0"
|
get_status "prod0"
|
||||||
exit
|
exit
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ COUNT=50
|
|||||||
mkdir -p output
|
mkdir -p output
|
||||||
while true; do
|
while true; do
|
||||||
|
|
||||||
if python show-last-messages --api-key=$API_KEY --user=$BOT_EMAIL --streams="$STREAMS" --count=$COUNT; then
|
if python show-last-messages --api-key="$API_KEY" --user="$BOT_EMAIL" --streams="$STREAMS" --count="$COUNT"; then
|
||||||
echo "[`date`] Success";
|
echo "[`date`] Success";
|
||||||
mv output-candidate.html output/zulip.html
|
mv output-candidate.html output/zulip.html
|
||||||
touch output/zulip.html
|
touch output/zulip.html
|
||||||
|
|||||||
@@ -21,5 +21,5 @@ if [ $# = 2 ]; then
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
for dist in $DISTS; do
|
for dist in $DISTS; do
|
||||||
reprepro --ignore=wrongdistribution include$type $dist "$1"
|
reprepro --ignore=wrongdistribution "include$TYPE" "$dist" "$1"
|
||||||
done
|
done
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ if [ -z "$hostname" ]; then
|
|||||||
echo "USAGE: $0 server type hostname [branch]"
|
echo "USAGE: $0 server type hostname [branch]"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
if ! $(echo "$hostname" | grep -q zulip); then
|
if ! echo "$hostname" | grep -q zulip; then
|
||||||
echo "USAGE: $0 server type hostname [branch]"
|
echo "USAGE: $0 server type hostname [branch]"
|
||||||
echo "Hostname must have zulip in it."
|
echo "Hostname must have zulip in it."
|
||||||
exit 1
|
exit 1
|
||||||
@@ -35,11 +35,11 @@ fi
|
|||||||
# Force RSA keys. We do this because the ECDSA key is not printed on syslog,
|
# Force RSA keys. We do this because the ECDSA key is not printed on syslog,
|
||||||
# and our puppet configuration does not use ECDSA. If we don't do this,
|
# and our puppet configuration does not use ECDSA. If we don't do this,
|
||||||
# we'll get key errors after puppet apply.
|
# we'll get key errors after puppet apply.
|
||||||
SSH_OPTS="-o HostKeyAlgorithms=ssh-rsa"
|
SSH_OPTS=(-o HostKeyAlgorithms=ssh-rsa)
|
||||||
|
|
||||||
set +e
|
set +e
|
||||||
|
|
||||||
ssh $SSH_OPTS "$server" -t -i "$amazon_key_file" -ladmin -o "ControlMaster=no" <<EOF
|
ssh "${SSH_OPTS[@]}" "$server" -t -i "$amazon_key_file" -ladmin -o "ControlMaster=no" <<EOF
|
||||||
sudo sed -i 's/PermitRootLogin no/PermitRootLogin yes/g' /etc/ssh/sshd_config
|
sudo sed -i 's/PermitRootLogin no/PermitRootLogin yes/g' /etc/ssh/sshd_config
|
||||||
sudo mkdir -p ~root/.ssh && sudo cp .ssh/authorized_keys ~root/.ssh/authorized_keys
|
sudo mkdir -p ~root/.ssh && sudo cp .ssh/authorized_keys ~root/.ssh/authorized_keys
|
||||||
sudo service ssh restart
|
sudo service ssh restart
|
||||||
@@ -48,7 +48,7 @@ EOF
|
|||||||
|
|
||||||
set -e
|
set -e
|
||||||
|
|
||||||
ssh $SSH_OPTS "$server" -t -i "$amazon_key_file" -lroot <<EOF
|
ssh "${SSH_OPTS[@]}" "$server" -t -i "$amazon_key_file" -lroot <<EOF
|
||||||
resize2fs /dev/xvda1
|
resize2fs /dev/xvda1
|
||||||
echo "$hostname" > /etc/hostname
|
echo "$hostname" > /etc/hostname
|
||||||
sed -i 's/localhost$/localhost $hostname/' /etc/hosts
|
sed -i 's/localhost$/localhost $hostname/' /etc/hosts
|
||||||
@@ -69,9 +69,9 @@ EOF
|
|||||||
|
|
||||||
# Give new server git access
|
# Give new server git access
|
||||||
# TODO: Don't give servers push access to our git!
|
# TODO: Don't give servers push access to our git!
|
||||||
scp $SSH_OPTS -i "$amazon_key_file" "$server_private_key_file" root@"$server":/root/.ssh/id_rsa
|
scp "${SSH_OPTS[@]}" -i "$amazon_key_file" "$server_private_key_file" root@"$server":/root/.ssh/id_rsa
|
||||||
|
|
||||||
ssh $SSH_OPTS "$server" -t -i "$amazon_key_file" -lroot <<EOF
|
ssh "${SSH_OPTS[@]}" "$server" -t -i "$amazon_key_file" -lroot <<EOF
|
||||||
chmod 600 /root/.ssh/id_rsa
|
chmod 600 /root/.ssh/id_rsa
|
||||||
# Setup initial known_hosts including git server
|
# Setup initial known_hosts including git server
|
||||||
cat > /root/.ssh/known_hosts <<EOF2
|
cat > /root/.ssh/known_hosts <<EOF2
|
||||||
@@ -85,9 +85,7 @@ cd /root/zulip
|
|||||||
git checkout $branch
|
git checkout $branch
|
||||||
EOF
|
EOF
|
||||||
|
|
||||||
ssh $SSH_OPTS "$server" -t -i "$amazon_key_file" -lroot <<EOF
|
ssh "${SSH_OPTS[@]}" "$server" -t -i "$amazon_key_file" -lroot <<EOF
|
||||||
cp -a /root/zulip/puppet/zulip/files/puppet.conf /etc/puppet/
|
|
||||||
|
|
||||||
userdel admin
|
userdel admin
|
||||||
passwd -d root
|
passwd -d root
|
||||||
mkdir /etc/zulip
|
mkdir /etc/zulip
|
||||||
@@ -107,8 +105,8 @@ fi
|
|||||||
EOF
|
EOF
|
||||||
|
|
||||||
# TODO: Don't give servers push access to our git!
|
# TODO: Don't give servers push access to our git!
|
||||||
scp $SSH_OPTS -i "$amazon_key_file" "$server_private_key_file" zulip@"$server":/home/zulip/.ssh/id_rsa
|
scp "${SSH_OPTS[@]}" -i "$amazon_key_file" "$server_private_key_file" zulip@"$server":/home/zulip/.ssh/id_rsa
|
||||||
ssh $SSH_OPTS "$server" -t -i "$amazon_key_file" -lzulip <<EOF
|
ssh "${SSH_OPTS[@]}" "$server" -t -i "$amazon_key_file" -lzulip <<EOF
|
||||||
chmod 600 /home/zulip/.ssh/id_rsa
|
chmod 600 /home/zulip/.ssh/id_rsa
|
||||||
EOF
|
EOF
|
||||||
|
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
#!/bin/sh -xe
|
#!/bin/sh -xe
|
||||||
|
|
||||||
# This is a really simple wrapper script, pretty much for documenting clarity
|
# This is a really simple wrapper script, pretty much for documenting clarity
|
||||||
`dirname $0`/../tools/generate-fixtures --force
|
"`dirname "$0"`/../tools/generate-fixtures" --force
|
||||||
|
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ function migration_status {
|
|||||||
|
|
||||||
template_grep_error_code=$(echo "SELECT 1 from pg_database WHERE datname='zulip_test_template';" | python manage.py dbshell --settings=zproject.test_settings | grep -q "1 row"; echo $?)
|
template_grep_error_code=$(echo "SELECT 1 from pg_database WHERE datname='zulip_test_template';" | python manage.py dbshell --settings=zproject.test_settings | grep -q "1 row"; echo $?)
|
||||||
|
|
||||||
if [ $template_grep_error_code == "0" ]; then
|
if [ "$template_grep_error_code" == "0" ]; then
|
||||||
migration_status zerver/fixtures/available-migrations
|
migration_status zerver/fixtures/available-migrations
|
||||||
if [ -e zerver/fixtures/migration-status ] &&
|
if [ -e zerver/fixtures/migration-status ] &&
|
||||||
cmp -s zerver/fixtures/available-migrations zerver/fixtures/migration-status &&
|
cmp -s zerver/fixtures/available-migrations zerver/fixtures/migration-status &&
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ VAGRANTUSERNAME=$(whoami)
|
|||||||
|
|
||||||
if [[ $# == 0 ]]; then
|
if [[ $# == 0 ]]; then
|
||||||
USERNAME=zulip
|
USERNAME=zulip
|
||||||
PASSWORD=$($(dirname $0)/../bin/get-django-setting LOCAL_DATABASE_PASSWORD)
|
PASSWORD=$("$(dirname "$0")/../bin/get-django-setting" LOCAL_DATABASE_PASSWORD)
|
||||||
DBNAME=zulip
|
DBNAME=zulip
|
||||||
SEARCH_PATH="$USERNAME",public
|
SEARCH_PATH="$USERNAME",public
|
||||||
elif [[ $# == 4 ]]; then
|
elif [[ $# == 4 ]]; then
|
||||||
@@ -28,7 +28,7 @@ fi
|
|||||||
|
|
||||||
DBNAME_BASE=${DBNAME}_base
|
DBNAME_BASE=${DBNAME}_base
|
||||||
|
|
||||||
$ROOT_POSTGRES $DEFAULT_DB << EOF
|
$ROOT_POSTGRES "$DEFAULT_DB" << EOF
|
||||||
CREATE USER $USERNAME;
|
CREATE USER $USERNAME;
|
||||||
ALTER USER $USERNAME PASSWORD '$PASSWORD';
|
ALTER USER $USERNAME PASSWORD '$PASSWORD';
|
||||||
ALTER USER $USERNAME CREATEDB;
|
ALTER USER $USERNAME CREATEDB;
|
||||||
@@ -42,28 +42,28 @@ EOF
|
|||||||
umask go-rw
|
umask go-rw
|
||||||
PGPASS_PREFIX="*:*:*:$USERNAME:"
|
PGPASS_PREFIX="*:*:*:$USERNAME:"
|
||||||
PGPASS_ESCAPED_PREFIX="*:\*:\*:$USERNAME:"
|
PGPASS_ESCAPED_PREFIX="*:\*:\*:$USERNAME:"
|
||||||
if ! $(grep -q "$PGPASS_ESCAPED_PREFIX" ~/.pgpass); then
|
if ! grep -q "$PGPASS_ESCAPED_PREFIX" ~/.pgpass; then
|
||||||
echo $PGPASS_PREFIX$PASSWORD >> ~/.pgpass
|
echo "$PGPASS_PREFIX$PASSWORD" >> ~/.pgpass
|
||||||
else
|
else
|
||||||
sed -i "s/$PGPASS_ESCAPED_PREFIX.*\$/$PGPASS_PREFIX$PASSWORD/" ~/.pgpass
|
sed -i "s/$PGPASS_ESCAPED_PREFIX.*\$/$PGPASS_PREFIX$PASSWORD/" ~/.pgpass
|
||||||
fi
|
fi
|
||||||
chmod go-rw ~/.pgpass
|
chmod go-rw ~/.pgpass
|
||||||
|
|
||||||
psql -h localhost postgres $USERNAME <<EOF
|
psql -h localhost postgres "$USERNAME" <<EOF
|
||||||
DROP DATABASE IF EXISTS $DBNAME;
|
DROP DATABASE IF EXISTS $DBNAME;
|
||||||
DROP DATABASE IF EXISTS $DBNAME_BASE;
|
DROP DATABASE IF EXISTS $DBNAME_BASE;
|
||||||
CREATE DATABASE $DBNAME_BASE
|
CREATE DATABASE $DBNAME_BASE
|
||||||
EOF
|
EOF
|
||||||
|
|
||||||
psql -h localhost $DBNAME_BASE $USERNAME <<EOF
|
psql -h localhost "$DBNAME_BASE" "$USERNAME" <<EOF
|
||||||
CREATE SCHEMA zulip;
|
CREATE SCHEMA zulip;
|
||||||
EOF
|
EOF
|
||||||
|
|
||||||
$ROOT_POSTGRES $DBNAME_BASE << EOF
|
$ROOT_POSTGRES "$DBNAME_BASE" << EOF
|
||||||
CREATE EXTENSION tsearch_extras SCHEMA zulip;
|
CREATE EXTENSION tsearch_extras SCHEMA zulip;
|
||||||
EOF
|
EOF
|
||||||
|
|
||||||
psql -h localhost postgres $USERNAME <<EOF
|
psql -h localhost postgres "$USERNAME" <<EOF
|
||||||
CREATE DATABASE $DBNAME TEMPLATE $DBNAME_BASE;
|
CREATE DATABASE $DBNAME TEMPLATE $DBNAME_BASE;
|
||||||
EOF
|
EOF
|
||||||
|
|
||||||
|
|||||||
@@ -1,3 +1,3 @@
|
|||||||
#!/bin/bash -xe
|
#!/bin/bash -xe
|
||||||
|
|
||||||
$(dirname $0)/postgres-init-db zulip_test $($(dirname $0)/../bin/get-django-setting LOCAL_DATABASE_PASSWORD) zulip_test zulip,public
|
"$(dirname "$0")/postgres-init-db" zulip_test "$("$(dirname "$0")/../bin/get-django-setting" LOCAL_DATABASE_PASSWORD)" zulip_test zulip,public
|
||||||
|
|||||||
@@ -6,11 +6,11 @@ export NODE_PATH=/usr/lib/nodejs:static
|
|||||||
|
|
||||||
INDEX_JS=zerver/tests/frontend/node/index.js
|
INDEX_JS=zerver/tests/frontend/node/index.js
|
||||||
NODEJS=$(which nodejs || which node)
|
NODEJS=$(which nodejs || which node)
|
||||||
if [ f$1 = fcover ]; then
|
if [ "$1" = "cover" ]; then
|
||||||
# Run a coverage test with Istanbul.
|
# Run a coverage test with Istanbul.
|
||||||
istanbul cover $INDEX_JS
|
istanbul cover "$INDEX_JS"
|
||||||
else
|
else
|
||||||
# Normal testing, no coverage analysis.
|
# Normal testing, no coverage analysis.
|
||||||
# Run the index.js test runner, which runs all the other tests.
|
# Run the index.js test runner, which runs all the other tests.
|
||||||
$NODEJS --stack-trace-limit=100 $INDEX_JS
|
"$NODEJS" --stack-trace-limit=100 "$INDEX_JS"
|
||||||
fi
|
fi
|
||||||
|
|||||||
@@ -693,7 +693,7 @@ def finish_google_oauth2(request):
|
|||||||
)
|
)
|
||||||
if resp.status_code != 200:
|
if resp.status_code != 200:
|
||||||
raise Exception('Could not convert google pauth2 code to access_token\r%r' % resp.text)
|
raise Exception('Could not convert google pauth2 code to access_token\r%r' % resp.text)
|
||||||
access_token = resp.json['access_token']
|
access_token = resp.json()['access_token']
|
||||||
|
|
||||||
resp = requests.get(
|
resp = requests.get(
|
||||||
'https://www.googleapis.com/plus/v1/people/me',
|
'https://www.googleapis.com/plus/v1/people/me',
|
||||||
@@ -701,7 +701,7 @@ def finish_google_oauth2(request):
|
|||||||
)
|
)
|
||||||
if resp.status_code != 200:
|
if resp.status_code != 200:
|
||||||
raise Exception('Google login failed making API call\r%r' % resp.text)
|
raise Exception('Google login failed making API call\r%r' % resp.text)
|
||||||
body = resp.json
|
body = resp.json()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
full_name = body['name']['formatted']
|
full_name = body['name']['formatted']
|
||||||
|
|||||||
@@ -91,7 +91,6 @@ class GoogleMobileOauth2Backend(ZulipAuthMixin):
|
|||||||
https://developers.google.com/+/mobile/android/sign-in#server-side_access_for_your_app
|
https://developers.google.com/+/mobile/android/sign-in#server-side_access_for_your_app
|
||||||
https://developers.google.com/accounts/docs/CrossClientAuth#offlineAccess
|
https://developers.google.com/accounts/docs/CrossClientAuth#offlineAccess
|
||||||
|
|
||||||
This backend is not currently supported on voyager.
|
|
||||||
"""
|
"""
|
||||||
def authenticate(self, google_oauth2_token=None, return_data={}):
|
def authenticate(self, google_oauth2_token=None, return_data={}):
|
||||||
try:
|
try:
|
||||||
|
|||||||
@@ -20,9 +20,17 @@ ADMIN_DOMAIN = 'example.com'
|
|||||||
AUTHENTICATION_BACKENDS = (
|
AUTHENTICATION_BACKENDS = (
|
||||||
# 'zproject.backends.EmailAuthBackend', # Email and password
|
# 'zproject.backends.EmailAuthBackend', # Email and password
|
||||||
# 'zproject.backends.ZulipRemoteUserBackend', # Local SSO
|
# 'zproject.backends.ZulipRemoteUserBackend', # Local SSO
|
||||||
# 'zproject.backends.GoogleBackend', # Google Apps
|
# 'zproject.backends.GoogleMobileOauth2Backend', # Google Apps, setup below
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# Google Oauth requires a bit of configuration; you will need to go to
|
||||||
|
# https://console.developers.google.com, setup an Oauth2 client ID
|
||||||
|
# that allows redirects to
|
||||||
|
# e.g. https://zulip.example.com/accounts/login/google/done/ put your
|
||||||
|
# client secret as "google_oauth2_client_secret" in
|
||||||
|
# zulip-secrets.conf, and your cleitn ID here:
|
||||||
|
# GOOGLE_OAUTH2_CLIENT_ID=<your client ID from Google>
|
||||||
|
|
||||||
# If you are using the ZulipRemoteUserBackend authentication backend,
|
# If you are using the ZulipRemoteUserBackend authentication backend,
|
||||||
# set this to your domain (e.g. if REMOTE_USER is "username" and the
|
# set this to your domain (e.g. if REMOTE_USER is "username" and the
|
||||||
# corresponding email address is "username@example.com", set
|
# corresponding email address is "username@example.com", set
|
||||||
@@ -34,9 +42,13 @@ SSO_APPEND_DOMAIN = None
|
|||||||
# outgoing port must be 587. The EMAIL_HOST is prepopulated
|
# outgoing port must be 587. The EMAIL_HOST is prepopulated
|
||||||
# for GMail servers, change it for other hosts, or leave it unset
|
# for GMail servers, change it for other hosts, or leave it unset
|
||||||
# or empty to skip sending email.
|
# or empty to skip sending email.
|
||||||
|
#
|
||||||
|
# A common problem you may encounter when trying to get this working
|
||||||
|
# is many hosting providers block outgoing SMTP traffic.
|
||||||
EMAIL_HOST = 'smtp.gmail.com'
|
EMAIL_HOST = 'smtp.gmail.com'
|
||||||
EMAIL_HOST_USER = ''
|
EMAIL_HOST_USER = ''
|
||||||
EMAIL_HOST_PASSWORD = ''
|
# If you're using password auth, you will need to put the password in
|
||||||
|
# /etc/zulip/zulip-secrets.conf as email_host_password.
|
||||||
EMAIL_PORT = 587
|
EMAIL_PORT = 587
|
||||||
EMAIL_USE_TLS = True
|
EMAIL_USE_TLS = True
|
||||||
|
|
||||||
|
|||||||
@@ -86,8 +86,7 @@ else:
|
|||||||
AUTHENTICATION_BACKENDS = ('zproject.backends.DevAuthBackend',)
|
AUTHENTICATION_BACKENDS = ('zproject.backends.DevAuthBackend',)
|
||||||
# Add some of the below if you're testing other backends
|
# Add some of the below if you're testing other backends
|
||||||
# AUTHENTICATION_BACKENDS = ('zproject.backends.EmailAuthBackend',
|
# AUTHENTICATION_BACKENDS = ('zproject.backends.EmailAuthBackend',
|
||||||
# 'zproject.backends.GoogleMobileOauth2Backend',
|
# 'zproject.backends.GoogleMobileOauth2Backend',)
|
||||||
# 'zproject.backends.GoogleBackend')
|
|
||||||
EXTERNAL_URI_SCHEME = "http://"
|
EXTERNAL_URI_SCHEME = "http://"
|
||||||
EMAIL_GATEWAY_PATTERN = "%s@" + EXTERNAL_HOST
|
EMAIL_GATEWAY_PATTERN = "%s@" + EXTERNAL_HOST
|
||||||
ADMIN_DOMAIN = "zulip.com"
|
ADMIN_DOMAIN = "zulip.com"
|
||||||
|
|||||||
Reference in New Issue
Block a user