mirror of
https://github.com/zulip/zulip.git
synced 2025-11-05 06:23:38 +00:00
Clients can now request to receive only certain kinds of events, although they always receive restart events. (imported from commit 1e72981f8fe763829ab2abde1e35f94cad5c34e4)
215 lines
7.0 KiB
Python
215 lines
7.0 KiB
Python
from django.conf import settings
|
|
from collections import deque
|
|
from tornado.ioloop import PeriodicCallback
|
|
import os
|
|
import time
|
|
import socket
|
|
import logging
|
|
import simplejson
|
|
import requests
|
|
import cPickle as pickle
|
|
import atexit
|
|
import sys
|
|
import signal
|
|
|
|
IDLE_EVENT_QUEUE_TIMEOUT_SECS = 60 * 10
|
|
|
|
class ClientDescriptor(object):
|
|
def __init__(self, user_profile_id, id, event_types=None, apply_markdown=True):
|
|
self.user_profile_id = user_profile_id
|
|
self.current_handler = None
|
|
self.event_queue = EventQueue(id)
|
|
self.event_types = event_types
|
|
self.last_connection_time = time.time()
|
|
self.apply_markdown = apply_markdown
|
|
|
|
def add_event(self, event):
|
|
if self.current_handler is not None:
|
|
self.current_handler._request._time_restarted = time.time()
|
|
|
|
self.event_queue.push(event)
|
|
self.check_connection()
|
|
if self.current_handler is not None:
|
|
try:
|
|
self.current_handler.humbug_finish(dict(result='success', msg='',
|
|
events=[event],
|
|
queue_id=self.event_queue.id),
|
|
self.current_handler._request,
|
|
apply_markdown=self.apply_markdown)
|
|
return
|
|
except socket.error:
|
|
pass
|
|
|
|
def accepts_event_type(self, type):
|
|
if self.event_types is None:
|
|
return True
|
|
return type in self.event_types
|
|
|
|
def idle(self, now):
|
|
self.check_connection()
|
|
return (self.current_handler is None
|
|
and now - self.last_connection_time >= IDLE_EVENT_QUEUE_TIMEOUT_SECS)
|
|
|
|
def connect_handler(self, handler):
|
|
self.current_handler = handler
|
|
self.last_connection_time = time.time()
|
|
|
|
def disconnect_handler(self):
|
|
self.current_handler = None
|
|
|
|
def check_connection(self):
|
|
if (self.current_handler is not None
|
|
and self.current_handler.request.connection.stream.closed()):
|
|
self.current_handler = None
|
|
|
|
class EventQueue(object):
|
|
def __init__(self, id):
|
|
self.queue = deque()
|
|
self.next_event_id = 0
|
|
self.id = id
|
|
|
|
def push(self, event):
|
|
event['id'] = self.next_event_id
|
|
self.next_event_id += 1
|
|
self.queue.append(event)
|
|
|
|
def pop(self):
|
|
return self.queue.popleft()
|
|
|
|
def empty(self):
|
|
return len(self.queue) == 0
|
|
|
|
def prune(self, through_id):
|
|
while not self.empty() and self.queue[0]['id'] <= through_id:
|
|
self.pop()
|
|
|
|
def contents(self):
|
|
return list(self.queue)
|
|
|
|
# maps queue ids to client descriptors
|
|
clients = {}
|
|
# maps user id to list of client descriptors
|
|
user_clients = {}
|
|
|
|
next_queue_id = 0
|
|
|
|
def allocate_client_descriptor(user_profile_id, event_types, apply_markdown):
|
|
global next_queue_id
|
|
id = str(settings.SERVER_GENERATION) + ':' + str(next_queue_id)
|
|
next_queue_id += 1
|
|
client = ClientDescriptor(user_profile_id, id, event_types, apply_markdown)
|
|
clients[id] = client
|
|
user_clients.setdefault(user_profile_id, []).append(client)
|
|
return client
|
|
|
|
EVENT_QUEUE_GC_FREQ_MSECS = 1000 * 60 * 5
|
|
|
|
def gc_event_queues():
|
|
start = time.time()
|
|
to_remove = set()
|
|
affected_users = set()
|
|
for (id, client) in clients.iteritems():
|
|
if client.idle(start):
|
|
to_remove.add(id)
|
|
affected_users.add(client.user_profile_id)
|
|
|
|
for id in to_remove:
|
|
del clients[id]
|
|
|
|
for user_id in affected_users:
|
|
new_client_list = filter(lambda c: c.event_queue.id not in to_remove,
|
|
user_clients[user_id])
|
|
user_clients[user_id] = new_client_list
|
|
|
|
logging.info(('Tornado removed %d idle event queues owned by %d users in %.3fs.'
|
|
+ ' Now %d active queues')
|
|
% (len(to_remove), len(affected_users), time.time() - start,
|
|
len(clients)))
|
|
|
|
PERSISTENT_QUEUE_FILENAME = os.path.join(os.path.dirname(__file__),
|
|
"..", "event_queues.pickle")
|
|
|
|
def dump_event_queues():
|
|
start = time.time()
|
|
# Remove unpickle-able attributes
|
|
for client in clients.itervalues():
|
|
client.current_handler = None
|
|
|
|
with file(PERSISTENT_QUEUE_FILENAME, "w") as stored_queues:
|
|
pickle.dump(clients, stored_queues)
|
|
|
|
logging.info('Tornado dumped %d event queues in %.3fs'
|
|
% (len(clients), time.time() - start))
|
|
|
|
def load_event_queues():
|
|
global clients
|
|
start = time.time()
|
|
try:
|
|
with file(PERSISTENT_QUEUE_FILENAME, "r") as stored_queues:
|
|
clients = pickle.load(stored_queues)
|
|
except (IOError, EOFError):
|
|
pass
|
|
|
|
for client in clients.itervalues():
|
|
user_clients.setdefault(client.user_profile_id, []).append(client)
|
|
|
|
logging.info('Tornado loaded %d event queues in %.3fs'
|
|
% (len(clients), time.time() - start))
|
|
|
|
def send_restart_events():
|
|
event = dict(type='restart', server_generation=settings.SERVER_GENERATION)
|
|
for client in clients.itervalues():
|
|
# All clients get restart events
|
|
client.add_event(event.copy())
|
|
|
|
def setup_event_queue(io_loop):
|
|
load_event_queues()
|
|
atexit.register(dump_event_queues)
|
|
# Make sure we dump event queues even if we exit via signal
|
|
signal.signal(signal.SIGTERM, lambda signum, stack: sys.exit(1))
|
|
|
|
try:
|
|
os.remove(PERSISTENT_QUEUE_FILENAME)
|
|
except OSError:
|
|
pass
|
|
|
|
# Set up event queue garbage collection
|
|
pc = PeriodicCallback(gc_event_queues, EVENT_QUEUE_GC_FREQ_MSECS, io_loop)
|
|
pc.start()
|
|
|
|
send_restart_events()
|
|
|
|
# The following functions are called from Django
|
|
|
|
def request_event_queue(user_profile, apply_markdown, event_types=None):
|
|
if settings.TORNADO_SERVER:
|
|
req = {'dont_block' : 'true',
|
|
'apply_markdown': simplejson.dumps(apply_markdown),
|
|
'client' : 'internal'}
|
|
if event_types is not None:
|
|
req['event_types'] = simplejson.dumps(event_types)
|
|
resp = requests.get(settings.TORNADO_SERVER + '/api/v1/events',
|
|
auth=requests.auth.HTTPBasicAuth(user_profile.user.email,
|
|
user_profile.api_key),
|
|
params=req)
|
|
|
|
resp.raise_for_status()
|
|
|
|
return resp.json['queue_id']
|
|
|
|
return None
|
|
|
|
def get_user_events(user_profile, queue_id, last_event_id):
|
|
if settings.TORNADO_SERVER:
|
|
resp = requests.get(settings.TORNADO_SERVER + '/api/v1/events',
|
|
auth=requests.auth.HTTPBasicAuth(user_profile.user.email,
|
|
user_profile.api_key),
|
|
params={'queue_id' : queue_id,
|
|
'last_event_id': last_event_id,
|
|
'dont_block' : 'true',
|
|
'client' : 'internal'})
|
|
|
|
resp.raise_for_status()
|
|
|
|
return resp.json['events']
|