Added a multi-worker approach to the Tornado WSGI container.

The WSGI container is a good option to wrap a multi-modal webapp
(Flask + websocket routes), but it's constrained to a single-process
approach and queued/pre-buffered requests. That makes performance poor
when handling requests that may take a few seconds to complete.
This commit is contained in:
Fabio Manganiello 2023-05-13 01:26:18 +02:00
parent b7b93edbae
commit 71401a4936
Signed by: blacklight
GPG key ID: D90FBA7F76362774

View file

@ -1,10 +1,12 @@
import multiprocessing
import os import os
import pathlib import pathlib
import secrets import secrets
import threading import threading
import time
from multiprocessing import Process from multiprocessing import Process
from typing import Mapping, Optional from typing import List, Mapping, Optional
from tornado.wsgi import WSGIContainer from tornado.wsgi import WSGIContainer
from tornado.web import Application, FallbackHandler from tornado.web import Application, FallbackHandler
@ -186,6 +188,7 @@ class HttpBackend(Backend):
""" """
_DEFAULT_HTTP_PORT = 8008 _DEFAULT_HTTP_PORT = 8008
"""The default listen port for the webserver."""
def __init__( def __init__(
self, self,
@ -193,6 +196,7 @@ class HttpBackend(Backend):
bind_address: str = '0.0.0.0', bind_address: str = '0.0.0.0',
resource_dirs: Optional[Mapping[str, str]] = None, resource_dirs: Optional[Mapping[str, str]] = None,
secret_key_file: Optional[str] = None, secret_key_file: Optional[str] = None,
num_workers: Optional[int] = None,
**kwargs, **kwargs,
): ):
""" """
@ -204,12 +208,13 @@ class HttpBackend(Backend):
the value is the absolute path to expose. the value is the absolute path to expose.
:param secret_key_file: Path to the file containing the secret key that will be used by Flask :param secret_key_file: Path to the file containing the secret key that will be used by Flask
(default: ``~/.local/share/platypush/flask.secret.key``). (default: ``~/.local/share/platypush/flask.secret.key``).
:param num_workers: Number of worker processes to use (default: ``(cpu_count * 2) + 1``).
""" """
super().__init__(**kwargs) super().__init__(**kwargs)
self.port = port self.port = port
self.server_proc = None self._workers: List[Process] = []
self._service_registry_thread = None self._service_registry_thread = None
self.bind_address = bind_address self.bind_address = bind_address
self._io_loop: Optional[IOLoop] = None self._io_loop: Optional[IOLoop] = None
@ -227,6 +232,7 @@ class HttpBackend(Backend):
or os.path.join(Config.get('workdir'), 'flask.secret.key') # type: ignore or os.path.join(Config.get('workdir'), 'flask.secret.key') # type: ignore
) )
self.local_base_url = f'http://localhost:{self.port}' self.local_base_url = f'http://localhost:{self.port}'
self.num_workers = num_workers or (multiprocessing.cpu_count() * 2) + 1
def send_message(self, *_, **__): def send_message(self, *_, **__):
self.logger.warning('Use cURL or any HTTP client to query the HTTP backend') self.logger.warning('Use cURL or any HTTP client to query the HTTP backend')
@ -240,18 +246,27 @@ class HttpBackend(Backend):
self._io_loop.stop() self._io_loop.stop()
self._io_loop.close() self._io_loop.close()
if self.server_proc: workers = self._workers.copy()
self.server_proc.terminate() for worker in workers:
self.server_proc.join(timeout=10) worker.terminate()
if self.server_proc.is_alive():
self.server_proc.kill() stop_timeout = 5
if self.server_proc.is_alive(): wait_start_time = time.time()
wait_max_time = wait_start_time + stop_timeout
for i, worker in enumerate(workers[::-1]):
worker.join(timeout=max(0, wait_max_time - time.time()))
if worker.is_alive():
worker.kill()
if worker.is_alive():
self.logger.info( self.logger.info(
'HTTP server process may be still alive at termination' 'HTTP worker %s may be still alive at termination', worker.name
) )
else: else:
self.logger.info('HTTP server process terminated') self._workers.pop(i)
self.logger.info('HTTP workers terminated')
if self._service_registry_thread and self._service_registry_thread.is_alive(): if self._service_registry_thread and self._service_registry_thread.is_alive():
self._service_registry_thread.join(timeout=5) self._service_registry_thread.join(timeout=5)
self._service_registry_thread = None self._service_registry_thread = None
@ -282,15 +297,12 @@ class HttpBackend(Backend):
raise e raise e
def _web_server_proc(self): def _web_server_proc(self):
def proc():
self.logger.info('Starting local web server on port %s', self.port)
assert isinstance( assert isinstance(
self.bus, RedisBus self.bus, RedisBus
), 'The HTTP backend only works if backed by a Redis bus' ), 'The HTTP backend only works if backed by a Redis bus'
application.config['redis_queue'] = self.bus.redis_queue application.config['redis_queue'] = self.bus.redis_queue
application.secret_key = self._get_secret_key() application.secret_key = self._get_secret_key()
container = WSGIContainer(application) container = WSGIContainer(application)
server = Application( server = Application(
[ [
@ -299,7 +311,7 @@ class HttpBackend(Backend):
] ]
) )
server.listen(address=self.bind_address, port=self.port) server.listen(address=self.bind_address, port=self.port, reuse_port=True)
self._io_loop = IOLoop.instance() self._io_loop = IOLoop.instance()
try: try:
@ -308,8 +320,6 @@ class HttpBackend(Backend):
if not self.should_stop(): if not self.should_stop():
raise e raise e
return proc
def _register_service(self): def _register_service(self):
try: try:
self.register_service(port=self.port) self.register_service(port=self.port)
@ -325,9 +335,22 @@ class HttpBackend(Backend):
self._service_registry_thread.start() self._service_registry_thread.start()
def _run_web_server(self): def _run_web_server(self):
self.server_proc = Process(target=self._web_server_proc(), name='WebServer') self.logger.info(
self.server_proc.start() 'Starting local web server on port %s with %d service workers',
self.server_proc.join() self.port,
self.num_workers,
)
workers = [
Process(target=self._web_server_proc, name=f'WebWorker#{i + 1}')
for i in range(self.num_workers)
]
for worker in workers:
worker.start()
for worker in workers:
worker.join()
def run(self): def run(self):
super().run() super().run()