mirror of
https://github.com/django/daphne.git
synced 2025-07-04 03:43:25 +03:00
Update runserver to autoreload w/daphne in main thread
This commit is contained in:
parent
5ddeed4a25
commit
aff9ca2f13
|
@ -52,6 +52,9 @@ class ChannelLayerManager(object):
|
||||||
self.backends[key] = self.make_backend(key)
|
self.backends[key] = self.make_backend(key)
|
||||||
return self.backends[key]
|
return self.backends[key]
|
||||||
|
|
||||||
|
def __contains__(self, key):
|
||||||
|
return key in self.configs
|
||||||
|
|
||||||
|
|
||||||
class ChannelLayerWrapper(object):
|
class ChannelLayerWrapper(object):
|
||||||
"""
|
"""
|
||||||
|
|
|
@ -20,6 +20,11 @@ class AsgiRequest(http.HttpRequest):
|
||||||
dict, and wraps request body handling.
|
dict, and wraps request body handling.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
# Exception that will cause any handler to skip around response
|
||||||
|
# transmission and presume something else will do it later.
|
||||||
|
class ResponseLater(Exception):
|
||||||
|
pass
|
||||||
|
|
||||||
def __init__(self, message):
|
def __init__(self, message):
|
||||||
self.message = message
|
self.message = message
|
||||||
self.reply_channel = self.message.reply_channel
|
self.reply_channel = self.message.reply_channel
|
||||||
|
@ -27,7 +32,8 @@ class AsgiRequest(http.HttpRequest):
|
||||||
self._post_parse_error = False
|
self._post_parse_error = False
|
||||||
self.resolver_match = None
|
self.resolver_match = None
|
||||||
# Path info
|
# Path info
|
||||||
self.path = self.message['path']
|
# TODO: probably needs actual URL decoding
|
||||||
|
self.path = self.message['path'].decode("ascii")
|
||||||
self.script_name = self.message.get('root_path', '')
|
self.script_name = self.message.get('root_path', '')
|
||||||
if self.script_name:
|
if self.script_name:
|
||||||
# TODO: Better is-prefix checking, slash handling?
|
# TODO: Better is-prefix checking, slash handling?
|
||||||
|
@ -38,7 +44,7 @@ class AsgiRequest(http.HttpRequest):
|
||||||
self.method = self.message['method'].upper()
|
self.method = self.message['method'].upper()
|
||||||
self.META = {
|
self.META = {
|
||||||
"REQUEST_METHOD": self.method,
|
"REQUEST_METHOD": self.method,
|
||||||
"QUERY_STRING": self.message.get('query_string', ''),
|
"QUERY_STRING": self.message.get('query_string', '').decode("ascii"),
|
||||||
# Old code will need these for a while
|
# Old code will need these for a while
|
||||||
"wsgi.multithread": True,
|
"wsgi.multithread": True,
|
||||||
"wsgi.multiprocess": True,
|
"wsgi.multiprocess": True,
|
||||||
|
@ -151,6 +157,10 @@ class AsgiHandler(base.BaseHandler):
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
response = http.HttpResponseBadRequest()
|
response = http.HttpResponseBadRequest()
|
||||||
|
except AsgiRequest.ResponseLater:
|
||||||
|
# The view has promised something else
|
||||||
|
# will send a response at a later time
|
||||||
|
return
|
||||||
else:
|
else:
|
||||||
response = self.get_response(request)
|
response = self.get_response(request)
|
||||||
# Transform response into messages, which we yield back to caller
|
# Transform response into messages, which we yield back to caller
|
||||||
|
@ -158,9 +168,10 @@ class AsgiHandler(base.BaseHandler):
|
||||||
# TODO: file_to_stream
|
# TODO: file_to_stream
|
||||||
yield message
|
yield message
|
||||||
|
|
||||||
def encode_response(self, response):
|
@classmethod
|
||||||
|
def encode_response(cls, response):
|
||||||
"""
|
"""
|
||||||
Encodes a Django HTTP response into an ASGI http.response message(s).
|
Encodes a Django HTTP response into ASGI http.response message(s).
|
||||||
"""
|
"""
|
||||||
# Collect cookies into headers.
|
# Collect cookies into headers.
|
||||||
# Note that we have to preserve header case as there are some non-RFC
|
# Note that we have to preserve header case as there are some non-RFC
|
||||||
|
@ -181,19 +192,19 @@ class AsgiHandler(base.BaseHandler):
|
||||||
response_headers.append(
|
response_headers.append(
|
||||||
(
|
(
|
||||||
'Set-Cookie',
|
'Set-Cookie',
|
||||||
six.binary_type(c.output(header='')),
|
c.output(header='').encode("ascii"),
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
# Make initial response message
|
# Make initial response message
|
||||||
message = {
|
message = {
|
||||||
"status": response.status_code,
|
"status": response.status_code,
|
||||||
"status_text": response.reason_phrase,
|
"status_text": response.reason_phrase.encode("ascii"),
|
||||||
"headers": response_headers,
|
"headers": response_headers,
|
||||||
}
|
}
|
||||||
# Streaming responses need to be pinned to their iterator
|
# Streaming responses need to be pinned to their iterator
|
||||||
if response.streaming:
|
if response.streaming:
|
||||||
for part in response.streaming_content:
|
for part in response.streaming_content:
|
||||||
for chunk in self.chunk_bytes(part):
|
for chunk in cls.chunk_bytes(part):
|
||||||
message['content'] = chunk
|
message['content'] = chunk
|
||||||
message['more_content'] = True
|
message['more_content'] = True
|
||||||
yield message
|
yield message
|
||||||
|
@ -205,13 +216,14 @@ class AsgiHandler(base.BaseHandler):
|
||||||
# Other responses just need chunking
|
# Other responses just need chunking
|
||||||
else:
|
else:
|
||||||
# Yield chunks of response
|
# Yield chunks of response
|
||||||
for chunk, last in self.chunk_bytes(response.content):
|
for chunk, last in cls.chunk_bytes(response.content):
|
||||||
message['content'] = chunk
|
message['content'] = chunk
|
||||||
message['more_content'] = not last
|
message['more_content'] = not last
|
||||||
yield message
|
yield message
|
||||||
message = {}
|
message = {}
|
||||||
|
|
||||||
def chunk_bytes(self, data):
|
@classmethod
|
||||||
|
def chunk_bytes(cls, data):
|
||||||
"""
|
"""
|
||||||
Chunks some data into chunks based on the current ASGI channel layer's
|
Chunks some data into chunks based on the current ASGI channel layer's
|
||||||
message size and reasonable defaults.
|
message size and reasonable defaults.
|
||||||
|
@ -221,10 +233,10 @@ class AsgiHandler(base.BaseHandler):
|
||||||
position = 0
|
position = 0
|
||||||
while position < len(data):
|
while position < len(data):
|
||||||
yield (
|
yield (
|
||||||
data[position:position + self.chunk_size],
|
data[position:position + cls.chunk_size],
|
||||||
(position + self.chunk_size) >= len(data),
|
(position + cls.chunk_size) >= len(data),
|
||||||
)
|
)
|
||||||
position += self.chunk_size
|
position += cls.chunk_size
|
||||||
|
|
||||||
|
|
||||||
class ViewConsumer(object):
|
class ViewConsumer(object):
|
||||||
|
@ -232,8 +244,10 @@ class ViewConsumer(object):
|
||||||
Dispatches channel HTTP requests into django's URL/View system.
|
Dispatches channel HTTP requests into django's URL/View system.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
handler_class = AsgiHandler
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
self.handler = AsgiHandler()
|
self.handler = self.handler_class()
|
||||||
|
|
||||||
def __call__(self, message):
|
def __call__(self, message):
|
||||||
for reply_message in self.handler(message):
|
for reply_message in self.handler(message):
|
||||||
|
|
|
@ -2,15 +2,28 @@ import logging
|
||||||
|
|
||||||
|
|
||||||
def setup_logger(name, verbosity=1):
|
def setup_logger(name, verbosity=1):
|
||||||
|
"""
|
||||||
|
Basic logger for runserver etc.
|
||||||
|
"""
|
||||||
|
|
||||||
formatter = logging.Formatter(fmt='%(asctime)s - %(levelname)s - %(module)s - %(message)s')
|
formatter = logging.Formatter(fmt='%(asctime)s - %(levelname)s - %(module)s - %(message)s')
|
||||||
|
|
||||||
handler = logging.StreamHandler()
|
handler = logging.StreamHandler()
|
||||||
handler.setFormatter(formatter)
|
handler.setFormatter(formatter)
|
||||||
|
|
||||||
|
# Set up main logger
|
||||||
logger = logging.getLogger(name)
|
logger = logging.getLogger(name)
|
||||||
logger.setLevel(logging.INFO)
|
logger.setLevel(logging.INFO)
|
||||||
logger.addHandler(handler)
|
logger.addHandler(handler)
|
||||||
if verbosity > 1:
|
if verbosity > 1:
|
||||||
logger.setLevel(logging.DEBUG)
|
logger.setLevel(logging.DEBUG)
|
||||||
|
logger.debug("Logging set to DEBUG")
|
||||||
|
|
||||||
|
# Set up daphne protocol loggers
|
||||||
|
for module in ["daphne.ws_protocol", "daphne.http_protocol"]:
|
||||||
|
daphne_logger = logging.getLogger()
|
||||||
|
daphne_logger.addHandler(handler)
|
||||||
|
daphne_logger.setLevel(logging.DEBUG if verbosity > 1 else logging.INFO)
|
||||||
|
|
||||||
logger.propagate = False
|
logger.propagate = False
|
||||||
return logger
|
return logger
|
||||||
|
|
|
@ -1,5 +1,7 @@
|
||||||
|
import os
|
||||||
import threading
|
import threading
|
||||||
|
|
||||||
|
from django.utils import autoreload
|
||||||
from django.core.management.commands.runserver import \
|
from django.core.management.commands.runserver import \
|
||||||
Command as RunserverCommand
|
Command as RunserverCommand
|
||||||
|
|
||||||
|
@ -17,21 +19,18 @@ class Command(RunserverCommand):
|
||||||
super(Command, self).handle(*args, **options)
|
super(Command, self).handle(*args, **options)
|
||||||
|
|
||||||
def run(self, *args, **options):
|
def run(self, *args, **options):
|
||||||
# Don't autoreload for now
|
|
||||||
self.inner_run(None, **options)
|
|
||||||
|
|
||||||
def inner_run(self, *args, **options):
|
|
||||||
# Check a handler is registered for http reqs; if not, add default one
|
# Check a handler is registered for http reqs; if not, add default one
|
||||||
self.channel_layer = channel_layers[DEFAULT_CHANNEL_LAYER]
|
self.channel_layer = channel_layers[DEFAULT_CHANNEL_LAYER]
|
||||||
if not self.channel_layer.registry.consumer_for_channel("http.request"):
|
if not self.channel_layer.registry.consumer_for_channel("http.request"):
|
||||||
self.channel_layer.registry.add_consumer(ViewConsumer(), ["http.request"])
|
self.channel_layer.registry.add_consumer(ViewConsumer(), ["http.request"])
|
||||||
# Note that this is the channel-enabled one on the console
|
# Helpful note to say this is the Channels runserver
|
||||||
self.logger.info("Worker thread running, channels enabled")
|
self.logger.info("Worker thread running, channels enabled")
|
||||||
# Launch a worker thread
|
# Launch worker as subthread (including autoreload logic)
|
||||||
worker = WorkerThread(self.channel_layer)
|
worker = WorkerThread(self.channel_layer)
|
||||||
worker.daemon = True
|
worker.daemon = True
|
||||||
worker.start()
|
worker.start()
|
||||||
# Launch server in main thread
|
# Launch server in main thread (Twisted doesn't like being in a
|
||||||
|
# subthread, and it doesn't need to autoreload as there's no user code)
|
||||||
from daphne.server import Server
|
from daphne.server import Server
|
||||||
Server(
|
Server(
|
||||||
channel_layer=self.channel_layer,
|
channel_layer=self.channel_layer,
|
||||||
|
@ -50,4 +49,8 @@ class WorkerThread(threading.Thread):
|
||||||
self.channel_layer = channel_layer
|
self.channel_layer = channel_layer
|
||||||
|
|
||||||
def run(self):
|
def run(self):
|
||||||
Worker(channel_layer=self.channel_layer).run()
|
worker = Worker(channel_layer=self.channel_layer)
|
||||||
|
# We need to set run_main so it doesn't try to relaunch the entire
|
||||||
|
# program - that will make Daphne run twice.
|
||||||
|
os.environ["RUN_MAIN"] = "true"
|
||||||
|
autoreload.main(worker.run)
|
||||||
|
|
Loading…
Reference in New Issue
Block a user