Dirty Arbiters¶
Dirty Arbiters provide a separate process pool for executing long-running, blocking operations (AI model loading, heavy computation) without blocking HTTP workers. This feature is inspired by Erlang's dirty schedulers.
Overview¶
Traditional Gunicorn workers are designed to handle HTTP requests quickly. Long-running operations like loading ML models or performing heavy computation can block these workers, reducing the server's ability to handle concurrent requests.
Dirty Arbiters solve this by providing:
- Separate worker pool - Completely separate from HTTP workers, can be killed/restarted independently
- Stateful workers - Loaded resources persist in dirty worker memory
- Message-passing IPC - Communication via Unix sockets with JSON serialization
- Explicit API - Clear
execute()calls (no hidden IPC) - Asyncio-based - Clean concurrent handling with streaming support
Design Philosophy¶
Dirty Arbiters follow several key design principles:
Separate Process Hierarchy¶
Unlike threads or in-process pools, Dirty Arbiters use a fully separate process tree:
- Isolation - A crash or memory leak in a dirty worker cannot affect HTTP workers
- Independent lifecycle - Dirty workers can be killed/restarted without affecting request handling
- Resource accounting - OS-level memory limits can be applied per-process
- Clean shutdown - Each process tree can be signaled and terminated independently
Erlang Inspiration¶
The name and concept come from Erlang's "dirty schedulers" - special schedulers that handle operations that would block normal schedulers. In Erlang, dirty schedulers run NIFs (Native Implemented Functions) that can't yield. Similarly, Gunicorn's Dirty Arbiters handle Python operations that would block HTTP workers.
Why Asyncio¶
The Dirty Arbiter uses asyncio for its core loop rather than the main arbiter's select-based approach:
- Non-blocking IPC - Can handle many concurrent client connections efficiently
- Concurrent request routing - Multiple requests can be dispatched to workers simultaneously
- Streaming support - Native async generators for streaming responses
- Clean signal handling - Signals integrate cleanly via
loop.add_signal_handler()
Stateful Applications¶
Traditional WSGI apps are request-scoped - they're invoked per-request and don't maintain state between requests. Dirty apps are different:
- Long-lived - Apps persist in worker memory for the worker's lifetime
- Pre-loaded resources - Models, connections, and caches stay loaded
- Explicit state management - Apps control their own lifecycle via
init()andclose()
This makes dirty apps ideal for ML inference, where loading a model once and reusing it for many requests is essential.
Architecture¶
+-------------------+
| Main Arbiter |
| (manages both) |
+--------+----------+
|
SIGTERM/SIGHUP/SIGUSR1 (forwarded)
|
+----------------------+----------------------+
| |
+-----v-----+ +------v------+
| HTTP | | Dirty |
| Workers | | Arbiter |
+-----------+ +------+------+
| |
| Unix Socket IPC SIGTERM/SIGHUP
| /tmp/gunicorn_dirty_<pid>.sock |
+------------------>---------------------->---+
+-----------+-----------+
| | |
+-----v---+ +-----v---+ +-----v---+
| Dirty | | Dirty | | Dirty |
| Worker | | Worker | | Worker |
+---------+ +---------+ +---------+
^ | ^ | ^ |
| | | | | |
Heartbeat (mtime every dirty_timeout/2)
| | | | | |
+---+--------+---+-------+---+
|
All workers load all dirty apps
[MLApp, ImageApp, ...]
Process Relationships¶
| Component | Parent | Communication |
|---|---|---|
| Main Arbiter | init/systemd | Signals from OS |
| HTTP Workers | Main Arbiter | Pipes, signals |
| Dirty Arbiter | Main Arbiter | Signals, exit status |
| Dirty Workers | Dirty Arbiter | Unix socket, signals, WorkerTmp |
Configuration¶
Add these settings to your Gunicorn configuration file or command line:
# gunicorn.conf.py
dirty_apps = [
"myapp.ml:MLApp",
"myapp.images:ImageApp",
]
dirty_workers = 2 # Number of dirty workers
dirty_timeout = 300 # Task timeout in seconds
dirty_threads = 1 # Threads per worker
dirty_graceful_timeout = 30 # Shutdown timeout
Or via command line:
gunicorn myapp:app \
--dirty-app myapp.ml:MLApp \
--dirty-app myapp.images:ImageApp \
--dirty-workers 2 \
--dirty-timeout 300
Configuration Options¶
| Setting | Default | Description |
|---|---|---|
dirty_apps |
[] |
List of dirty app import paths |
dirty_workers |
0 |
Number of dirty workers (0 = disabled) |
dirty_timeout |
300 |
Task timeout in seconds |
dirty_threads |
1 |
Threads per dirty worker |
dirty_graceful_timeout |
30 |
Graceful shutdown timeout |
Creating a Dirty App¶
Dirty apps inherit from DirtyApp and implement three methods:
# myapp/dirty.py
from gunicorn.dirty import DirtyApp
class MLApp(DirtyApp):
"""Dirty application for ML workloads."""
def __init__(self):
self.models = {}
def init(self):
"""Called once at dirty worker startup."""
# Pre-load commonly used models
self.models['default'] = self._load_model('base-model')
def __call__(self, action, *args, **kwargs):
"""Dispatch to action methods."""
method = getattr(self, action, None)
if method is None:
raise ValueError(f"Unknown action: {action}")
return method(*args, **kwargs)
def load_model(self, name):
"""Load a model into memory."""
if name not in self.models:
self.models[name] = self._load_model(name)
return {"loaded": True, "name": name}
def inference(self, model_name, input_text):
"""Run inference on loaded model."""
model = self.models.get(model_name)
if not model:
raise ValueError(f"Model not loaded: {model_name}")
return model.predict(input_text)
def _load_model(self, name):
import torch
model = torch.load(f"models/{name}.pt")
return model
def close(self):
"""Cleanup on shutdown."""
for model in self.models.values():
del model
DirtyApp Interface¶
| Method | Description |
|---|---|
init() |
Called once when dirty worker starts, after instantiation. Load resources here. |
__call__(action, *args, **kwargs) |
Handle requests from HTTP workers. |
close() |
Called when dirty worker shuts down. Cleanup resources. |
Initialization Sequence¶
When a dirty worker starts, initialization happens in this order:
- Fork - Worker process is forked from dirty arbiter
dirty_post_fork(arbiter, worker)- Hook called immediately after fork- App instantiation - Each dirty app class is instantiated (
__init__) app.init()- Called for each app after instantiation (load models, resources)dirty_worker_init(worker)- Hook called after ALL apps are initialized- Run loop - Worker starts accepting requests from HTTP workers
This means:
- Use
__init__for basic setup (initialize empty containers, store config) - Use
init()for heavy loading (ML models, database connections, large files) - The
dirty_worker_inithook fires only after all apps have completed theirinit()calls
Using from HTTP Workers¶
Sync Workers (sync, gthread)¶
from gunicorn.dirty import get_dirty_client
def my_view(request):
client = get_dirty_client()
# Load a model
client.execute("myapp.ml:MLApp", "load_model", "gpt-4")
# Run inference
result = client.execute(
"myapp.ml:MLApp",
"inference",
"gpt-4",
input_text=request.data
)
return result
Async Workers (ASGI)¶
from gunicorn.dirty import get_dirty_client_async
async def my_view(request):
client = await get_dirty_client_async()
# Non-blocking execution
await client.execute_async("myapp.ml:MLApp", "load_model", "gpt-4")
result = await client.execute_async(
"myapp.ml:MLApp",
"inference",
"gpt-4",
input_text=request.data
)
return result
Streaming¶
Dirty Arbiters support streaming responses for use cases like LLM token generation, where data is produced incrementally. This enables real-time delivery of results without waiting for complete execution.
Streaming with Generators¶
Any dirty app action that returns a generator (sync or async) automatically streams chunks to the client:
# myapp/llm.py
from gunicorn.dirty import DirtyApp
class LLMApp(DirtyApp):
def init(self):
from transformers import pipeline
self.generator = pipeline("text-generation", model="gpt2")
def generate(self, prompt):
"""Sync streaming - yields tokens."""
for token in self.generator(prompt, stream=True):
yield token["generated_text"]
async def generate_async(self, prompt):
"""Async streaming - yields tokens."""
import openai
client = openai.AsyncOpenAI()
stream = await client.chat.completions.create(
model="gpt-4",
messages=[{"role": "user", "content": prompt}],
stream=True
)
async for chunk in stream:
if chunk.choices[0].delta.content:
yield chunk.choices[0].delta.content
def close(self):
pass
Client Streaming API¶
Use stream() for sync workers and stream_async() for async workers:
Sync Workers (sync, gthread):
from gunicorn.dirty import get_dirty_client
def generate_view(request):
client = get_dirty_client()
def generate_response():
for chunk in client.stream("myapp.llm:LLMApp", "generate", request.prompt):
yield chunk
return StreamingResponse(generate_response())
Async Workers (ASGI):
from gunicorn.dirty import get_dirty_client_async
async def generate_view(request):
client = await get_dirty_client_async()
async def generate_response():
async for chunk in client.stream_async("myapp.llm:LLMApp", "generate", request.prompt):
yield chunk
return StreamingResponse(generate_response())
Streaming Protocol¶
Streaming uses a simple protocol with three message types:
- Chunk (
type: "chunk") - Contains partial data - End (
type: "end") - Signals stream completion - Error (
type: "error") - Signals error during streaming
Example message flow:
Client -> Arbiter -> Worker: request
Worker -> Arbiter -> Client: chunk (data: "Hello")
Worker -> Arbiter -> Client: chunk (data: " ")
Worker -> Arbiter -> Client: chunk (data: "World")
Worker -> Arbiter -> Client: end
Error Handling in Streams¶
Errors during streaming are delivered as error messages:
def generate_view(request):
client = get_dirty_client()
try:
for chunk in client.stream("myapp.llm:LLMApp", "generate", prompt):
yield chunk
except DirtyError as e:
# Error occurred mid-stream
yield f"\n[Error: {e.message}]"
Best Practices for Streaming¶
- Use async generators for I/O-bound streaming - e.g., API calls to external services
- Use sync generators for CPU-bound streaming - e.g., local model inference
- Yield frequently - Heartbeats are sent during streaming to keep workers alive
- Keep chunks small - Smaller chunks provide better perceived latency
- Handle client disconnection - Streams continue even if client disconnects; design accordingly
Flask Example¶
from flask import Flask, Response
from gunicorn.dirty import get_dirty_client
app = Flask(__name__)
@app.route("/chat", methods=["POST"])
def chat():
prompt = request.json.get("prompt")
client = get_dirty_client()
def stream():
for token in client.stream("myapp.llm:LLMApp", "generate", prompt):
yield f"data: {token}\n\n"
return Response(stream(), content_type="text/event-stream")
FastAPI Example¶
from fastapi import FastAPI
from fastapi.responses import StreamingResponse
from gunicorn.dirty import get_dirty_client_async
app = FastAPI()
@app.post("/chat")
async def chat(prompt: str):
client = await get_dirty_client_async()
async def stream():
async for token in client.stream_async("myapp.llm:LLMApp", "generate", prompt):
yield f"data: {token}\n\n"
return StreamingResponse(stream(), media_type="text/event-stream")
Lifecycle Hooks¶
Dirty Arbiters provide hooks for customization:
# gunicorn.conf.py
def on_dirty_starting(arbiter):
"""Called just before the dirty arbiter starts."""
print("Dirty arbiter starting...")
def dirty_post_fork(arbiter, worker):
"""Called just after a dirty worker is forked."""
print(f"Dirty worker {worker.pid} forked")
def dirty_worker_init(worker):
"""Called after a dirty worker initializes all apps."""
print(f"Dirty worker {worker.pid} initialized")
def dirty_worker_exit(arbiter, worker):
"""Called when a dirty worker exits."""
print(f"Dirty worker {worker.pid} exiting")
on_dirty_starting = on_dirty_starting
dirty_post_fork = dirty_post_fork
dirty_worker_init = dirty_worker_init
dirty_worker_exit = dirty_worker_exit
Signal Handling¶
Dirty Arbiters integrate with the main arbiter's signal handling. Signals are forwarded from the main arbiter to the dirty arbiter, which then propagates them to workers.
Signal Flow¶
Main Arbiter Dirty Arbiter Dirty Workers
| | |
SIGTERM/SIGHUP/SIGUSR1 ------> signal_handler() |
| | |
| call_soon_threadsafe() |
| | |
| handle_signal() |
| | |
| +------> os.kill(worker, sig) |
| |
Signal Reference¶
| Signal | At Dirty Arbiter | At Dirty Workers | Notes |
|---|---|---|---|
SIGTERM |
Sets self.alive = False, waits for graceful shutdown |
Exits after completing current request | Graceful shutdown with timeout |
SIGQUIT |
Immediate exit via sys.exit(0) |
Killed immediately | Fast shutdown, no cleanup |
SIGHUP |
Kills all workers, spawns new ones | Exits immediately | Hot reload of workers |
SIGUSR1 |
Reopens log files, forwards to workers | Reopens log files | Log rotation support |
SIGCHLD |
Handled by event loop, triggers reap | N/A | Worker death detection |
SIGINT |
Same as SIGTERM | Same as SIGTERM | Ctrl-C handling |
Forwarded Signals¶
The main arbiter forwards these signals to the dirty arbiter process:
- SIGTERM - Graceful shutdown of entire process tree
- SIGHUP - Worker reload (main arbiter reloads HTTP workers, dirty arbiter reloads dirty workers)
- SIGUSR1 - Log rotation across all processes
Async Signal Handling¶
The dirty arbiter uses asyncio's signal integration for safe handling in the event loop:
# Signals are registered with the event loop
loop.add_signal_handler(signal.SIGTERM, self.signal_handler, signal.SIGTERM)
def signal_handler(self, sig):
# Use call_soon_threadsafe for thread-safe event loop integration
self.loop.call_soon_threadsafe(self.handle_signal, sig)
This pattern ensures signals don't interrupt asyncio operations mid-execution, preventing race conditions and partial state updates.
Liveness and Health Monitoring¶
Dirty Arbiters implement multiple layers of health monitoring to ensure workers remain responsive and orphaned processes are cleaned up.
Heartbeat Mechanism¶
Each dirty worker maintains a "worker tmp" file whose mtime serves as a heartbeat:
Worker Lifecycle:
1. Worker spawns, creates WorkerTmp file
2. Worker touches file every (dirty_timeout / 2) seconds
3. Arbiter checks all worker mtimes every 1 second
4. If mtime > dirty_timeout seconds old, worker is killed
This file-based heartbeat has several advantages:
- OS-level tracking - No IPC required, works even if worker is stuck in C code
- Crash detection - Arbiter notices immediately when worker stops updating
- Graceful recovery - Worker killed with SIGKILL, arbiter spawns replacement
Timeout Detection¶
The arbiter's monitoring loop checks worker health every second:
# Pseudocode for worker monitoring
for worker in self.workers:
mtime = worker.tmp.last_update()
if time.time() - mtime > self.dirty_timeout:
log.warning(f"Worker {worker.pid} timed out, killing")
os.kill(worker.pid, signal.SIGKILL)
When a worker is killed:
SIGCHLDis delivered to the arbiter- Arbiter reaps the worker process
dirty_worker_exithook is called- A new worker is spawned to maintain
dirty_workerscount
Parent Death Detection¶
Dirty arbiters monitor their parent process (the main arbiter) to detect orphaning:
# In the dirty arbiter's main loop
if os.getppid() != self.parent_pid:
log.info("Parent died, shutting down")
self.alive = False
This check runs every iteration of the event loop (typically sub-millisecond). When parent death is detected:
- Arbiter sets
self.alive = False - All workers are sent SIGTERM
- Arbiter waits for graceful shutdown (up to
dirty_graceful_timeout) - Remaining workers are sent SIGKILL
- Arbiter exits
Orphan Cleanup¶
To handle edge cases where the dirty arbiter itself crashes, a well-known PID file is used:
PID file location: /tmp/gunicorn_dirty_<main_arbiter_pid>.pid
On startup, the dirty arbiter:
- Checks if PID file exists
- If yes, reads the old PID and attempts to kill it (
SIGTERM) - Waits briefly for cleanup
- Writes its own PID to the file
- On exit, removes the PID file
This ensures that if a dirty arbiter crashes and the main arbiter restarts it, the old orphaned process is terminated.
Respawn Behavior¶
| Component | Respawn Trigger | Respawn Behavior |
|---|---|---|
| Dirty Worker | Exit, timeout, or crash | Immediate respawn to maintain dirty_workers count |
| Dirty Arbiter | Exit or crash | Main arbiter respawns if not shutting down |
The dirty arbiter maintains a target worker count and continuously spawns workers until the target is reached:
while len(self.workers) < self.num_workers:
self.spawn_worker()
Monitoring Recommendations¶
For production deployments, consider:
- Log monitoring - Watch for "Worker timed out" messages indicating hung workers
- Process monitoring - Use systemd or supervisord to monitor the main arbiter
- Metrics - Track respawn frequency to detect unstable workers
# Check for recent worker timeouts
grep "Worker.*timed out" /var/log/gunicorn.log | tail -20
# Monitor process tree
watch -n 1 'pstree -p $(cat gunicorn.pid)'
Error Handling¶
The dirty client raises specific exceptions:
from gunicorn.dirty import (
DirtyError,
DirtyTimeoutError,
DirtyConnectionError,
DirtyAppError,
DirtyAppNotFoundError,
)
try:
result = client.execute("myapp.ml:MLApp", "inference", "model", data)
except DirtyTimeoutError:
# Operation timed out
pass
except DirtyAppNotFoundError:
# App not loaded in dirty workers
pass
except DirtyAppError as e:
# Error during app execution
print(f"App error: {e.message}, traceback: {e.traceback}")
except DirtyConnectionError:
# Connection to dirty arbiter failed
pass
Best Practices¶
- Pre-load commonly used resources in
init()to avoid cold starts - Set appropriate timeouts based on your workload
- Handle errors gracefully - dirty workers may restart
- Use meaningful action names for easier debugging
- Keep responses JSON-serializable - results are passed via IPC
Monitoring¶
Monitor dirty workers using standard process monitoring:
# Check dirty arbiter and workers
ps aux | grep "dirty"
# View logs
tail -f gunicorn.log | grep dirty
Example: Image Processing¶
# myapp/images.py
from gunicorn.dirty import DirtyApp
from PIL import Image
import io
class ImageApp(DirtyApp):
def init(self):
# Pre-import heavy libraries
import cv2
self.cv2 = cv2
def resize(self, image_data, width, height):
"""Resize an image."""
img = Image.open(io.BytesIO(image_data))
resized = img.resize((width, height))
buffer = io.BytesIO()
resized.save(buffer, format='PNG')
return buffer.getvalue()
def thumbnail(self, image_data, size=128):
"""Create a thumbnail."""
img = Image.open(io.BytesIO(image_data))
img.thumbnail((size, size))
buffer = io.BytesIO()
img.save(buffer, format='JPEG')
return buffer.getvalue()
def close(self):
pass
Usage:
from gunicorn.dirty import get_dirty_client
def upload_image(request):
client = get_dirty_client()
# Create thumbnail in dirty worker
thumbnail = client.execute(
"myapp.images:ImageApp",
"thumbnail",
request.files['image'].read(),
size=256
)
return save_thumbnail(thumbnail)
Complete Examples¶
For full working examples with Docker deployment, see:
- Embedding Service Example - FastAPI-based text embedding API using sentence-transformers with dirty workers for ML model management.
- Streaming Chat Example - Simulated LLM chat with token-by-token SSE streaming, demonstrating dirty worker generators and real-time response delivery.