Pyramid and Celery are both fantastic projects that compliment each other well: Pyramid processes synchronous web requests, while Celery performs asynchronous tasks in the background. Unfortunately, due to differences in structure and configuration, it's very difficult to integrate the two together. Configuration, clients, etc. available in Pyramid views may be unavailable in tasks, or may need to be accessed in a different way. Configuration and functionality may have to be duplicated in order to be shared between Pyramid and Celery. Pyramid Tasks aims to bridge this gap by creating a full Pyramid application in the Celery worker and providing a request object to every task. You can use the same configuration for Celery that you do Pyramid, including a Paste-style ini file. Bringing parity to Pyramid and Celery means you can write code for Pyramid and have the code just work in Celery.
To see Pyramid Tasks in action, check out the sample app.
To use Pyramid Tasks, you should first be familiar with Pyramid and Celery.
You can install Pyramid Tasks from PyPI:
pip install pyramid-tasks
Include Pyramid Tasks in your application using config.include
, or add it to pyramid.includes
in your ini file.
config.include('pyramid_tasks')
When you import Pyramid Tasks into your application, a new Celery application is created.
All settings prefixed with celery.
are put into Celery's configuration.
As settings from a .ini file are all strings, values are coerced as necessary.
Nested settings are supported by chaining dots, e.g. celery.broker_transport_options.queue_name_prefix
.
For example, the following simple celeryconf.py
:
broker_url = 'redis://'
broker_transport_options = {
'visibility_timeout': 3600,
}
result_backend = 'redis://'
Is equivalent to the following .ini file:
celery.broker_url = redis://
celery.broker_transport_options.visibility_timeout = 3600
celery.result_backend = redis://
If you're running Pyramid via Paste (i.e. an ini file and possibly pserve
),
you can run a Celery worker using the same ini file.
celery -A pyramid_tasks --ini config.ini
This will create a Pyramid app via the same process pserve
does, allowing you to share configuration between the two environments.
You can also create a Celery app using config.make_celery_app()
, just like you use config.make_wsgi_app()
.
If you add app = config.make_celery_app()
to celery.py
in your project's package, you can invoke celery -A myproject worker
to boot a worker.
To see both methods of running a worker in action, take a look at the sample app.
To register a new task, call config.register_task
with the task function.
You can also use the pyramid_tasks.task
decorator as long as you run a scan (config.scan()
) on the package, just like Pyramid's view_config
decorator.
register_task
and @task
take the same arguments as Celery.task.
For example, a simple Pyramid app with a task might look like the following:
from pyramid.config import Configurator
def add(request, x, y):
return x + y
with Configurator() as config:
config.register_task(add, name='add')
Once a task is registered, you can add it to the work queue using request.defer_task
.
This takes the task function or a string of the name of the task as the first argument.
The remaining arguments (positional and keyword) will be passed to the task.
When the task is invoked by a Celery worker, a request object will be created and passed as the first argument.
This request object will share the same configuration as requests in the Pyramid application.
This means it will have the same or similar methods, registry, etc.
However, it is not the request that invoked the task and properties such as url
, GET
, etc. will not be present.
To use these values in your task, pass them in as arguments.
Let's take our simple Pyramid app and add a view that invokes the task.
from pyramid.config import Configurator
def add_view(context, request):
request.defer_task(add, int(request.GET['x']), int(request.GET['y']))
return 'OK\n'
def add(request, x, y):
return x + y
with Configurator() as config:
config.add_route('root', '/')
config.add_view(add_view, route_name='root')
config.register_task(add, name='add')
You can also use request.defer_task_with_options
to pass options into Celery.
See the Celery docs for details on what options are available.
For example:
def add_view(context, request):
args = int(request.GET['x']), int(request.GET['y'])
request.defer_task_with_options(add, args=args, countdown=5)
return 'OK\n'
request.defer_task
returns a Celery AsyncResult object.
You can use this object to check if the task has completed (AsyncResult.ready()
) and to get the return value of the task (AsyncResult.result
).
See the Celery docs for more information.
AsyncResult
also has an id
property.
If you store this property somewhere, such as a client session, you can use request.get_task_result(id)
to return a new AsyncResult
object.
pyramid_tm is the recommended way of adding transaction management to Pyramid.
For example, the Pyramid cookiecutter
uses pyramid_tm
and zope.sqlalchemy
to integrate SQLAlchemy into Pyramid.
Pyramid Tasks includes built-in support for pyramid_tm.
It can be enabled by including pyramid_tasks.transaction
in your project.
This must be included after Pyramid Tasks, but doesn't need to be included before pyramid_tm.
To see Pyramid Tasks, pyramid_tm, and SQLAlchemy in action, check out the SQLAlchemy sample app.
Pyramid Tasks supports Celery Beat for running periodic tasks.
After registering a task, use config.add_periodic_task
to schedule the task.
The arguments mirror Celery.add_periodic_task:
config.add_periodic_task(
5.0, # Run every five seconds
'mytask',
('foo', 'bar'), # Position arguments passed to task
{'fizz': 'buzz'}, # Keyword arguments passed to task
)
You can also use celery.schedules.crontab as the first argument to use crontab-style scheduling.
You can run the Beat scheduler the same way you run the Celery worker.
celery -A pyramid_tasks beat --ini config.ini
To see Celery Beat in action, check out the beat sample app.
Task Derivers are analogous to Pyramid's View Derivers. They allow you to transform the task before registering it, such as wrapping the task in a transaction or adding metric collection.
A task deriver is a callable that takes two arguments: A task function and an info object. The deriver should return a task callable. The info object has the following attributes:
registry
— The registry for the current Pyramid application.package
— The package where the configuration statement was found.name
— The name of the task.options
— The options passed in to the register task action.original_func
— The original task function.
You can register a new task deriver with the Configurator.add_task_deriver
method.
The first argument is the task deriver.
The second argument is the name.
If omitted, the name of the task deriver function will be used.
It also optionally takes over
and under
arguments, which work the same as with Pyramid's view deriver.
For example, here's a simple task deriver that wraps the task in a database transaction:
def transaction_deriver(task, info):
def wrapped(request, *args, **kwargs):
with request.db:
task(request, *args, **kwargs)
return wrapped
def includeme(config):
config.add_task_deriver(transaction_deriver)
You can pass in options when registering the task to configure your task derivers.
For example, here's the same transaction task deriver as above, but now will only wrap the task if the in_transaction
option is set.
def transaction_deriver(task, info):
def wrapped(request, *args, **kwargs):
with request.db:
task(request, *args, **kwargs)
if info.options.get('in_transaction', False):
return wrapped
else:
return task
Celery will accept any keyword arguments passed in, so no configuration is necessary to use your own options. All options will be set as attributes on the task object.
Pyramid Task also fires events using Pyramid's event system.
Currently the only event is pyramid_task.events.BeforeDeferTask
, which will fire when calling defer_task
or defer_task_with_options
.
The event contains the following attributes:
request
— The current request.task
— The task being deferred.args
— The arguments being passed to the task.kwargs
— The keyword arguments being passed to the task.options
— The options being passed intoTask.apply_async
.
You can modify options
in-place and the changes will be reflected in the apply_async
call.
For example, here's an event subscriber that adds the current user ID to the headers.
def add_headers(event):
headers = event.options.setdefault('headers', {})
headers.setdefault('user_id', event.request.authenticated_userid)
def includeme(config):
config.add_subscriber(add_headers, BeforeDeferTask)
The user ID will now be accessible from request.current_task.request.user_id
.
Celery by default uses a pre-fork worker model,
meaning the application will be initialized and then forked to launch the desired number of workers.
This can cause issues with some libraries, especially ones utilizing file descriptors such as database connections.
For example, SQLAlchemy requires disposing connections on fork.
You can do this by subscribing to the pyramid_task.events.CeleryWorkerProcessInit
event.
config.add_subscriber(lambda _: engine.pool.recreate(), CeleryWorkerProcessInit)
The event includes the current application registry in the registry
property.
CeleryWorkerProcessInit
is triggered by Celery's worker_process_init
signal, so use it in the same situations you would that signal.
Pyramid Tasks is heavily inspired by the code of PyPA's Warehouse project.