delay_task), countdown and eta;
>>> # Run 10 seconds into the future. >>> res = apply_async(MyTask, countdown=10);>>> # Run 1 day from now >>> res = apply_async(MyTask, eta=datetime.now() + ... timedelta(days=1)
Now unlinks the pidfile if it’s stale.
Lots of more tests.
Now compatible with carrot >= 0.5.0.
IMPORTANT The subtask_ids attribute on the TaskSetResult instance has been removed. To get this information instead use:
>>> subtask_ids = [subtask.task_id for subtask in ts_res.subtasks]
Taskset.run() now respects extra message options from the task class.
return value. This means you can’t use the celery.result.AsyncResult to check if the task is done, or get its return value. Only use if you need the performance and is able live without these features. Any exceptions raised will store the return value/status as usual.
emails for that task.
work, so using the --detach argument results in an exception being raised.)
To start sending statistics start celeryd with the --statistics option. Then after a while you can dump the results by running python manage.py celerystats. See celery.monitoring for more information.
restarted if it crashes). To use this start celeryd with the --supervised option (or alternatively -S).
views.apply: View applying a task. Example:
http://e.com/celery/apply/task_name/arg1/arg2//?kwarg1=a&kwarg2=b
**NOTE** Use with caution, preferably not make this publicly
accessible without ensuring your code is safe!
Refactored celery.task. It’s now split into three modules:
celery.task
Contains apply_async, delay_task, discard_all, and task shortcuts, plus imports objects from celery.task.base and celery.task.builtins
celery.task.base
Contains task base classes: Task, PeriodicTask, TaskSet
celery.task.builtins
Built-in tasks: PingTask, AsynchronousMapTask, ExecuteRemoteTask, ++.
basic.get. This means we’re no longer polling the broker for new messages.
available on the system.
name, so
>>> tasks.register(func, task_name="mytask")
has to be replaced with:
>>> tasks.register(func, name="mytask")
The daemon now correctly runs if the pidlock is stale.
Now compatible with carrot 0.4.5
Default AMQP connnection timeout is now 4 seconds.
AsyncResult.read() was always returning True.
doesn’t break.
celery.view: JSON responses now properly set its mime-type.
can re-use the same AMQP connection if you want to execute more than one task.
Handle failures in task_status view such that it won’t throw 500s.
Fixed typo AMQP_SERVER in documentation to AMQP_HOST.
Worker exception e-mails sent to admins now works properly.
the preferred Django version installed.
PickledObject field.
it’s the term used in the documentation from now on.
properly at exit. (So Ctrl-C works again).
Now depends on python-daemon.
Removed dependency to simplejson
if the Django cache backend is memcached/libmemcached.
executed.
for periodic tasks to execute.
messages in the queue.
celeryd: The --wakeup-after option was not handled as a float.
of blocking the TaskController loop.
Default QUEUE_WAKEUP_AFTER has been lowered to 0.1 (was 0.3)
NOTE This is a development version, for the stable release, please see versions 0.2.x.
VERY IMPORTANT: Pickle is now the encoder used for serializing task arguments, so be sure to flush your task queue before you upgrade.
instance, which lets you inspect the status and return values of a taskset as it was a single entity.
IMPORTANT Celery now depends on carrot >= 0.4.1.
To turn off this feature, set SEND_CELERY_TASK_ERROR_EMAILS to False in your settings.py. Thanks to Grégoire Cachet.
You can now run the celery daemon by using manage.py:
$ python manage.py celeryd
Thanks to Grégoire Cachet.
keys for tasks. This means we have introduced celery.task.apply_async, a new way of executing tasks.
You can use celery.task.delay and celery.Task.delay like usual, but if you want greater control over the message sent, you want celery.task.apply_async and celery.Task.apply_async.
This also means the AMQP configuration has changed. Some settings has been renamed, while others are new:
CELERY_AMQP_EXCHANGE
CELERY_AMQP_PUBLISHER_ROUTING_KEY
CELERY_AMQP_CONSUMER_ROUTING_KEY
CELERY_AMQP_CONSUMER_QUEUE
CELERY_AMQP_EXCHANGE_TYPE
See the entry Can I send some tasks to only some servers? in the FAQ for more information.
and backtraces are dumped. Thanks to Grégoire Cachet.
this solving the “MySQL connection died?” exceptions. Thanks to Vitaly Babiy and Jirka Vejrazka.
now can pass complex python objects to tasks as arguments.
Removed dependency on yadayada.
Added a FAQ, see docs/faq.rst.
Thanks Vitaly Babiy.
Renamed the TaskDaemon to WorkController.
celery.pool.TaskPool.
stability.
get_result() now tries to recreate something looking like the original exception.
Forgot to add yadayada to install requirements.
Now deletes all expired task results, not just those marked as done.
Able to load the Tokyo Tyrant backend class without django configuration, can specify tyrant settings directly in the class constructor.
Improved API documentation
Now using the Sphinx documentation system, you can build the html documentation by doing
$ cd docs
$ make html
and the result will be in docs/.build/html.
delay_task() etc. now returns celery.task.AsyncResult object, which lets you check the result and any failure that might have happened. It kind of works like the multiprocessing.AsyncResult class returned by multiprocessing.Pool.map_async.
Added dmap() and dmap_async(). This works like the multiprocessing.Pool versions except they are tasks distributed to the celery server. Example:
>>> from celery.task import dmap >>> import operator >>> dmap(operator.add, [[2, 2], [4, 4], [8, 8]]) >>> [4, 8, 16]>>> from celery.task import dmap_async >>> import operator >>> result = dmap_async(operator.add, [[2, 2], [4, 4], [8, 8]]) >>> result.ready() False >>> time.sleep(1) >>> result.ready() True >>> result.result [4, 8, 16]
Refactored the task metadata cache and database backends, and added a new backend for Tokyo Tyrant. You can set the backend in your django settings file. e.g
CELERY_BACKEND = “database”; # Uses the database
CELERY_BACKEND = “cache”; # Uses the django cache framework
CELERY_BACKEND = “tyrant”; # Uses Tokyo Tyrant TT_HOST = “localhost”; # Hostname for the Tokyo Tyrant server. TT_PORT = 6657; # Port of the Tokyo Tyrant server.
- Introducing TaskSet. A set of subtasks is executed and you can find out how many, or if all them, are done (excellent for progress bars and such)
- Now catches all exceptions when running Task.__call__, so the daemon doesn’t die. This does’t happen for pure functions yet, only Task classes.
- autodiscover() now works with zipped eggs.
- celeryd: Now adds curernt working directory to sys.path for convenience.
- The run_every attribute of PeriodicTask classes can now be a datetime.timedelta() object.
- celeryd: You can now set the DJANGO_PROJECT_DIR variable for celeryd and it will add that to sys.path for easy launching.
- Can now check if a task has been executed or not via HTTP.
You can do this by including the celery urls.py into your project,
>>> url(r'^celery/$', include("celery.urls"))then visiting the following url,:
http://mysite/celery/$task_id/done/this will return a JSON dictionary like e.g:
>>> {"task": {"id": $task_id, "executed": true}}
- delay_task now returns string id, not uuid.UUID instance.
- Now has PeriodicTasks, to have cron like functionality.
- Project changed name from crunchy to celery. The details of the name change request is in docs/name_change_request.txt.