No Description

Ask Solem 620b03faca Program to stress test celery/carrot in different ways to find connection leakage etc. 16 years ago
bin 9677c67764 Forgot to import sys in bin/celeryd 16 years ago
celery 98e6babb3b Bumped verison to 0.1.15, updated changelog and added link to API documentation in README.rst 16 years ago
contrib 620b03faca Program to stress test celery/carrot in different ways to find connection leakage etc. 16 years ago
docs 6787df2dac Fix celery daemon documentaion. 16 years ago
testproj 1cda75ac78 Test coverage now at 95% (that's without celeryd/worker.py/process/platform) 16 years ago
.gitignore 2068f1897b Added .build to gitignore 16 years ago
AUTHORS 39bb72c5ab Added runeh and antont as name contributors. 16 years ago
Changelog 98e6babb3b Bumped verison to 0.1.15, updated changelog and added link to API documentation in README.rst 16 years ago
LICENSE 8dd7ac781a And here comes the code 16 years ago
MANIFEST.in cf0ebe2664 Updated MANIFEST.in 16 years ago
README.rst 98e6babb3b Bumped verison to 0.1.15, updated changelog and added link to API documentation in README.rst 16 years ago
TODO 0f7dc467e1 Added more TODO items. 16 years ago
setup.py 6787df2dac Fix celery daemon documentaion. 16 years ago

README.rst

============================================
celery - Distributed Task Queue for Django.
============================================

:Authors:
Ask Solem (askh@opera.com)
:Version: 0.1.15

Introduction
------------

``celery`` is a distributed task queue framework for Django.
More information will follow.

Be sure to also consult the `API Reference Documentation`_

.. _`API Reference Docmentation`: http://ask.github.com/celery/

Installation
=============

You can install ``celery`` either via the Python Package Index (PyPI)
or from source.

To install using ``pip``,::

$ pip install celery

To install using ``easy_install``,::

$ easy_install celery

If you have downloaded a source tarball you can install it
by doing the following,::

$ python setup.py build
# python setup.py install # as root

Usage
=====

Have to write a cool tutorial, but here is some simple usage info.

*Note* You need to have a AMQP message broker running, like `RabbitMQ`_,
and you need to have the amqp server setup in your settings file, as described
in the `carrot distribution README`_.

*Note* If you're running ``SQLite`` as the database backend, ``celeryd`` will
only be able to process one message at a time, this because ``SQLite`` doesn't
allow concurrent writes.

.. _`RabbitMQ`: http://www.rabbitmq.com
.. _`carrot distribution README`: http://pypi.python.org/pypi/carrot/0.3.3


Defining tasks
--------------

>>> from celery.task import tasks
>>> from celery.log import setup_logger
>>> def do_something(some_arg, **kwargs):
... logger = setup_logger(**kwargs)
... logger.info("Did something: %s" % some_arg)
>>> task.register(do_something, "do_something")

Tell the celery daemon to run a task
-------------------------------------

>>> from celery.task import delay_task
>>> delay_task("do_something", some_arg="foo bar baz")


Running the celery daemon
--------------------------

::

$ cd mydjangoproject
$ env DJANGO_SETTINGS_MODULE=settings celeryd
[....]
[2009-04-23 17:44:05,115: INFO/Process-1] Did something: foo bar baz
[2009-04-23 17:44:05,118: INFO/MainProcess] Waiting for queue.




Autodiscovery of tasks
-----------------------

``celery`` has an autodiscovery feature like the Django Admin, that
automatically loads any ``tasks.py`` module in the applications listed
in ``settings.INSTALLED_APPS``.

A good place to add this command could be in your ``urls.py``,
::

from celery.task import tasks
tasks.autodiscover()



Then you can add new tasks in your applications ``tasks.py`` module,
::

from celery.task import tasks
from celery.log import setup_logger
from clickcounter.models import ClickCount

def increment_click(for_url, **kwargs):
logger = setup_logger(**kwargs)
clicks_for_url, cr = ClickCount.objects.get_or_create(url=for_url)
clicks_for_url.clicks = clicks_for_url.clicks + 1
clicks_for_url.save()
logger.info("Incremented click count for %s (not at %d)" % (
for_url, clicks_for_url.clicks)
tasks.register(increment_click, "increment_click")


Periodic Tasks
---------------

Periodic tasks are tasks that are run every ``n`` seconds. They don't
support extra arguments. Here's an example of a periodic task:


>>> from celery.task import tasks, PeriodicTask
>>> from datetime import timedelta
>>> class MyPeriodicTask(PeriodicTask):
... name = "foo.my-periodic-task"
... run_every = timedelta(seconds=30)
...
... def run(self, **kwargs):
... logger = self.get_logger(**kwargs)
... logger.info("Running periodic task!")
...
>>> tasks.register(MyPeriodicTask)


For periodic tasks to work you need to add ``celery`` to ``INSTALLED_APPS``,
and issue a ``syncdb``.

License
=======

This software is licensed under the ``New BSD License``. See the ``LICENSE``
file in the top distribution directory for the full license text.

.. # vim: syntax=rst expandtab tabstop=4 shiftwidth=4 shiftround