123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530 |
- .. _configuration:
- ============================
- Configuration and defaults
- ============================
- This document describes the configuration options available.
- If you're using the default loader, you must create the :file:`celeryconfig.py`
- module and make sure it is available on the Python path.
- .. contents::
- :local:
- :depth: 2
- .. _conf-example:
- Example configuration file
- ==========================
- This is an example configuration file to get you started.
- It should contain all you need to run a basic Celery set-up.
- .. code-block:: python
- ## Broker settings.
- BROKER_URL = "amqp://guest:guest@localhost:5672//"
- # List of modules to import when celery starts.
- CELERY_IMPORTS = ("myapp.tasks", )
- ## Using the database to store task state and results.
- CELERY_RESULT_BACKEND = "database"
- CELERY_RESULT_DBURI = "sqlite:///mydatabase.db"
- CELERY_ANNOTATIONS = {"tasks.add": {"rate_limit": "10/s"}}
- Configuration Directives
- ========================
- .. _conf-datetime:
- Time and date settings
- ----------------------
- .. setting:: CELERY_ENABLE_UTC
- CELERY_ENABLE_UTC
- ~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.5
- If enabled dates and times in messages will be converted to use
- the UTC timezone.
- Note that workers running Celery versions below 2.5 will assume a local
- timezone for all messages, so only enable if all workers have been
- upgraded.
- Enabled by default since version 3.0.
- .. setting:: CELERY_TIMEZONE
- CELERY_TIMEZONE
- ---------------
- Configure Celery to use a custom time zone.
- The timezone value can be any time zone supported by the :mod:`pytz`
- library. :mod:`pytz` must be installed for the selected zone
- to be used.
- If not set then the systems default local time zone is used.
- .. warning::
- Celery requires the :mod:`pytz` library to be installed,
- when using custom time zones (other than UTC). You can
- install it using :program:`pip` or :program:`easy_install`::
- $ pip install pytz
- Pytz is a library that defines the timzones of the world,
- it changes quite frequently so it is not included in the Python Standard
- Library.
- .. _conf-tasks:
- Task settings
- -------------
- .. setting:: CELERY_ANNOTATIONS
- CELERY_ANNOTATIONS
- ~~~~~~~~~~~~~~~~~~
- This setting can be used to rewrite any task attribute from the
- configuration. The setting can be a dict, or a list of annotation
- objects that filter for tasks and return a map of attributes
- to change.
- This will change the ``rate_limit`` attribute for the ``tasks.add``
- task:
- .. code-block:: python
- CELERY_ANNOTATIONS = {"tasks.add": {"rate_limit": "10/s"}}
- or change the same for all tasks:
- .. code-block:: python
- CELERY_ANNOTATIONS = {"*": {"rate_limit": "10/s"}}
- You can change methods too, for example the ``on_failure`` handler:
- .. code-block:: python
- def my_on_failure(self, exc, task_id, args, kwargs, einfo):
- print("Oh no! Task failed: %r" % (exc, ))
- CELERY_ANNOTATIONS = {"*": {"on_failure": my_on_failure}}
- If you need more flexibility then you can use objects
- instead of a dict to choose which tasks to annotate:
- .. code-block:: python
- class MyAnnotate(object):
- def annotate(self, task):
- if task.name.startswith("tasks."):
- return {"rate_limit": "10/s"}
- CELERY_ANNOTATIONS = (MyAnnotate(), {...})
- .. _conf-concurrency:
- Concurrency settings
- --------------------
- .. setting:: CELERYD_CONCURRENCY
- CELERYD_CONCURRENCY
- ~~~~~~~~~~~~~~~~~~~
- The number of concurrent worker processes/threads/green threads executing
- tasks.
- If you're doing mostly I/O you can have more processes,
- but if mostly CPU-bound, try to keep it close to the
- number of CPUs on your machine. If not set, the number of CPUs/cores
- on the host will be used.
- Defaults to the number of available CPUs.
- .. setting:: CELERYD_PREFETCH_MULTIPLIER
- CELERYD_PREFETCH_MULTIPLIER
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~
- How many messages to prefetch at a time multiplied by the number of
- concurrent processes. The default is 4 (four messages for each
- process). The default setting is usually a good choice, however -- if you
- have very long running tasks waiting in the queue and you have to start the
- workers, note that the first worker to start will receive four times the
- number of messages initially. Thus the tasks may not be fairly distributed
- to the workers.
- .. _conf-result-backend:
- Task result backend settings
- ----------------------------
- .. setting:: CELERY_RESULT_BACKEND
- CELERY_RESULT_BACKEND
- ~~~~~~~~~~~~~~~~~~~~~
- :Deprecated aliases: ``CELERY_BACKEND``
- The backend used to store task results (tombstones).
- Disabled by default.
- Can be one of the following:
- * database
- Use a relational database supported by `SQLAlchemy`_.
- See :ref:`conf-database-result-backend`.
- * cache
- Use `memcached`_ to store the results.
- See :ref:`conf-cache-result-backend`.
- * mongodb
- Use `MongoDB`_ to store the results.
- See :ref:`conf-mongodb-result-backend`.
- * redis
- Use `Redis`_ to store the results.
- See :ref:`conf-redis-result-backend`.
- * amqp
- Send results back as AMQP messages
- See :ref:`conf-amqp-result-backend`.
- * cassandra
- Use `Cassandra`_ to store the results.
- See :ref:`conf-cassandra-result-backend`.
- .. warning:
- While the AMQP result backend is very efficient, you must make sure
- you only receive the same result once. See :doc:`userguide/calling`).
- .. _`SQLAlchemy`: http://sqlalchemy.org
- .. _`memcached`: http://memcached.org
- .. _`MongoDB`: http://mongodb.org
- .. _`Redis`: http://code.google.com/p/redis/
- .. _`Cassandra`: http://cassandra.apache.org/
- .. setting:: CELERY_RESULT_SERIALIZER
- CELERY_RESULT_SERIALIZER
- ~~~~~~~~~~~~~~~~~~~~~~~~
- Result serialization format. Default is `"pickle"`. See
- :ref:`calling-serializers` for information about supported
- serialization formats.
- .. _conf-database-result-backend:
- Database backend settings
- -------------------------
- .. setting:: CELERY_RESULT_DBURI
- CELERY_RESULT_DBURI
- ~~~~~~~~~~~~~~~~~~~
- Please see `Supported Databases`_ for a table of supported databases.
- To use this backend you need to configure it with an
- `Connection String`_, some examples include:
- .. code-block:: python
- # sqlite (filename)
- CELERY_RESULT_DBURI = "sqlite:///celerydb.sqlite"
- # mysql
- CELERY_RESULT_DBURI = "mysql://scott:tiger@localhost/foo"
- # postgresql
- CELERY_RESULT_DBURI = "postgresql://scott:tiger@localhost/mydatabase"
- # oracle
- CELERY_RESULT_DBURI = "oracle://scott:tiger@127.0.0.1:1521/sidname"
- See `Connection String`_ for more information about connection
- strings.
- .. setting:: CELERY_RESULT_ENGINE_OPTIONS
- CELERY_RESULT_ENGINE_OPTIONS
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- To specify additional SQLAlchemy database engine options you can use
- the :setting:`CELERY_RESULT_ENGINE_OPTIONS` setting::
- # echo enables verbose logging from SQLAlchemy.
- CELERY_RESULT_ENGINE_OPTIONS = {"echo": True}
- .. setting:: CELERY_RESULT_DB_SHORT_LIVED_SESSIONS
- CELERY_RESULT_DB_SHORT_LIVED_SESSIONS = True
- Short lived sessions are disabled by default. If enabled they can drastically reduce
- performance, especially on systems processing lots of tasks. This option is useful
- on low-traffic workers that experience errors as a result of cached database connections
- going stale through inactivity. For example, intermittent errors like
- `(OperationalError) (2006, 'MySQL server has gone away')` can be fixed by enabling
- short lived sessions. This option only affects the database backend.
- .. _`Supported Databases`:
- http://www.sqlalchemy.org/docs/core/engines.html#supported-databases
- .. _`Connection String`:
- http://www.sqlalchemy.org/docs/core/engines.html#database-urls
- Example configuration
- ~~~~~~~~~~~~~~~~~~~~~
- .. code-block:: python
- CELERY_RESULT_BACKEND = "database"
- CELERY_RESULT_DBURI = "mysql://user:password@host/dbname"
- .. _conf-amqp-result-backend:
- AMQP backend settings
- ---------------------
- .. note::
- The AMQP backend requires RabbitMQ 1.1.0 or higher to automatically
- expire results. If you are running an older version of RabbitmQ
- you should disable result expiration like this:
- CELERY_TASK_RESULT_EXPIRES = None
- .. setting:: CELERY_RESULT_EXCHANGE
- CELERY_RESULT_EXCHANGE
- ~~~~~~~~~~~~~~~~~~~~~~
- Name of the exchange to publish results in. Default is `"celeryresults"`.
- .. setting:: CELERY_RESULT_EXCHANGE_TYPE
- CELERY_RESULT_EXCHANGE_TYPE
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~
- The exchange type of the result exchange. Default is to use a `direct`
- exchange.
- .. setting:: CELERY_RESULT_PERSISTENT
- CELERY_RESULT_PERSISTENT
- ~~~~~~~~~~~~~~~~~~~~~~~~
- If set to :const:`True`, result messages will be persistent. This means the
- messages will not be lost after a broker restart. The default is for the
- results to be transient.
- Example configuration
- ~~~~~~~~~~~~~~~~~~~~~
- .. code-block:: python
- CELERY_RESULT_BACKEND = "amqp"
- CELERY_TASK_RESULT_EXPIRES = 18000 # 5 hours.
- .. _conf-cache-result-backend:
- Cache backend settings
- ----------------------
- .. note::
- The cache backend supports the `pylibmc`_ and `python-memcached`
- libraries. The latter is used only if `pylibmc`_ is not installed.
- .. setting:: CELERY_CACHE_BACKEND
- CELERY_CACHE_BACKEND
- ~~~~~~~~~~~~~~~~~~~~
- Using a single memcached server:
- .. code-block:: python
- CELERY_CACHE_BACKEND = 'memcached://127.0.0.1:11211/'
- Using multiple memcached servers:
- .. code-block:: python
- CELERY_RESULT_BACKEND = "cache"
- CELERY_CACHE_BACKEND = 'memcached://172.19.26.240:11211;172.19.26.242:11211/'
- .. setting:: CELERY_CACHE_BACKEND_OPTIONS
- The "dummy" backend stores the cache in memory only:
- CELERY_CACHE_BACKEND = "dummy"
- CELERY_CACHE_BACKEND_OPTIONS
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- You can set pylibmc options using the :setting:`CELERY_CACHE_BACKEND_OPTIONS`
- setting:
- .. code-block:: python
- CELERY_CACHE_BACKEND_OPTIONS = {"binary": True,
- "behaviors": {"tcp_nodelay": True}}
- .. _`pylibmc`: http://sendapatch.se/projects/pylibmc/
- .. _conf-redis-result-backend:
- Redis backend settings
- ----------------------
- Configuring the backend URL
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~
- .. note::
- The Redis backend requires the :mod:`redis` library:
- http://pypi.python.org/pypi/redis/
- To install the redis package use `pip` or `easy_install`::
- $ pip install redis
- This backend requires the :setting:`CELERY_RESULT_BACKEND`
- setting to be set to a Redis URL::
- CELERY_RESULT_BACKEND = "redis://:password@host:port/db"
- For example::
- CELERY_RESULT_BACKEND = "redis://localhost/0"
- which is the same as::
- CELERY_RESULT_BACKEND = "redis://"
- The fields of the URL is defined as folows:
- - *host*
- Host name or IP address of the Redis server. e.g. `"localhost"`.
- - *port*
- Port to the Redis server. Default is 6379.
- - *db*
- Database number to use. Default is 0.
- The db can include an optional leading slash.
- - *password*
- Password used to connect to the database.
- .. setting:: CELERY_REDIS_MAX_CONNECTIONS
- CELERY_REDIS_MAX_CONNECTIONS
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- Maximum number of connections available in the Redis connection
- pool used for sending and retrieving results.
- .. _conf-mongodb-result-backend:
- MongoDB backend settings
- ------------------------
- .. note::
- The MongoDB backend requires the :mod:`pymongo` library:
- http://github.com/mongodb/mongo-python-driver/tree/master
- .. setting:: CELERY_MONGODB_BACKEND_SETTINGS
- CELERY_MONGODB_BACKEND_SETTINGS
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- This is a dict supporting the following keys:
- * host
- Host name of the MongoDB server. Defaults to "localhost".
- * port
- The port the MongoDB server is listening to. Defaults to 27017.
- * user
- User name to authenticate to the MongoDB server as (optional).
- * password
- Password to authenticate to the MongoDB server (optional).
- * database
- The database name to connect to. Defaults to "celery".
- * taskmeta_collection
- The collection name to store task meta data.
- Defaults to "celery_taskmeta".
- .. _example-mongodb-result-config:
- Example configuration
- ~~~~~~~~~~~~~~~~~~~~~
- .. code-block:: python
- CELERY_RESULT_BACKEND = "mongodb"
- CELERY_MONGODB_BACKEND_SETTINGS = {
- "host": "192.168.1.100",
- "port": 30000,
- "database": "mydb",
- "taskmeta_collection": "my_taskmeta_collection",
- }
- .. _conf-cassandra-result-backend:
- Cassandra backend settings
- --------------------------
- .. note::
- The Cassandra backend requires the :mod:`pycassa` library:
- http://pypi.python.org/pypi/pycassa/
- To install the pycassa package use `pip` or `easy_install`::
- $ pip install pycassa
- This backend requires the following configuration directives to be set.
- .. setting:: CASSANDRA_SERVERS
- CASSANDRA_SERVERS
- ~~~~~~~~~~~~~~~~~
- List of ``host:port`` Cassandra servers. e.g. ``["localhost:9160]"``.
- .. setting:: CASSANDRA_KEYSPACE
- CASSANDRA_KEYSPACE
- ~~~~~~~~~~~~~~~~~~
- The keyspace in which to store the results. e.g. ``"tasks_keyspace"``.
- .. setting:: CASSANDRA_COLUMN_FAMILY
- CASSANDRA_COLUMN_FAMILY
- ~~~~~~~~~~~~~~~~~~~~~~~
- The column family in which to store the results. eg ``"tasks"``
- .. setting:: CASSANDRA_READ_CONSISTENCY
- CASSANDRA_READ_CONSISTENCY
- ~~~~~~~~~~~~~~~~~~~~~~~~~~
- The read consistency used. Values can be ``"ONE"``, ``"QUORUM"`` or ``"ALL"``.
- .. setting:: CASSANDRA_WRITE_CONSISTENCY
- CASSANDRA_WRITE_CONSISTENCY
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~
- The write consistency used. Values can be ``"ONE"``, ``"QUORUM"`` or ``"ALL"``.
- .. setting:: CASSANDRA_DETAILED_MODE
- CASSANDRA_DETAILED_MODE
- ~~~~~~~~~~~~~~~~~~~~~~~
- Enable or disable detailed mode. Default is :const:`False`.
- This mode allows to use the power of Cassandra wide columns to
- store all states for a task as a wide column, instead of only the last one.
- To use this mode, you need to configure your ColumnFamily to
- use the ``TimeUUID`` type as a comparator::
- create column family task_results with comparator = TimeUUIDType;
- Example configuration
- ~~~~~~~~~~~~~~~~~~~~~
- .. code-block:: python
- CASSANDRA_SERVERS = ["localhost:9160"]
- CASSANDRA_KEYSPACE = "celery"
- CASSANDRA_COLUMN_FAMILY = "task_results"
- CASSANDRA_READ_CONSISTENCY = "ONE"
- CASSANDRA_WRITE_CONSISTENCY = "ONE"
- CASSANDRA_DETAILED_MODE = True
- .. _conf-messaging:
- Message Routing
- ---------------
- .. _conf-messaging-routing:
- .. setting:: CELERY_QUEUES
- CELERY_QUEUES
- ~~~~~~~~~~~~~
- The mapping of queues the worker consumes from. This is a dictionary
- of queue name/options. See :ref:`guide-routing` for more information.
- The default is a queue/exchange/binding key of `"celery"`, with
- exchange type `direct`.
- You don't have to care about this unless you want custom routing facilities.
- .. setting:: CELERY_ROUTES
- CELERY_ROUTES
- ~~~~~~~~~~~~~
- A list of routers, or a single router used to route tasks to queues.
- When deciding the final destination of a task the routers are consulted
- in order. See :ref:`routers` for more information.
- .. setting:: CELERY_WORKER_DIRECT
- CELERY_WORKER_DIRECT
- ~~~~~~~~~~~~~~~~~~~~
- This option enables so that every worker has a dedicated queue,
- so that tasks can be routed to specific workers.
- The queue name for each worker is automatically generated based on
- the worker hostname and a ``.dq`` suffix, using the ``C.dq`` exchange.
- For example the queue name for the worker with hostname ``w1.example.com``
- becomes::
- w1.example.com.dq
- Then you can route the task to the task by specifying the hostname
- as the routung key and the ``C.dq`` exchange::
- CELERY_ROUTES = {
- 'tasks.add': {'exchange': 'C.dq', 'routing_key': 'w1.example.com'}
- }
- This setting is mandatory if you want to use the ``move_to_worker`` features
- of :mod:`celery.contrib.migrate`.
- .. setting:: CELERY_CREATE_MISSING_QUEUES
- CELERY_CREATE_MISSING_QUEUES
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- If enabled (default), any queues specified that is not defined in
- :setting:`CELERY_QUEUES` will be automatically created. See
- :ref:`routing-automatic`.
- .. setting:: CELERY_DEFAULT_QUEUE
- CELERY_DEFAULT_QUEUE
- ~~~~~~~~~~~~~~~~~~~~
- The name of the default queue used by `.apply_async` if the message has
- no route or no custom queue has been specified.
- This queue must be listed in :setting:`CELERY_QUEUES`.
- If :setting:`CELERY_QUEUES` is not specified then it this automatically
- created containing one queue entry, where this name is used as the name of
- that queue.
- The default is: `celery`.
- .. seealso::
- :ref:`routing-changing-default-queue`
- .. setting:: CELERY_DEFAULT_EXCHANGE
- CELERY_DEFAULT_EXCHANGE
- ~~~~~~~~~~~~~~~~~~~~~~~
- Name of the default exchange to use when no custom exchange is
- specified for a key in the :setting:`CELERY_QUEUES` setting.
- The default is: `celery`.
- .. setting:: CELERY_DEFAULT_EXCHANGE_TYPE
- CELERY_DEFAULT_EXCHANGE_TYPE
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- Default exchange type used when no custom exchange type is specified.
- for a key in the :setting:`CELERY_QUEUES` setting.
- The default is: `direct`.
- .. setting:: CELERY_DEFAULT_ROUTING_KEY
- CELERY_DEFAULT_ROUTING_KEY
- ~~~~~~~~~~~~~~~~~~~~~~~~~~
- The default routing key used when no custom routing key
- is specified for a key in the :setting:`CELERY_QUEUES` setting.
- The default is: `celery`.
- .. setting:: CELERY_DEFAULT_DELIVERY_MODE
- CELERY_DEFAULT_DELIVERY_MODE
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- Can be `transient` or `persistent`. The default is to send
- persistent messages.
- .. _conf-broker-settings:
- Broker Settings
- ---------------
- .. setting:: BROKER_TRANSPORT
- BROKER_TRANSPORT
- ~~~~~~~~~~~~~~~~
- :Aliases: ``BROKER_BACKEND``
- :Deprecated aliases: ``CARROT_BACKEND``
- .. setting:: BROKER_URL
- BROKER_URL
- ~~~~~~~~~~
- Default broker URL. This must be an URL in the form of::
- transport://userid:password@hostname:port/virtual_host
- Only the scheme part (``transport://``) is required, the rest
- is optional, and defaults to the specific transports default values.
- The transport part is the broker implementation to use, and the
- default is ``amqp``, but there are many other choices including
- ``librabbitmq``, ``amqplib``, ``redis``, ``beanstalk``,
- ``sqlalchemy``, ``django``, ``mongodb``, ``couchdb`` and ``pika``.
- It can also be a fully qualified path to your own transport implementation.
- See the Kombu documentation for more information about broker URLs.
- .. setting:: BROKER_USE_SSL
- BROKER_USE_SSL
- ~~~~~~~~~~~~~~
- Use SSL to connect to the broker. Off by default. This may not be supported
- by all transports.
- .. setting:: BROKER_POOL_LIMIT
- BROKER_POOL_LIMIT
- ~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.3
- The maximum number of connections that can be open in the connection pool.
- The pool is enabled by default since version 2.5, with a default limit of ten
- connections. This number can be tweaked depending on the number of
- threads/greenthreads (eventlet/gevent) using a connection. For example
- running eventlet with 1000 greenlets that use a connection to the broker,
- contention can arise and you should consider increasing the limit.
- If set to :const:`None` or 0 the connection pool will be disabled and
- connections will be established and closed for every use.
- Default (since 2.5) is to use a pool of 10 connections.
- .. setting:: BROKER_CONNECTION_TIMEOUT
- BROKER_CONNECTION_TIMEOUT
- ~~~~~~~~~~~~~~~~~~~~~~~~~
- The default timeout in seconds before we give up establishing a connection
- to the AMQP server. Default is 4 seconds.
- .. setting:: BROKER_CONNECTION_RETRY
- BROKER_CONNECTION_RETRY
- ~~~~~~~~~~~~~~~~~~~~~~~
- Automatically try to re-establish the connection to the AMQP broker if lost.
- The time between retries is increased for each retry, and is
- not exhausted before :setting:`BROKER_CONNECTION_MAX_RETRIES` is
- exceeded.
- This behavior is on by default.
- .. setting:: BROKER_CONNECTION_MAX_RETRIES
- BROKER_CONNECTION_MAX_RETRIES
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- Maximum number of retries before we give up re-establishing a connection
- to the AMQP broker.
- If this is set to :const:`0` or :const:`None`, we will retry forever.
- Default is 100 retries.
- .. setting:: BROKER_TRANSPORT_OPTIONS
- BROKER_TRANSPORT_OPTIONS
- ~~~~~~~~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.2
- A dict of additional options passed to the underlying transport.
- See your transport user manual for supported options (if any).
- .. _conf-task-execution:
- Task execution settings
- -----------------------
- .. setting:: CELERY_ALWAYS_EAGER
- CELERY_ALWAYS_EAGER
- ~~~~~~~~~~~~~~~~~~~
- If this is :const:`True`, all tasks will be executed locally by blocking until
- the task returns. ``apply_async()`` and ``Task.delay()`` will return
- an :class:`~celery.result.EagerResult` instance, which emulates the API
- and behavior of :class:`~celery.result.AsyncResult`, except the result
- is already evaluated.
- That is, tasks will be executed locally instead of being sent to
- the queue.
- .. setting:: CELERY_EAGER_PROPAGATES_EXCEPTIONS
- CELERY_EAGER_PROPAGATES_EXCEPTIONS
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- If this is :const:`True`, eagerly executed tasks (applied by `task.apply()`,
- or when the :setting:`CELERY_ALWAYS_EAGER` setting is enabled), will
- propagate exceptions.
- It's the same as always running ``apply()`` with ``throw=True``.
- .. setting:: CELERY_IGNORE_RESULT
- CELERY_IGNORE_RESULT
- ~~~~~~~~~~~~~~~~~~~~
- Whether to store the task return values or not (tombstones).
- If you still want to store errors, just not successful return values,
- you can set :setting:`CELERY_STORE_ERRORS_EVEN_IF_IGNORED`.
- .. setting:: CELERY_MESSAGE_COMPRESSION
- CELERY_MESSAGE_COMPRESSION
- ~~~~~~~~~~~~~~~~~~~~~~~~~~
- Default compression used for task messages.
- Can be ``"gzip"``, ``"bzip2"`` (if available), or any custom
- compression schemes registered in the Kombu compression registry.
- The default is to send uncompressed messages.
- .. setting:: CELERY_TASK_RESULT_EXPIRES
- CELERY_TASK_RESULT_EXPIRES
- ~~~~~~~~~~~~~~~~~~~~~~~~~~
- Time (in seconds, or a :class:`~datetime.timedelta` object) for when after
- stored task tombstones will be deleted.
- A built-in periodic task will delete the results after this time
- (:class:`celery.task.backend_cleanup`).
- .. note::
- For the moment this only works with the amqp, database, cache, redis and MongoDB
- backends.
- When using the database or MongoDB backends, `celerybeat` must be
- running for the results to be expired.
- .. setting:: CELERY_MAX_CACHED_RESULTS
- CELERY_MAX_CACHED_RESULTS
- ~~~~~~~~~~~~~~~~~~~~~~~~~
- Result backends caches ready results used by the client.
- This is the total number of results to cache before older results are evicted.
- The default is 5000.
- .. setting:: CELERY_TRACK_STARTED
- CELERY_TRACK_STARTED
- ~~~~~~~~~~~~~~~~~~~~
- If :const:`True` the task will report its status as "started" when the
- task is executed by a worker. The default value is :const:`False` as
- the normal behaviour is to not report that level of granularity. Tasks
- are either pending, finished, or waiting to be retried. Having a "started"
- state can be useful for when there are long running tasks and there is a
- need to report which task is currently running.
- .. setting:: CELERY_TASK_SERIALIZER
- CELERY_TASK_SERIALIZER
- ~~~~~~~~~~~~~~~~~~~~~~
- A string identifying the default serialization method to use. Can be
- `pickle` (default), `json`, `yaml`, `msgpack` or any custom serialization
- methods that have been registered with :mod:`kombu.serialization.registry`.
- .. seealso::
- :ref:`calling-serializers`.
- .. setting:: CELERY_TASK_PUBLISH_RETRY
- CELERY_TASK_PUBLISH_RETRY
- ~~~~~~~~~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.2
- Decides if publishing task messages will be retried in the case
- of connection loss or other connection errors.
- See also :setting:`CELERY_TASK_PUBLISH_RETRY_POLICY`.
- Disabled by default.
- .. setting:: CELERY_TASK_PUBLISH_RETRY_POLICY
- CELERY_TASK_PUBLISH_RETRY_POLICY
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.2
- Defines the default policy when retrying publishing a task message in
- the case of connection loss or other connection errors.
- See :ref:`calling-retry` for more information.
- .. setting:: CELERY_DEFAULT_RATE_LIMIT
- CELERY_DEFAULT_RATE_LIMIT
- ~~~~~~~~~~~~~~~~~~~~~~~~~
- The global default rate limit for tasks.
- This value is used for tasks that does not have a custom rate limit
- The default is no rate limit.
- .. setting:: CELERY_DISABLE_RATE_LIMITS
- CELERY_DISABLE_RATE_LIMITS
- ~~~~~~~~~~~~~~~~~~~~~~~~~~
- Disable all rate limits, even if tasks has explicit rate limits set.
- .. setting:: CELERY_ACKS_LATE
- CELERY_ACKS_LATE
- ~~~~~~~~~~~~~~~~
- Late ack means the task messages will be acknowledged **after** the task
- has been executed, not *just before*, which is the default behavior.
- .. seealso::
- FAQ: :ref:`faq-acks_late-vs-retry`.
- .. _conf-celeryd:
- Worker: celeryd
- ---------------
- .. setting:: CELERY_IMPORTS
- CELERY_IMPORTS
- ~~~~~~~~~~~~~~
- A sequence of modules to import when the celery daemon starts.
- This is used to specify the task modules to import, but also
- to import signal handlers and additional remote control commands, etc.
- .. setting:: CELERY_INCLUDE
- CELERY_INCLUDE
- ~~~~~~~~~~~~~~
- Exact same semantics as :setting:`CELERY_IMPORTS`, but can be used as a means
- to have different import categories.
- .. setting:: CELERYD_FORCE_EXECV
- CELERYD_FORCE_EXECV
- ~~~~~~~~~~~~~~~~~~~
- On Unix the processes pool will fork, so that child processes
- start with the same memory as the parent process.
- This can cause problems as there is a known deadlock condition
- with pthread locking primitives when `fork()` is combined with threads.
- You should enable this setting if you are experiencing hangs (deadlocks),
- especially in combination with time limits or having a max tasks per child limit.
- This option will be enabled by default in a later version.
- This is not a problem on Windows, as it does not have `fork()`.
- .. setting:: CELERYD_WORKER_LOST_WAIT
- CELERYD_WORKER_LOST_WAIT
- ~~~~~~~~~~~~~~~~~~~~~~~~
- In some cases a worker may be killed without proper cleanup,
- and the worker may have published a result before terminating.
- This value specifies how long we wait for any missing results before
- raising a :exc:`@WorkerLostError` exception.
- Default is 10.0
- .. setting:: CELERYD_MAX_TASKS_PER_CHILD
- CELERYD_MAX_TASKS_PER_CHILD
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~
- Maximum number of tasks a pool worker process can execute before
- it's replaced with a new one. Default is no limit.
- .. setting:: CELERYD_TASK_TIME_LIMIT
- CELERYD_TASK_TIME_LIMIT
- ~~~~~~~~~~~~~~~~~~~~~~~
- Task hard time limit in seconds. The worker processing the task will
- be killed and replaced with a new one when this is exceeded.
- .. setting:: CELERYD_TASK_SOFT_TIME_LIMIT
- CELERYD_TASK_SOFT_TIME_LIMIT
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- Task soft time limit in seconds.
- The :exc:`~@SoftTimeLimitExceeded` exception will be
- raised when this is exceeded. The task can catch this to
- e.g. clean up before the hard time limit comes.
- Example:
- .. code-block:: python
- from celery.exceptions import SoftTimeLimitExceeded
- @celery.task()
- def mytask():
- try:
- return do_work()
- except SoftTimeLimitExceeded:
- cleanup_in_a_hurry()
- .. setting:: CELERY_STORE_ERRORS_EVEN_IF_IGNORED
- CELERY_STORE_ERRORS_EVEN_IF_IGNORED
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- If set, the worker stores all task errors in the result store even if
- :attr:`Task.ignore_result <celery.task.base.Task.ignore_result>` is on.
- .. setting:: CELERYD_STATE_DB
- CELERYD_STATE_DB
- ~~~~~~~~~~~~~~~~
- Name of the file used to stores persistent worker state (like revoked tasks).
- Can be a relative or absolute path, but be aware that the suffix `.db`
- may be appended to the file name (depending on Python version).
- Can also be set via the :option:`--statedb` argument to
- :mod:`~celery.bin.celeryd`.
- Not enabled by default.
- .. setting:: CELERYD_TIMER_PRECISION
- CELERYD_TIMER_PRECISION
- ~~~~~~~~~~~~~~~~~~~~~~~
- Set the maximum time in seconds that the ETA scheduler can sleep between
- rechecking the schedule. Default is 1 second.
- Setting this value to 1 second means the schedulers precision will
- be 1 second. If you need near millisecond precision you can set this to 0.1.
- .. _conf-error-mails:
- Error E-Mails
- -------------
- .. setting:: CELERY_SEND_TASK_ERROR_EMAILS
- CELERY_SEND_TASK_ERROR_EMAILS
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- The default value for the `Task.send_error_emails` attribute, which if
- set to :const:`True` means errors occurring during task execution will be
- sent to :setting:`ADMINS` by email.
- Disabled by default.
- .. setting:: ADMINS
- ADMINS
- ~~~~~~
- List of `(name, email_address)` tuples for the administrators that should
- receive error emails.
- .. setting:: SERVER_EMAIL
- SERVER_EMAIL
- ~~~~~~~~~~~~
- The email address this worker sends emails from.
- Default is celery@localhost.
- .. setting:: EMAIL_HOST
- EMAIL_HOST
- ~~~~~~~~~~
- The mail server to use. Default is `"localhost"`.
- .. setting:: EMAIL_HOST_USER
- EMAIL_HOST_USER
- ~~~~~~~~~~~~~~~
- User name (if required) to log on to the mail server with.
- .. setting:: EMAIL_HOST_PASSWORD
- EMAIL_HOST_PASSWORD
- ~~~~~~~~~~~~~~~~~~~
- Password (if required) to log on to the mail server with.
- .. setting:: EMAIL_PORT
- EMAIL_PORT
- ~~~~~~~~~~
- The port the mail server is listening on. Default is `25`.
- .. setting:: EMAIL_USE_SSL
- EMAIL_USE_SSL
- ~~~~~~~~~~~~~
- Use SSL when connecting to the SMTP server. Disabled by default.
- .. setting:: EMAIL_USE_TLS
- EMAIL_USE_TLS
- ~~~~~~~~~~~~~
- Use TLS when connecting to the SMTP server. Disabled by default.
- .. setting:: EMAIL_TIMEOUT
- EMAIL_TIMEOUT
- ~~~~~~~~~~~~~
- Timeout in seconds for when we give up trying to connect
- to the SMTP server when sending emails.
- The default is 2 seconds.
- .. _conf-example-error-mail-config:
- Example E-Mail configuration
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- This configuration enables the sending of error emails to
- george@vandelay.com and kramer@vandelay.com:
- .. code-block:: python
- # Enables error emails.
- CELERY_SEND_TASK_ERROR_EMAILS = True
- # Name and email addresses of recipients
- ADMINS = (
- ("George Costanza", "george@vandelay.com"),
- ("Cosmo Kramer", "kosmo@vandelay.com"),
- )
- # Email address used as sender (From field).
- SERVER_EMAIL = "no-reply@vandelay.com"
- # Mailserver configuration
- EMAIL_HOST = "mail.vandelay.com"
- EMAIL_PORT = 25
- # EMAIL_HOST_USER = "servers"
- # EMAIL_HOST_PASSWORD = "s3cr3t"
- .. _conf-events:
- Events
- ------
- .. setting:: CELERY_SEND_EVENTS
- CELERY_SEND_EVENTS
- ~~~~~~~~~~~~~~~~~~
- Send events so the worker can be monitored by tools like `celerymon`.
- .. setting:: CELERY_SEND_TASK_SENT_EVENT
- CELERY_SEND_TASK_SENT_EVENT
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.2
- If enabled, a `task-sent` event will be sent for every task so tasks can be
- tracked before they are consumed by a worker.
- Disabled by default.
- .. setting:: CELERY_EVENT_SERIALIZER
- CELERY_EVENT_SERIALIZER
- ~~~~~~~~~~~~~~~~~~~~~~~
- Message serialization format used when sending event messages.
- Default is `"json"`. See :ref:`calling-serializers`.
- .. _conf-broadcast:
- Broadcast Commands
- ------------------
- .. setting:: CELERY_BROADCAST_QUEUE
- CELERY_BROADCAST_QUEUE
- ~~~~~~~~~~~~~~~~~~~~~~
- Name prefix for the queue used when listening for broadcast messages.
- The workers host name will be appended to the prefix to create the final
- queue name.
- Default is `"celeryctl"`.
- .. setting:: CELERY_BROADCAST_EXCHANGE
- CELERY_BROADCAST_EXCHANGE
- ~~~~~~~~~~~~~~~~~~~~~~~~~
- Name of the exchange used for broadcast messages.
- Default is `"celeryctl"`.
- .. setting:: CELERY_BROADCAST_EXCHANGE_TYPE
- CELERY_BROADCAST_EXCHANGE_TYPE
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- Exchange type used for broadcast messages. Default is `"fanout"`.
- .. _conf-logging:
- Logging
- -------
- .. setting:: CELERYD_HIJACK_ROOT_LOGGER
- CELERYD_HIJACK_ROOT_LOGGER
- ~~~~~~~~~~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.2
- By default any previously configured logging options will be reset,
- because the Celery programs "hijacks" the root logger.
- If you want to customize your own logging then you can disable
- this behavior.
- .. note::
- Logging can also be customized by connecting to the
- :signal:`celery.signals.setup_logging` signal.
- .. setting:: CELERYD_LOG_COLOR
- CELERYD_LOG_COLOR
- ~~~~~~~~~~~~~~~~~
- Enables/disables colors in logging output by the Celery apps.
- By default colors are enabled if
- 1) the app is logging to a real terminal, and not a file.
- 2) the app is not running on Windows.
- .. setting:: CELERYD_LOG_FORMAT
- CELERYD_LOG_FORMAT
- ~~~~~~~~~~~~~~~~~~
- The format to use for log messages.
- Default is `[%(asctime)s: %(levelname)s/%(processName)s] %(message)s`
- See the Python :mod:`logging` module for more information about log
- formats.
- .. setting:: CELERYD_TASK_LOG_FORMAT
- CELERYD_TASK_LOG_FORMAT
- ~~~~~~~~~~~~~~~~~~~~~~~
- The format to use for log messages logged in tasks. Can be overridden using
- the :option:`--loglevel` option to :mod:`~celery.bin.celeryd`.
- Default is::
- [%(asctime)s: %(levelname)s/%(processName)s]
- [%(task_name)s(%(task_id)s)] %(message)s
- See the Python :mod:`logging` module for more information about log
- formats.
- .. setting:: CELERY_REDIRECT_STDOUTS
- CELERY_REDIRECT_STDOUTS
- ~~~~~~~~~~~~~~~~~~~~~~~
- If enabled `stdout` and `stderr` will be redirected
- to the current logger.
- Enabled by default.
- Used by :program:`celeryd` and :program:`celerybeat`.
- .. setting:: CELERY_REDIRECT_STDOUTS_LEVEL
- CELERY_REDIRECT_STDOUTS_LEVEL
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- The log level output to `stdout` and `stderr` is logged as.
- Can be one of :const:`DEBUG`, :const:`INFO`, :const:`WARNING`,
- :const:`ERROR` or :const:`CRITICAL`.
- Default is :const:`WARNING`.
- .. _conf-security:
- Security
- --------
- .. setting:: CELERY_SECURITY_KEY
- CELERY_SECURITY_KEY
- ~~~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.5
- The relative or absolute path to a file containing the private key
- used to sign messages when :ref:`message-signing` is used.
- .. setting:: CELERY_SECURITY_CERTIFICATE
- CELERY_SECURITY_CERTIFICATE
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.5
- The relative or absolute path to an X.509 certificate file
- used to sign messages when :ref:`message-signing` is used.
- .. setting:: CELERY_SECURITY_CERT_STORE
- CELERY_SECURITY_CERT_STORE
- ~~~~~~~~~~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.5
- The directory containing X.509 certificates used for
- :ref:`message-signing`. Can be a glob with wildcards,
- (for example :file:`/etc/certs/*.pem`).
- .. _conf-custom-components:
- Custom Component Classes (advanced)
- -----------------------------------
- .. setting:: CELERYD_BOOT_STEPS
- CELERYD_BOOT_STEPS
- ~~~~~~~~~~~~~~~~~~
- This setting enables you to add additional components to the worker process.
- It should be a list of module names with :class:`celery.abstract.Component`
- classes, that augments functionality in the worker.
- .. setting:: CELERYD_POOL
- CELERYD_POOL
- ~~~~~~~~~~~~
- Name of the pool class used by the worker.
- You can use a custom pool class name, or select one of
- the built-in aliases: ``processes``, ``eventlet``, ``gevent``.
- Default is ``processes``.
- .. setting:: CELERYD_AUTOSCALER
- CELERYD_AUTOSCALER
- ~~~~~~~~~~~~~~~~~~
- .. versionadded:: 2.2
- Name of the autoscaler class to use.
- Default is ``"celery.worker.autoscale.Autoscaler"``.
- .. setting:: CELERYD_AUTORELOADER
- CELERYD_AUTORELOADER
- ~~~~~~~~~~~~~~~~~~~~
- Name of the autoreloader class used by the worker to reload
- Python modules and files that have changed.
- Default is: ``"celery.worker.autoreload.Autoreloader"``.
- .. setting:: CELERYD_CONSUMER
- CELERYD_CONSUMER
- ~~~~~~~~~~~~~~~~
- Name of the consumer class used by the worker.
- Default is :class:`celery.worker.consumer.Consumer`
- .. setting:: CELERYD_MEDIATOR
- CELERYD_MEDIATOR
- ~~~~~~~~~~~~~~~~
- Name of the mediator class used by the worker.
- Default is :class:`celery.worker.controllers.Mediator`.
- .. setting:: CELERYD_TIMER
- CELERYD_TIMER
- ~~~~~~~~~~~~~~~~~~~~~
- Name of the ETA scheduler class used by the worker.
- Default is :class:`celery.utils.timer2.Timer`, or one overrided
- by the pool implementation.
- .. _conf-celerybeat:
- Periodic Task Server: celerybeat
- --------------------------------
- .. setting:: CELERYBEAT_SCHEDULE
- CELERYBEAT_SCHEDULE
- ~~~~~~~~~~~~~~~~~~~
- The periodic task schedule used by :mod:`~celery.bin.celerybeat`.
- See :ref:`beat-entries`.
- .. setting:: CELERYBEAT_SCHEDULER
- CELERYBEAT_SCHEDULER
- ~~~~~~~~~~~~~~~~~~~~
- The default scheduler class. Default is
- `"celery.beat.PersistentScheduler"`.
- Can also be set via the :option:`-S` argument to
- :mod:`~celery.bin.celerybeat`.
- .. setting:: CELERYBEAT_SCHEDULE_FILENAME
- CELERYBEAT_SCHEDULE_FILENAME
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- Name of the file used by `PersistentScheduler` to store the last run times
- of periodic tasks. Can be a relative or absolute path, but be aware that the
- suffix `.db` may be appended to the file name (depending on Python version).
- Can also be set via the :option:`--schedule` argument to
- :mod:`~celery.bin.celerybeat`.
- .. setting:: CELERYBEAT_MAX_LOOP_INTERVAL
- CELERYBEAT_MAX_LOOP_INTERVAL
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
- The maximum number of seconds :mod:`~celery.bin.celerybeat` can sleep
- between checking the schedule.
- The default for this value is scheduler specific.
- For the default celerybeat scheduler the value is 300 (5 minutes),
- but for e.g. the django-celery database scheduler it is 5 seconds
- because the schedule may be changed externally, and so it must take
- changes to the schedule into account.
- Also when running celerybeat embedded (:option:`-B`) on Jython as a thread
- the max interval is overridden and set to 1 so that it's possible
- to shut down in a timely manner.
- .. _conf-celerymon:
- Monitor Server: celerymon
- -------------------------
- .. setting:: CELERYMON_LOG_FORMAT
- CELERYMON_LOG_FORMAT
- ~~~~~~~~~~~~~~~~~~~~
- The format to use for log messages.
- Default is `[%(asctime)s: %(levelname)s/%(processName)s] %(message)s`
- See the Python :mod:`logging` module for more information about log
- formats.
|