123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676 |
- import sys
- import warnings
- from celery import conf
- from celery.backends import default_backend
- from celery.exceptions import MaxRetriesExceededError, RetryTaskError
- from celery.execute import apply_async, apply
- from celery.log import setup_task_logger
- from celery.messaging import TaskPublisher, TaskConsumer
- from celery.messaging import establish_connection as _establish_connection
- from celery.registry import tasks
- from celery.result import BaseAsyncResult, EagerResult
- from celery.schedules import maybe_schedule
- from celery.utils.timeutils import timedelta_seconds
- from celery.task.sets import TaskSet, subtask
- PERIODIC_DEPRECATION_TEXT = """\
- Periodic task classes has been deprecated and will be removed
- in celery v3.0.
- Please use the CELERYBEAT_SCHEDULE setting instead:
- CELERYBEAT_SCHEDULE = {
- name: dict(task=task_name, schedule=run_every,
- args=(), kwargs={}, options={}, relative=False)
- }
- """
- def _unpickle_task(name):
- return tasks[name]
- class TaskType(type):
- """Metaclass for tasks.
- Automatically registers the task in the task registry, except
- if the ``abstract`` attribute is set.
- If no ``name`` attribute is provided, the name is automatically
- set to the name of the module it was defined in, and the class name.
- """
- def __new__(cls, name, bases, attrs):
- super_new = super(TaskType, cls).__new__
- task_module = attrs["__module__"]
- # Abstract class, remove the abstract attribute so
- # any class inheriting from this won't be abstract by default.
- if attrs.pop("abstract", None) or not attrs.get("autoregister", True):
- return super_new(cls, name, bases, attrs)
- # Automatically generate missing name.
- if not attrs.get("name"):
- task_module = sys.modules[task_module]
- task_name = ".".join([task_module.__name__, name])
- attrs["name"] = task_name
- # Because of the way import happens (recursively)
- # we may or may not be the first time the task tries to register
- # with the framework. There should only be one class for each task
- # name, so we always return the registered version.
- task_name = attrs["name"]
- if task_name not in tasks:
- task_cls = super_new(cls, name, bases, attrs)
- tasks.register(task_cls)
- return tasks[task_name].__class__
- class Task(object):
- """A celery task.
- All subclasses of :class:`Task` must define the :meth:`run` method,
- which is the actual method the ``celery`` daemon executes.
- The :meth:`run` method can take use of the default keyword arguments,
- as listed in the :meth:`run` documentation.
- The resulting class is callable, which if called will apply the
- :meth:`run` method.
- .. attribute:: name
- Name of the task.
- .. attribute:: abstract
- If :const:`True` the task is an abstract base class.
- .. attribute:: type
- The type of task, currently this can be ``regular``, or ``periodic``,
- however if you want a periodic task, you should subclass
- :class:`PeriodicTask` instead.
- .. attribute:: queue
- Select a destination queue for this task. The queue needs to exist
- in :setting:`CELERY_QUEUES`. The ``routing_key``, ``exchange`` and
- ``exchange_type`` attributes will be ignored if this is set.
- .. attribute:: routing_key
- Override the global default ``routing_key`` for this task.
- .. attribute:: exchange
- Override the global default ``exchange`` for this task.
- .. attribute:: exchange_type
- Override the global default exchange type for this task.
- .. attribute:: delivery_mode
- Override the global default delivery mode for this task.
- By default this is set to ``2`` (persistent). You can change this
- to ``1`` to get non-persistent behavior, which means the messages
- are lost if the broker is restarted.
- .. attribute:: mandatory
- Mandatory message routing. An exception will be raised if the task
- can't be routed to a queue.
- .. attribute:: immediate:
- Request immediate delivery. An exception will be raised if the task
- can't be routed to a worker immediately.
- .. attribute:: priority:
- The message priority. A number from ``0`` to ``9``, where ``0`` is the
- highest. Note that RabbitMQ doesn't support priorities yet.
- .. attribute:: max_retries
- Maximum number of retries before giving up.
- If set to :const:`None`, it will never stop retrying.
- .. attribute:: default_retry_delay
- Default time in seconds before a retry of the task should be
- executed. Default is a 3 minute delay.
- .. attribute:: rate_limit
- Set the rate limit for this task type, Examples: :const:`None` (no
- rate limit), ``"100/s"`` (hundred tasks a second), ``"100/m"``
- (hundred tasks a minute), ``"100/h"`` (hundred tasks an hour)
- .. attribute:: ignore_result
- Don't store the return value of this task.
- .. attribute:: store_errors_even_if_ignored
- If true, errors will be stored even if the task is configured
- to ignore results.
- .. attribute:: send_error_emails
- If true, an e-mail will be sent to the admins whenever
- a task of this type raises an exception.
- .. attribute:: error_whitelist
- List of exception types to send error e-mails for.
- .. attribute:: serializer
- The name of a serializer that has been registered with
- :mod:`carrot.serialization.registry`. Example: ``"json"``.
- .. attribute:: backend
- The result store backend used for this task.
- .. attribute:: autoregister
- If :const:`True` the task is automatically registered in the task
- registry, which is the default behaviour.
- .. attribute:: track_started
- If :const:`True` the task will report its status as "started"
- when the task is executed by a worker.
- The default value is ``False`` as the normal behaviour is to not
- report that level of granularity. Tasks are either pending, finished,
- or waiting to be retried. Having a "started" status can be useful for
- when there are long running tasks and there is a need to report which
- task is currently running.
- The global default can be overridden by the
- :setting:`CELERY_TRACK_STARTED` setting.
- .. attribute:: acks_late
- If set to :const:`True` messages for this task will be acknowledged
- **after** the task has been executed, not *just before*, which is
- the default behavior.
- Note that this means the task may be executed twice if the worker
- crashes in the middle of execution, which may be acceptable for some
- applications.
- The global default can be overriden by the :setting:`CELERY_ACKS_LATE`
- setting.
- """
- __metaclass__ = TaskType
- name = None
- abstract = True
- autoregister = True
- type = "regular"
- queue = None
- routing_key = None
- exchange = None
- exchange_type = conf.DEFAULT_EXCHANGE_TYPE
- delivery_mode = conf.DEFAULT_DELIVERY_MODE
- immediate = False
- mandatory = False
- priority = None
- ignore_result = conf.IGNORE_RESULT
- store_errors_even_if_ignored = conf.STORE_ERRORS_EVEN_IF_IGNORED
- send_error_emails = conf.CELERY_SEND_TASK_ERROR_EMAILS
- error_whitelist = conf.CELERY_TASK_ERROR_WHITELIST
- disable_error_emails = False # FIXME
- max_retries = 5
- default_retry_delay = 3 * 60
- serializer = conf.TASK_SERIALIZER
- rate_limit = conf.DEFAULT_RATE_LIMIT
- backend = default_backend
- track_started = conf.TRACK_STARTED
- acks_late = conf.ACKS_LATE
- MaxRetriesExceededError = MaxRetriesExceededError
- def __call__(self, *args, **kwargs):
- return self.run(*args, **kwargs)
- def __reduce__(self):
- return (_unpickle_task, (self.name, ), None)
- def run(self, *args, **kwargs):
- """The body of the task executed by the worker.
- The following standard keyword arguments are reserved and is passed
- by the worker if the function/method supports them:
- * task_id
- * task_name
- * task_retries
- * task_is_eager
- * logfile
- * loglevel
- * delivery_info
- Additional standard keyword arguments may be added in the future.
- To take these default arguments, the task can either list the ones
- it wants explicitly or just take an arbitrary list of keyword
- arguments (\*\*kwargs).
- """
- raise NotImplementedError("Tasks must define the run method.")
- @classmethod
- def get_logger(self, loglevel=None, logfile=None, **kwargs):
- """Get task-aware logger object.
- See :func:`celery.log.setup_task_logger`.
- """
- return setup_task_logger(loglevel=loglevel, logfile=logfile,
- task_kwargs=kwargs)
- @classmethod
- def establish_connection(self,
- connect_timeout=conf.BROKER_CONNECTION_TIMEOUT):
- """Establish a connection to the message broker."""
- return _establish_connection(connect_timeout=connect_timeout)
- @classmethod
- def get_publisher(self, connection=None, exchange=None,
- connect_timeout=conf.BROKER_CONNECTION_TIMEOUT,
- exchange_type=None):
- """Get a celery task message publisher.
- :rtype :class:`celery.messaging.TaskPublisher`:
- Please be sure to close the AMQP connection when you're done
- with this object, i.e.:
- >>> publisher = self.get_publisher()
- >>> # do something with publisher
- >>> publisher.connection.close()
- """
- if exchange is None:
- exchange = self.exchange
- if exchange_type is None:
- exchange_type = self.exchange_type
- connection = connection or self.establish_connection(connect_timeout)
- return TaskPublisher(connection=connection,
- exchange=exchange,
- exchange_type=exchange_type,
- routing_key=self.routing_key)
- @classmethod
- def get_consumer(self, connection=None,
- connect_timeout=conf.BROKER_CONNECTION_TIMEOUT):
- """Get a celery task message consumer.
- :rtype :class:`celery.messaging.TaskConsumer`:
- Please be sure to close the AMQP connection when you're done
- with this object. i.e.:
- >>> consumer = self.get_consumer()
- >>> # do something with consumer
- >>> consumer.connection.close()
- """
- connection = connection or self.establish_connection(connect_timeout)
- return TaskConsumer(connection=connection, exchange=self.exchange,
- routing_key=self.routing_key)
- @classmethod
- def delay(self, *args, **kwargs):
- """Shortcut to :meth:`apply_async`, with star arguments,
- but doesn't support the extra options.
- :param \*args: positional arguments passed on to the task.
- :param \*\*kwargs: keyword arguments passed on to the task.
- :returns :class:`celery.result.AsyncResult`:
- """
- return self.apply_async(args, kwargs)
- @classmethod
- def apply_async(self, args=None, kwargs=None, **options):
- """Delay this task for execution by the ``celery`` daemon(s).
- :param args: positional arguments passed on to the task.
- :param kwargs: keyword arguments passed on to the task.
- :keyword \*\*options: Any keyword arguments to pass on to
- :func:`celery.execute.apply_async`.
- See :func:`celery.execute.apply_async` for more information.
- :returns :class:`celery.result.AsyncResult`:
- """
- return apply_async(self, args, kwargs, **options)
- @classmethod
- def retry(self, args=None, kwargs=None, exc=None, throw=True, **options):
- """Retry the task.
- :param args: Positional arguments to retry with.
- :param kwargs: Keyword arguments to retry with.
- :keyword exc: Optional exception to raise instead of
- :exc:`~celery.exceptions.MaxRetriesExceededError` when the max
- restart limit has been exceeded.
- :keyword countdown: Time in seconds to delay the retry for.
- :keyword eta: Explicit time and date to run the retry at (must be a
- :class:`datetime.datetime` instance).
- :keyword \*\*options: Any extra options to pass on to
- meth:`apply_async`. See :func:`celery.execute.apply_async`.
- :keyword throw: If this is ``False``, do not raise the
- :exc:`~celery.exceptions.RetryTaskError` exception,
- that tells the worker to mark the task as being retried.
- Note that this means the task will be marked as failed
- if the task raises an exception, or successful if it
- returns.
- :raises celery.exceptions.RetryTaskError: To tell the worker that the
- task has been re-sent for retry. This always happens, unless
- the ``throw`` keyword argument has been explicitly set
- to ``False``, and is considered normal operation.
- Example
- >>> class TwitterPostStatusTask(Task):
- ...
- ... def run(self, username, password, message, **kwargs):
- ... twitter = Twitter(username, password)
- ... try:
- ... twitter.post_status(message)
- ... except twitter.FailWhale, exc:
- ... # Retry in 5 minutes.
- ... self.retry([username, password, message], kwargs,
- ... countdown=60 * 5, exc=exc)
- """
- if not kwargs:
- raise TypeError(
- "kwargs argument to retries can't be empty. "
- "Task must accept **kwargs, see http://bit.ly/cAx3Bg")
- delivery_info = kwargs.pop("delivery_info", {})
- options.setdefault("exchange", delivery_info.get("exchange"))
- options.setdefault("routing_key", delivery_info.get("routing_key"))
- options["retries"] = kwargs.pop("task_retries", 0) + 1
- options["task_id"] = kwargs.pop("task_id", None)
- options["countdown"] = options.get("countdown",
- self.default_retry_delay)
- max_exc = exc or self.MaxRetriesExceededError(
- "Can't retry %s[%s] args:%s kwargs:%s" % (
- self.name, options["task_id"], args, kwargs))
- max_retries = self.max_retries
- if max_retries is not None and options["retries"] > max_retries:
- raise max_exc
- # If task was executed eagerly using apply(),
- # then the retry must also be executed eagerly.
- if kwargs.get("task_is_eager", False):
- result = self.apply(args=args, kwargs=kwargs, **options)
- if isinstance(result, EagerResult):
- return result.get() # propogates exceptions.
- return result
- self.apply_async(args=args, kwargs=kwargs, **options)
- if throw:
- message = "Retry in %d seconds." % options["countdown"]
- raise RetryTaskError(message, exc)
- @classmethod
- def apply(self, args=None, kwargs=None, **options):
- """Execute this task locally, by blocking until the task
- has finished executing.
- :param args: positional arguments passed on to the task.
- :param kwargs: keyword arguments passed on to the task.
- :keyword throw: Re-raise task exceptions. Defaults to
- the :setting:`CELERY_EAGER_PROPAGATES_EXCEPTIONS` setting.
- :rtype :class:`celery.result.EagerResult`:
- See :func:`celery.execute.apply`.
- """
- return apply(self, args, kwargs, **options)
- @classmethod
- def AsyncResult(self, task_id):
- """Get AsyncResult instance for this kind of task.
- :param task_id: Task id to get result for.
- """
- return BaseAsyncResult(task_id, backend=self.backend)
- def update_state(self, task_id, state, meta=None):
- """Update task state.
- :param task_id: Id of the task to update.
- :param state: New state (:class:`str`).
- :param meta: State metadata (:class:`dict`).
- """
- self.backend.store_result(task_id, meta, state)
- def on_retry(self, exc, task_id, args, kwargs, einfo=None):
- """Retry handler.
- This is run by the worker when the task is to be retried.
- :param exc: The exception sent to :meth:`retry`.
- :param task_id: Unique id of the retried task.
- :param args: Original arguments for the retried task.
- :param kwargs: Original keyword arguments for the retried task.
- :keyword einfo: :class:`~celery.datastructures.ExceptionInfo` instance,
- containing the traceback.
- The return value of this handler is ignored.
- """
- pass
- def after_return(self, status, retval, task_id, args, kwargs, einfo=None):
- """Handler called after the task returns.
- :param status: Current task state.
- :param retval: Task return value/exception.
- :param task_id: Unique id of the task.
- :param args: Original arguments for the task that failed.
- :param kwargs: Original keyword arguments for the task that failed.
- :keyword einfo: :class:`~celery.datastructures.ExceptionInfo` instance,
- containing the traceback (if any).
- The return value of this handler is ignored.
- """
- pass
- def on_failure(self, exc, task_id, args, kwargs, einfo=None):
- """Error handler.
- This is run by the worker when the task fails.
- :param exc: The exception raised by the task.
- :param task_id: Unique id of the failed task.
- :param args: Original arguments for the task that failed.
- :param kwargs: Original keyword arguments for the task that failed.
- :keyword einfo: :class:`~celery.datastructures.ExceptionInfo` instance,
- containing the traceback.
- The return value of this handler is ignored.
- """
- pass
- def on_success(self, retval, task_id, args, kwargs):
- """Success handler.
- Run by the worker if the task executes successfully.
- :param retval: The return value of the task.
- :param task_id: Unique id of the executed task.
- :param args: Original arguments for the executed task.
- :param kwargs: Original keyword arguments for the executed task.
- The return value of this handler is ignored.
- """
- pass
- def execute(self, wrapper, pool, loglevel, logfile):
- """The method the worker calls to execute the task.
- :param wrapper: A :class:`~celery.worker.job.TaskRequest`.
- :param pool: A task pool.
- :param loglevel: Current loglevel.
- :param logfile: Name of the currently used logfile.
- """
- wrapper.execute_using_pool(pool, loglevel, logfile)
- def __repr__(self):
- """repr(task)"""
- try:
- kind = self.__class__.mro()[1].__name__
- except (AttributeError, IndexError): # pragma: no cover
- kind = "%s(Task)" % self.__class__.__name__
- return "<%s: %s (%s)>" % (kind, self.name, self.type)
- @classmethod
- def subtask(cls, *args, **kwargs):
- """Returns a :class:`~celery.task.sets.subtask` object for
- this task that wraps arguments and execution options
- for a single task invocation."""
- return subtask(cls, *args, **kwargs)
- @property
- def __name__(self):
- return self.__class__.__name__
- class PeriodicTask(Task):
- """A periodic task is a task that behaves like a :manpage:`cron` job.
- Results of periodic tasks are not stored by default.
- .. attribute:: run_every
- *REQUIRED* Defines how often the task is run (its interval),
- it can be a :class:`~datetime.timedelta` object, a
- :class:`~celery.task.schedules.crontab` object or an integer
- specifying the time in seconds.
- .. attribute:: relative
- If set to :const:`True`, run times are relative to the time when the
- server was started. This was the previous behaviour, periodic tasks
- are now scheduled by the clock.
- :raises NotImplementedError: if the :attr:`run_every` attribute is
- not defined.
- Example
- >>> from celery.task import tasks, PeriodicTask
- >>> from datetime import timedelta
- >>> class EveryThirtySecondsTask(PeriodicTask):
- ... run_every = timedelta(seconds=30)
- ...
- ... def run(self, **kwargs):
- ... logger = self.get_logger(**kwargs)
- ... logger.info("Execute every 30 seconds")
- >>> from celery.task import PeriodicTask
- >>> from celery.task.schedules import crontab
- >>> class EveryMondayMorningTask(PeriodicTask):
- ... run_every = crontab(hour=7, minute=30, day_of_week=1)
- ...
- ... def run(self, **kwargs):
- ... logger = self.get_logger(**kwargs)
- ... logger.info("Execute every Monday at 7:30AM.")
- >>> class EveryMorningTask(PeriodicTask):
- ... run_every = crontab(hours=7, minute=30)
- ...
- ... def run(self, **kwargs):
- ... logger = self.get_logger(**kwargs)
- ... logger.info("Execute every day at 7:30AM.")
- >>> class EveryQuarterPastTheHourTask(PeriodicTask):
- ... run_every = crontab(minute=15)
- ...
- ... def run(self, **kwargs):
- ... logger = self.get_logger(**kwargs)
- ... logger.info("Execute every 0:15 past the hour every day.")
- """
- abstract = True
- ignore_result = True
- type = "periodic"
- relative = False
- def __init__(self):
- if not hasattr(self, "run_every"):
- raise NotImplementedError(
- "Periodic tasks must have a run_every attribute")
- self.run_every = maybe_schedule(self.run_every, self.relative)
- # Periodic task classes is pending deprecation.
- warnings.warn(PendingDeprecationWarning(PERIODIC_DEPRECATION_TEXT))
- # For backward compatibility, add the periodic task to the
- # configuration schedule instead.
- conf.CELERYBEAT_SCHEDULE[self.name] = {
- "task": self.name,
- "schedule": self.run_every,
- "args": (),
- "kwargs": {},
- "options": {},
- "relative": self.relative,
- }
- super(PeriodicTask, self).__init__()
- def timedelta_seconds(self, delta):
- """Convert :class:`~datetime.timedelta` to seconds.
- Doesn't account for negative timedeltas.
- """
- return timedelta_seconds(delta)
- def is_due(self, last_run_at):
- """Returns tuple of two items ``(is_due, next_time_to_run)``,
- where next time to run is in seconds.
- See :meth:`celery.schedules.schedule.is_due` for more information.
- """
- return self.run_every.is_due(last_run_at)
- def remaining_estimate(self, last_run_at):
- """Returns when the periodic task should run next as a timedelta."""
- return self.run_every.remaining_estimate(last_run_at)
|