123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609 |
- from carrot.connection import DjangoBrokerConnection
- from celery.conf import AMQP_CONNECTION_TIMEOUT
- from celery.messaging import TaskPublisher, TaskConsumer
- from celery.log import setup_logger
- from celery.result import TaskSetResult, EagerResult
- from celery.execute import apply_async, delay_task, apply
- from celery.utils import gen_unique_id, get_full_cls_name
- from datetime import timedelta
- from celery.registry import tasks
- from celery.serialization import pickle
- from celery.exceptions import MaxRetriesExceededError, RetryTaskError
- class Task(object):
- """A task that can be delayed for execution by the ``celery`` daemon.
- All subclasses of :class:`Task` must define the :meth:`run` method,
- which is the actual method the ``celery`` daemon executes.
- The :meth:`run` method can take use of the default keyword arguments,
- as listed in the :meth:`run` documentation.
- The :meth:`run` method supports both positional, and keyword arguments.
- .. attribute:: name
- *REQUIRED* All subclasses of :class:`Task` has to define the
- :attr:`name` attribute. This is the name of the task, registered
- in the task registry, and passed to :func:`delay_task`.
- .. attribute:: type
- The type of task, currently this can be ``regular``, or ``periodic``,
- however if you want a periodic task, you should subclass
- :class:`PeriodicTask` instead.
- .. attribute:: routing_key
- Override the global default ``routing_key`` for this task.
- .. attribute:: exchange
- Override the global default ``exchange`` for this task.
- .. attribute:: mandatory
- If set, the message has mandatory routing. By default the message
- is silently dropped by the broker if it can't be routed to a queue.
- However - If the message is mandatory, an exception will be raised
- instead.
- .. attribute:: immediate:
- Request immediate delivery. If the message cannot be routed to a
- task worker immediately, an exception will be raised. This is
- instead of the default behaviour, where the broker will accept and
- queue the message, but with no guarantee that the message will ever
- be consumed.
- .. attribute:: priority:
- The message priority. A number from ``0`` to ``9``.
- .. attribute:: max_retries
- Maximum number of retries before giving up.
- .. attribute:: default_retry_delay
- Defeault time in seconds before a retry of the task should be
- executed. Default is a 1 minute delay.
- .. attribute:: ignore_result
- Don't store the status and return value. This means you can't
- use the :class:`celery.result.AsyncResult` to check if the task is
- done, or get its return value. Only use if you need the performance
- and is able live without these features. Any exceptions raised will
- store the return value/status as usual.
- .. attribute:: disable_error_emails
- Disable all error e-mails for this task (only applicable if
- ``settings.SEND_CELERY_ERROR_EMAILS`` is on.)
- :raises NotImplementedError: if the :attr:`name` attribute is not set.
- The resulting class is callable, which if called will apply the
- :meth:`run` method.
- Examples
- This is a simple task just logging a message,
- >>> from celery.task import tasks, Task
- >>> class MyTask(Task):
- ...
- ... def run(self, some_arg=None, **kwargs):
- ... logger = self.get_logger(**kwargs)
- ... logger.info("Running MyTask with arg some_arg=%s" %
- ... some_arg))
- ... return 42
- ... tasks.register(MyTask)
- You can delay the task using the classmethod :meth:`delay`...
- >>> result = MyTask.delay(some_arg="foo")
- >>> result.status # after some time
- 'DONE'
- >>> result.result
- 42
- ...or using the :func:`delay_task` function, by passing the name of
- the task.
- >>> from celery.task import delay_task
- >>> result = delay_task(MyTask.name, some_arg="foo")
- """
- name = None
- type = "regular"
- exchange = None
- routing_key = None
- immediate = False
- mandatory = False
- priority = None
- ignore_result = False
- disable_error_emails = False
- max_retries = 3
- default_retry_delay = 3 * 60
- MaxRetriesExceededError = MaxRetriesExceededError
- def __init__(self):
- if not self.__class__.name:
- self.__class__.name = get_full_cls_name(self.__class__)
- def __call__(self, *args, **kwargs):
- return self.run(*args, **kwargs)
- def run(self, *args, **kwargs):
- """The body of the task executed by the worker.
- The following standard keyword arguments are reserved and is passed
- by the worker if the function/method supports them:
- * task_id
- Unique id of the currently executing task.
- * task_name
- Name of the currently executing task (same as :attr:`name`)
- * task_retries
- How many times the current task has been retried
- (an integer starting at ``0``).
- * logfile
- Name of the worker log file.
- * loglevel
- The current loglevel, an integer mapping to one of the
- following values: ``logging.DEBUG``, ``logging.INFO``,
- ``logging.ERROR``, ``logging.CRITICAL``, ``logging.WARNING``,
- ``logging.FATAL``.
- Additional standard keyword arguments may be added in the future.
- To take these default arguments, the task can either list the ones
- it wants explicitly or just take an arbitrary list of keyword
- arguments (\*\*kwargs).
- Example using an explicit list of default arguments to take:
- .. code-block:: python
- def run(self, x, y, logfile=None, loglevel=None):
- self.get_logger(loglevel=loglevel, logfile=logfile)
- return x * y
- Example taking all default keyword arguments, and any extra arguments
- passed on by the caller:
- .. code-block:: python
- def run(self, x, y, **kwargs): # CORRECT!
- logger = self.get_logger(**kwargs)
- adjust = kwargs.get("adjust", 0)
- return x * y - adjust
- """
- raise NotImplementedError("Tasks must define a run method.")
- def get_logger(self, **kwargs):
- """Get process-aware logger object.
- See :func:`celery.log.setup_logger`.
- """
- logfile = kwargs.get("logfile")
- loglevel = kwargs.get("loglevel")
- return setup_logger(loglevel=loglevel, logfile=logfile)
- def get_publisher(self, connect_timeout=AMQP_CONNECTION_TIMEOUT):
- """Get a celery task message publisher.
- :rtype: :class:`celery.messaging.TaskPublisher`.
- Please be sure to close the AMQP connection when you're done
- with this object, i.e.:
- >>> publisher = self.get_publisher()
- >>> # do something with publisher
- >>> publisher.connection.close()
- """
- connection = DjangoBrokerConnection(connect_timeout=connect_timeout)
- return TaskPublisher(connection=connection,
- exchange=self.exchange,
- routing_key=self.routing_key)
- def get_consumer(self, connect_timeout=AMQP_CONNECTION_TIMEOUT):
- """Get a celery task message consumer.
- :rtype: :class:`celery.messaging.TaskConsumer`.
- Please be sure to close the AMQP connection when you're done
- with this object. i.e.:
- >>> consumer = self.get_consumer()
- >>> # do something with consumer
- >>> consumer.connection.close()
- """
- connection = DjangoBrokerConnection(connect_timeout=connect_timeout)
- return TaskConsumer(connection=connection, exchange=self.exchange,
- routing_key=self.routing_key)
- @classmethod
- def delay(cls, *args, **kwargs):
- """Delay this task for execution by the ``celery`` daemon(s).
- :param \*args: positional arguments passed on to the task.
- :param \*\*kwargs: keyword arguments passed on to the task.
- :rtype: :class:`celery.result.AsyncResult`
- See :func:`celery.execute.delay_task`.
- """
- return apply_async(cls, args, kwargs)
- @classmethod
- def apply_async(cls, args=None, kwargs=None, **options):
- """Delay this task for execution by the ``celery`` daemon(s).
- :param args: positional arguments passed on to the task.
- :param kwargs: keyword arguments passed on to the task.
- :keyword \*\*options: Any keyword arguments to pass on to
- :func:`celery.execute.apply_async`.
- See :func:`celery.execute.apply_async` for more information.
- :rtype: :class:`celery.result.AsyncResult`
- """
- return apply_async(cls, args, kwargs, **options)
- def retry(self, args, kwargs, exc=None, throw=True, **options):
- """Retry the task.
- :param args: Positional arguments to retry with.
- :param kwargs: Keyword arguments to retry with.
- :keyword exc: Optional exception to raise instead of
- :exc:`MaxRestartsExceededError` when the max restart limit has
- been exceeded.
- :keyword throw: Do not raise the
- :exc:`celery.exceptions.RetryTaskError` exception,
- that tells the worker that the task is to be retried.
- :keyword countdown: Time in seconds to delay the retry for.
- :keyword eta: Explicit time and date to run the retry at (must be a
- :class:`datetime.datetime` instance).
- :keyword \*\*options: Any extra options to pass on to
- meth:`apply_async`. See :func:`celery.execute.apply_async`.
- :raises celery.exceptions.RetryTaskError: To tell the worker that the
- task has been re-sent for retry. This always happens except if
- the ``throw`` keyword argument has been explicitly set
- to ``False``.
- Example
- >>> class TwitterPostStatusTask(Task):
- ...
- ... def run(self, username, password, message, **kwargs):
- ... twitter = Twitter(username, password)
- ... try:
- ... twitter.post_status(message)
- ... except twitter.FailWhale, exc:
- ... # Retry in 5 minutes.
- ... self.retry([username, password, message], kwargs,
- ... countdown=60 * 5, exc=exc)
- """
- options["retries"] = kwargs.pop("task_retries", 0) + 1
- options["task_id"] = kwargs.pop("task_id", None)
- options["countdown"] = options.get("countdown",
- self.default_retry_delay)
- max_exc = exc or self.MaxRetriesExceededError(
- "Can't retry %s[%s] args:%s kwargs:%s" % (
- self.name, options["task_id"], args, kwargs))
- if options["retries"] > self.max_retries:
- raise max_exc
- # If task was executed eagerly using apply(),
- # then the retry must also be executed eagerly.
- if kwargs.get("task_is_eager", False):
- result = self.apply(args=args, kwargs=kwargs, **options)
- if isinstance(result, EagerResult):
- # get() propogates any exceptions.
- return result.get()
- return result
- self.apply_async(args=args, kwargs=kwargs, **options)
- if throw:
- message = "Retry in %d seconds." % options["countdown"]
- raise RetryTaskError(message, exc)
- def on_retry(self, exc):
- """Retry handler.
- This is run by the worker when the task is to be retried.
- :param exc: The exception sent to :meth:`retry`.
- """
- pass
- def on_failure(self, exc):
- """Error handler.
- This is run by the worker when the task fails.
- :param exc: The exception raised by the task.
- """
- pass
- def on_success(self, retval):
- """Success handler.
- This is run by the worker when the task executed successfully.
- :param retval: The return value of the task.
-
- """
- pass
- @classmethod
- def apply(cls, args=None, kwargs=None, **options):
- """Execute this task at once, by blocking until the task
- has finished executing.
- :param args: positional arguments passed on to the task.
- :param kwargs: keyword arguments passed on to the task.
- :rtype: :class:`celery.result.EagerResult`
- See :func:`celery.execute.apply`.
- """
- return apply(cls, args, kwargs, **options)
- class ExecuteRemoteTask(Task):
- """Execute an arbitrary function or object.
- *Note* You probably want :func:`execute_remote` instead, which this
- is an internal component of.
- The object must be pickleable, so you can't use lambdas or functions
- defined in the REPL (that is the python shell, or ``ipython``).
- """
- name = "celery.execute_remote"
- def run(self, ser_callable, fargs, fkwargs, **kwargs):
- """
- :param ser_callable: A pickled function or callable object.
- :param fargs: Positional arguments to apply to the function.
- :param fkwargs: Keyword arguments to apply to the function.
- """
- callable_ = pickle.loads(ser_callable)
- return callable_(*fargs, **fkwargs)
- tasks.register(ExecuteRemoteTask)
- class AsynchronousMapTask(Task):
- """Task used internally by :func:`dmap_async` and
- :meth:`TaskSet.map_async`. """
- name = "celery.map_async"
- def run(self, serfunc, args, **kwargs):
- """The method run by ``celeryd``."""
- timeout = kwargs.get("timeout")
- return TaskSet.map(pickle.loads(serfunc), args, timeout=timeout)
- tasks.register(AsynchronousMapTask)
- class TaskSet(object):
- """A task containing several subtasks, making it possible
- to track how many, or when all of the tasks has been completed.
- :param task: The task class or name.
- Can either be a fully qualified task name, or a task class.
- :param args: A list of args, kwargs pairs.
- e.g. ``[[args1, kwargs1], [args2, kwargs2], ..., [argsN, kwargsN]]``
- .. attribute:: task_name
- The name of the task.
- .. attribute:: arguments
- The arguments, as passed to the task set constructor.
- .. attribute:: total
- Total number of tasks in this task set.
- Example
- >>> from djangofeeds.tasks import RefreshFeedTask
- >>> taskset = TaskSet(RefreshFeedTask, args=[
- ... [], {"feed_url": "http://cnn.com/rss"},
- ... [], {"feed_url": "http://bbc.com/rss"},
- ... [], {"feed_url": "http://xkcd.com/rss"}])
- >>> taskset_result = taskset.run()
- >>> list_of_return_values = taskset.join()
- """
- def __init__(self, task, args):
- try:
- task_name = task.name
- task_obj = task
- except AttributeError:
- task_name = task
- task_obj = tasks[task_name]
- self.task = task_obj
- self.task_name = task_name
- self.arguments = args
- self.total = len(args)
- def run(self, connect_timeout=AMQP_CONNECTION_TIMEOUT):
- """Run all tasks in the taskset.
- :returns: A :class:`celery.result.TaskSetResult` instance.
- Example
- >>> ts = TaskSet(RefreshFeedTask, [
- ... ["http://foo.com/rss", {}],
- ... ["http://bar.com/rss", {}],
- ... )
- >>> result = ts.run()
- >>> result.taskset_id
- "d2c9b261-8eff-4bfb-8459-1e1b72063514"
- >>> result.subtask_ids
- ["b4996460-d959-49c8-aeb9-39c530dcde25",
- "598d2d18-ab86-45ca-8b4f-0779f5d6a3cb"]
- >>> result.waiting()
- True
- >>> time.sleep(10)
- >>> result.ready()
- True
- >>> result.successful()
- True
- >>> result.failed()
- False
- >>> result.join()
- [True, True]
- """
- taskset_id = gen_unique_id()
- from celery.conf import ALWAYS_EAGER
- if ALWAYS_EAGER:
- subtasks = [apply(self.task, args, kwargs)
- for args, kwargs in self.arguments]
- return TaskSetResult(taskset_id, subtasks)
- conn = DjangoBrokerConnection(connect_timeout=connect_timeout)
- publisher = TaskPublisher(connection=conn,
- exchange=self.task.exchange)
- subtasks = [apply_async(self.task, args, kwargs,
- taskset_id=taskset_id, publisher=publisher)
- for args, kwargs in self.arguments]
- publisher.close()
- conn.close()
- return TaskSetResult(taskset_id, subtasks)
- def join(self, timeout=None):
- """Gather the results for all of the tasks in the taskset,
- and return a list with them ordered by the order of which they
- were called.
- :keyword timeout: The time in seconds, how long
- it will wait for results, before the operation times out.
- :raises TimeoutError: if ``timeout`` is not ``None``
- and the operation takes longer than ``timeout`` seconds.
- If any of the tasks raises an exception, the exception
- will be reraised by :meth:`join`.
- :returns: list of return values for all tasks in the taskset.
- """
- return self.run().join(timeout=timeout)
- @classmethod
- def remote_execute(cls, func, args):
- """Apply ``args`` to function by distributing the args to the
- celery server(s)."""
- pickled = pickle.dumps(func)
- arguments = [[[pickled, arg, {}], {}] for arg in args]
- return cls(ExecuteRemoteTask, arguments)
- @classmethod
- def map(cls, func, args, timeout=None):
- """Distribute processing of the arguments and collect the results."""
- remote_task = cls.remote_execute(func, args)
- return remote_task.join(timeout=timeout)
- @classmethod
- def map_async(cls, func, args, timeout=None):
- """Distribute processing of the arguments and collect the results
- asynchronously.
- :returns: :class:`celery.result.AsyncResult` instance.
- """
- serfunc = pickle.dumps(func)
- return AsynchronousMapTask.delay(serfunc, args, timeout=timeout)
- class PeriodicTask(Task):
- """A periodic task is a task that behaves like a :manpage:`cron` job.
- .. attribute:: run_every
- *REQUIRED* Defines how often the task is run (its interval),
- it can be either a :class:`datetime.timedelta` object or an
- integer specifying the time in seconds.
- :raises NotImplementedError: if the :attr:`run_every` attribute is
- not defined.
- You have to register the periodic task in the task registry.
- Example
- >>> from celery.task import tasks, PeriodicTask
- >>> from datetime import timedelta
- >>> class MyPeriodicTask(PeriodicTask):
- ... name = "my_periodic_task"
- ... run_every = timedelta(seconds=30)
- ...
- ... def run(self, **kwargs):
- ... logger = self.get_logger(**kwargs)
- ... logger.info("Running MyPeriodicTask")
- >>> tasks.register(MyPeriodicTask)
- """
- run_every = timedelta(days=1)
- type = "periodic"
- def __init__(self):
- if not self.run_every:
- raise NotImplementedError(
- "Periodic tasks must have a run_every attribute")
- # If run_every is a integer, convert it to timedelta seconds.
- # Operate on the original class attribute so anyone accessing
- # it directly gets the right value.
- if isinstance(self.__class__.run_every, int):
- self.__class__.run_every = timedelta(seconds=self.run_every)
- super(PeriodicTask, self).__init__()
|