# -*- coding: utf-8 -*- """ celery.app.builtins ~~~~~~~~~~~~~~~~~~~ Built-in tasks that are always available in all app instances. E.g. chord, group and xmap. """ from __future__ import absolute_import from collections import deque from future_builtins import map, zip from itertools import starmap from celery._state import get_current_worker_task from celery.utils import uuid #: global list of functions defining tasks that should be #: added to all apps. _shared_tasks = [] def shared_task(constructor): """Decorator that specifies that the decorated function is a function that generates a built-in task. The function will then be called for every new app instance created (lazily, so more exactly when the task registry for that app is needed). """ _shared_tasks.append(constructor) return constructor def load_shared_tasks(app): """Loads the built-in tasks for an app instance.""" for constructor in _shared_tasks: constructor(app) @shared_task def add_backend_cleanup_task(app): """The backend cleanup task can be used to clean up the default result backend. This task is also added do the periodic task schedule so that it is run every day at midnight, but :program:`celerybeat` must be running for this to be effective. Note that not all backends do anything for this, what needs to be done at cleanup is up to each backend, and some backends may even clean up in realtime so that a periodic cleanup is not necessary. """ @app.task(name='celery.backend_cleanup') def backend_cleanup(): app.backend.cleanup() return backend_cleanup @shared_task def add_unlock_chord_task(app): """The unlock chord task is used by result backends that doesn't have native chord support. It creates a task chain polling the header for completion. """ from celery.canvas import subtask from celery import result as _res @app.task(name='celery.chord_unlock', max_retries=None) def unlock_chord(group_id, callback, interval=1, propagate=False, max_retries=None, result=None): AR = _res.AsyncResult result = _res.GroupResult(group_id, [AR(r) for r in result]) j = result.join_native if result.supports_native_join else result.join if result.ready(): subtask(callback).delay(j(propagate=propagate)) else: unlock_chord.retry(countdown=interval, max_retries=max_retries) return unlock_chord @shared_task def add_map_task(app): from celery.canvas import subtask @app.task(name='celery.map') def xmap(task, it): task = subtask(task).type return list(map(task, it)) return xmap @shared_task def add_starmap_task(app): from celery.canvas import subtask @app.task(name='celery.starmap') def xstarmap(task, it): task = subtask(task).type return list(starmap(task, it)) return xstarmap @shared_task def add_chunk_task(app): from celery.canvas import chunks as _chunks @app.task(name='celery.chunks') def chunks(task, it, n): return _chunks.apply_chunks(task, it, n) return chunks @shared_task def add_group_task(app): _app = app from celery.canvas import maybe_subtask, subtask from celery.result import from_serializable class Group(app.Task): app = _app name = 'celery.group' accept_magic_kwargs = False def run(self, tasks, result, group_id, partial_args): app = self.app result = from_serializable(result) # any partial args are added to all tasks in the group taskit = (subtask(task).clone(partial_args) for i, task in enumerate(tasks)) if self.request.is_eager or app.conf.CELERY_ALWAYS_EAGER: return app.GroupResult(result.id, [task.apply(group_id=group_id) for task in taskit]) with app.producer_or_acquire() as pub: [task.apply_async(group_id=group_id, publisher=pub, add_to_parent=False) for task in taskit] parent = get_current_worker_task() if parent: parent.request.children.append(result) return result def prepare(self, options, tasks, args, **kwargs): options['group_id'] = group_id = \ options.setdefault('task_id', uuid()) def prepare_member(task): task = maybe_subtask(task) opts = task.options opts['group_id'] = group_id try: tid = opts['task_id'] except KeyError: tid = opts['task_id'] = uuid() return task, self.AsyncResult(tid) tasks, res = list(zip(*[prepare_member(task) for task in tasks])) return (tasks, self.app.GroupResult(group_id, res), group_id, args) def apply_async(self, partial_args=(), kwargs={}, **options): if self.app.conf.CELERY_ALWAYS_EAGER: return self.apply(partial_args, kwargs, **options) tasks, result, gid, args = self.prepare(options, args=partial_args, **kwargs) super(Group, self).apply_async((list(tasks), result.serializable(), gid, args), **options) return result def apply(self, args=(), kwargs={}, **options): return super(Group, self).apply( self.prepare(options, args=args, **kwargs), **options).get() return Group @shared_task def add_chain_task(app): from celery.canvas import chord, group, maybe_subtask _app = app class Chain(app.Task): app = _app name = 'celery.chain' accept_magic_kwargs = False def prepare_steps(self, args, tasks): steps = deque(tasks) next_step = prev_task = prev_res = None tasks, results = [], [] i = 0 while steps: # First task get partial args from chain. task = maybe_subtask(steps.popleft()) task = task.clone() if i else task.clone(args) i += 1 tid = task.options.get('task_id') if tid is None: tid = task.options['task_id'] = uuid() res = task.type.AsyncResult(tid) # automatically upgrade group(..) | s to chord(group, s) if isinstance(task, group): try: next_step = steps.popleft() except IndexError: next_step = None if next_step is not None: task = chord(task, body=next_step, task_id=tid) if prev_task: # link previous task to this task. prev_task.link(task) # set the results parent attribute. res.parent = prev_res results.append(res) tasks.append(task) prev_task, prev_res = task, res return tasks, results def apply_async(self, args=(), kwargs={}, group_id=None, chord=None, task_id=None, **options): if self.app.conf.CELERY_ALWAYS_EAGER: return self.apply(args, kwargs, **options) options.pop('publisher', None) tasks, results = self.prepare_steps(args, kwargs['tasks']) result = results[-1] if group_id: tasks[-1].set(group_id=group_id) if chord: tasks[-1].set(chord=chord) if task_id: tasks[-1].set(task_id=task_id) result = tasks[-1].type.AsyncResult(task_id) tasks[0].apply_async() return result def apply(self, args=(), kwargs={}, **options): tasks = [maybe_subtask(task).clone() for task in kwargs['tasks']] res = prev = None for task in tasks: res = task.apply((prev.get(), ) if prev else ()) res.parent, prev = prev, res return res return Chain @shared_task def add_chord_task(app): """Every chord is executed in a dedicated task, so that the chord can be used as a subtask, and this generates the task responsible for that.""" from celery import group from celery.canvas import maybe_subtask _app = app class Chord(app.Task): app = _app name = 'celery.chord' accept_magic_kwargs = False ignore_result = False def run(self, header, body, partial_args=(), interval=1, max_retries=None, propagate=False, eager=False, **kwargs): group_id = uuid() AsyncResult = self.app.AsyncResult prepare_member = self._prepare_member # - convert back to group if serialized if not isinstance(header, group): header = group([maybe_subtask(t) for t in header]) # - eager applies the group inline if eager: return header.apply(args=partial_args, task_id=group_id) results = [AsyncResult(prepare_member(task, body, group_id)) for task in header.tasks] # - fallback implementations schedules the chord_unlock task here app.backend.on_chord_apply(group_id, body, interval=interval, max_retries=max_retries, propagate=propagate, result=results) # - call the header group, returning the GroupResult. return header(*partial_args, task_id=group_id) def _prepare_member(self, task, body, group_id): opts = task.options # d.setdefault would work but generating uuid's are expensive try: task_id = opts['task_id'] except KeyError: task_id = opts['task_id'] = uuid() opts.update(chord=body, group_id=group_id) return task_id def apply_async(self, args=(), kwargs={}, task_id=None, **options): if self.app.conf.CELERY_ALWAYS_EAGER: return self.apply(args, kwargs, **options) group_id = options.pop('group_id', None) chord = options.pop('chord', None) header, body = (list(maybe_subtask(kwargs['header'])), maybe_subtask(kwargs['body'])) if group_id: body.set(group_id=group_id) if chord: body.set(chord=chord) callback_id = body.options.setdefault('task_id', task_id or uuid()) parent = super(Chord, self).apply_async((header, body, args), **options) body_result = self.AsyncResult(callback_id) body_result.parent = parent return body_result def apply(self, args=(), kwargs={}, propagate=True, **options): body = kwargs['body'] res = super(Chord, self).apply(args, dict(kwargs, eager=True), **options) return maybe_subtask(body).apply( args=(res.get(propagate=propagate).get(), )) return Chord