tasks.py 4.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200
  1. # -*- coding: utf-8 -*-
  2. from __future__ import absolute_import, unicode_literals
  3. from time import sleep
  4. from celery import chain, group, shared_task
  5. from celery.exceptions import SoftTimeLimitExceeded
  6. from celery.utils.log import get_task_logger
  7. from .conftest import get_redis_connection
  8. logger = get_task_logger(__name__)
  9. @shared_task
  10. def identity(x):
  11. return x
  12. @shared_task
  13. def add(x, y):
  14. """Add two numbers."""
  15. return x + y
  16. @shared_task
  17. def raise_error():
  18. """Deliberately raise an error."""
  19. raise ValueError("deliberate error")
  20. @shared_task(ignore_result=True)
  21. def add_ignore_result(x, y):
  22. """Add two numbers."""
  23. return x + y
  24. @shared_task
  25. def chain_add(x, y):
  26. (
  27. add.s(x, x) | add.s(y)
  28. ).apply_async()
  29. @shared_task
  30. def delayed_sum(numbers, pause_time=1):
  31. """Sum the iterable of numbers."""
  32. # Allow the task to be in STARTED state for
  33. # a limited period of time.
  34. sleep(pause_time)
  35. return sum(numbers)
  36. @shared_task
  37. def delayed_sum_with_soft_guard(numbers, pause_time=1):
  38. """Sum the iterable of numbers."""
  39. try:
  40. sleep(pause_time)
  41. return sum(numbers)
  42. except SoftTimeLimitExceeded:
  43. return 0
  44. @shared_task
  45. def tsum(nums):
  46. """Sum an iterable of numbers"""
  47. return sum(nums)
  48. @shared_task(bind=True)
  49. def add_replaced(self, x, y):
  50. """Add two numbers (via the add task)."""
  51. raise self.replace(add.s(x, y))
  52. @shared_task(bind=True)
  53. def add_to_all(self, nums, val):
  54. """Add the given value to all supplied numbers."""
  55. subtasks = [add.s(num, val) for num in nums]
  56. raise self.replace(group(*subtasks))
  57. @shared_task(bind=True)
  58. def add_to_all_to_chord(self, nums, val):
  59. for num in nums:
  60. self.add_to_chord(add.s(num, val))
  61. return 0
  62. @shared_task(bind=True)
  63. def add_chord_to_chord(self, nums, val):
  64. subtasks = [add.s(num, val) for num in nums]
  65. self.add_to_chord(group(subtasks) | tsum.s())
  66. return 0
  67. @shared_task
  68. def print_unicode(log_message='hå它 valmuefrø', print_message='hiöäüß'):
  69. """Task that both logs and print strings containing funny characters."""
  70. logger.warning(log_message)
  71. print(print_message)
  72. @shared_task
  73. def sleeping(i, **_):
  74. """Task sleeping for ``i`` seconds, and returning nothing."""
  75. sleep(i)
  76. @shared_task(bind=True)
  77. def ids(self, i):
  78. """Returns a tuple of ``root_id``, ``parent_id`` and
  79. the argument passed as ``i``."""
  80. return self.request.root_id, self.request.parent_id, i
  81. @shared_task(bind=True)
  82. def collect_ids(self, res, i):
  83. """Used as a callback in a chain or group where the previous tasks
  84. are :task:`ids`: returns a tuple of::
  85. (previous_result, (root_id, parent_id, i))
  86. """
  87. return res, (self.request.root_id, self.request.parent_id, i)
  88. @shared_task(bind=True, expires=60.0, max_retries=1)
  89. def retry_once(self):
  90. """Task that fails and is retried. Returns the number of retries."""
  91. if self.request.retries:
  92. return self.request.retries
  93. raise self.retry(countdown=0.1)
  94. @shared_task
  95. def redis_echo(message):
  96. """Task that appends the message to a redis list"""
  97. redis_connection = get_redis_connection()
  98. redis_connection.rpush('redis-echo', message)
  99. @shared_task(bind=True)
  100. def second_order_replace1(self, state=False):
  101. redis_connection = get_redis_connection()
  102. if not state:
  103. redis_connection.rpush('redis-echo', 'In A')
  104. new_task = chain(second_order_replace2.s(),
  105. second_order_replace1.si(state=True))
  106. raise self.replace(new_task)
  107. else:
  108. redis_connection.rpush('redis-echo', 'Out A')
  109. @shared_task(bind=True)
  110. def second_order_replace2(self, state=False):
  111. redis_connection = get_redis_connection()
  112. if not state:
  113. redis_connection.rpush('redis-echo', 'In B')
  114. new_task = chain(redis_echo.s("In/Out C"),
  115. second_order_replace2.si(state=True))
  116. raise self.replace(new_task)
  117. else:
  118. redis_connection.rpush('redis-echo', 'Out B')
  119. @shared_task(bind=True)
  120. def build_chain_inside_task(self):
  121. """Task to build a chain.
  122. This task builds a chain and returns the chain's AsyncResult
  123. to verify that Asyncresults are correctly converted into
  124. serializable objects"""
  125. test_chain = (
  126. add.s(1, 1) |
  127. add.s(2) |
  128. group(
  129. add.s(3),
  130. add.s(4)
  131. ) |
  132. add.s(5)
  133. )
  134. result = test_chain()
  135. return result
  136. class ExpectedException(Exception):
  137. pass
  138. @shared_task
  139. def fail(*args):
  140. raise ExpectedException('Task expected to fail')
  141. @shared_task
  142. def chord_error(*args):
  143. return args