__init__.py 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367
  1. # -*- coding: utf-8 -*-
  2. """
  3. celery.worker
  4. ~~~~~~~~~~~~~
  5. :class:`WorkController` can be used to instantiate in-process workers.
  6. The worker consists of several components, all managed by boot-steps
  7. (mod:`celery.abstract`).
  8. :copyright: (c) 2009 - 2012 by Ask Solem.
  9. :license: BSD, see LICENSE for more details.
  10. """
  11. from __future__ import absolute_import
  12. import atexit
  13. import logging
  14. import socket
  15. import sys
  16. import traceback
  17. from billiard import forking_enable
  18. from kombu.syn import detect_environment
  19. from kombu.utils.finalize import Finalize
  20. from celery import concurrency as _concurrency
  21. from celery import platforms
  22. from celery.app import app_or_default, set_default_app
  23. from celery.app.abstract import configurated, from_config
  24. from celery.exceptions import SystemTerminate
  25. from celery.utils.functional import noop
  26. from celery.utils.imports import qualname, reload_from_cwd
  27. from celery.utils.log import get_logger
  28. from celery.utils.threads import Event
  29. from celery.utils.timer2 import Schedule
  30. from . import abstract
  31. from . import state
  32. from .buckets import TaskBucket, FastQueue
  33. from .hub import BoundedSemaphore
  34. RUN = 0x1
  35. CLOSE = 0x2
  36. TERMINATE = 0x3
  37. logger = get_logger(__name__)
  38. class Namespace(abstract.Namespace):
  39. """This is the boot-step namespace of the :class:`WorkController`.
  40. It loads modules from :setting:`CELERYD_BOOT_STEPS`, and its
  41. own set of built-in boot-step modules.
  42. """
  43. name = "worker"
  44. builtin_boot_steps = ("celery.worker.autoscale",
  45. "celery.worker.autoreload",
  46. "celery.worker.consumer",
  47. "celery.worker.mediator")
  48. def modules(self):
  49. return (self.builtin_boot_steps
  50. + self.app.conf.CELERYD_BOOT_STEPS)
  51. class Pool(abstract.StartStopComponent):
  52. """The pool component.
  53. Describes how to initialize the worker pool, and starts and stops
  54. the pool during worker startup/shutdown.
  55. Adds attributes:
  56. * autoscale
  57. * pool
  58. * max_concurrency
  59. * min_concurrency
  60. """
  61. name = "worker.pool"
  62. requires = ("queues", )
  63. def __init__(self, w, autoscale=None, no_execv=False, **kwargs):
  64. w.autoscale = autoscale
  65. w.pool = None
  66. w.max_concurrency = None
  67. w.min_concurrency = w.concurrency
  68. w.no_execv = no_execv
  69. if w.autoscale:
  70. w.max_concurrency, w.min_concurrency = w.autoscale
  71. def create(self, w):
  72. threaded = not w.use_eventloop
  73. forking_enable(w.no_execv or not w.force_execv)
  74. procs = w.min_concurrency
  75. if not threaded:
  76. semaphore = w.semaphore = BoundedSemaphore(procs)
  77. pool = w.pool = self.instantiate(w.pool_cls, w.min_concurrency,
  78. initargs=(w.app, w.hostname),
  79. maxtasksperchild=w.max_tasks_per_child,
  80. timeout=w.task_time_limit,
  81. soft_timeout=w.task_soft_time_limit,
  82. putlocks=w.pool_putlocks and threaded,
  83. lost_worker_timeout=w.worker_lost_wait,
  84. with_task_thread=threaded,
  85. with_result_thread=threaded,
  86. with_supervisor_thread=threaded,
  87. semaphore=semaphore)
  88. return pool
  89. class Beat(abstract.StartStopComponent):
  90. """Component used to embed a celerybeat process.
  91. This will only be enabled if the ``beat``
  92. argument is set.
  93. """
  94. name = "worker.beat"
  95. def __init__(self, w, beat=False, **kwargs):
  96. self.enabled = w.beat = beat
  97. w.beat = None
  98. def create(self, w):
  99. from celery.beat import EmbeddedService
  100. b = w.beat = EmbeddedService(app=w.app,
  101. schedule_filename=w.schedule_filename,
  102. scheduler_cls=w.scheduler_cls)
  103. return b
  104. class Queues(abstract.Component):
  105. """This component initializes the internal queues
  106. used by the worker."""
  107. name = "worker.queues"
  108. def create(self, w):
  109. if not w.pool_cls.rlimit_safe:
  110. w.disable_rate_limits = True
  111. if w.disable_rate_limits:
  112. w.ready_queue = FastQueue()
  113. if w.use_eventloop:
  114. if w.pool_putlocks:
  115. w.ready_queue.put = w.process_task_sem
  116. else:
  117. w.ready_queue.put = w.process_task
  118. elif not w.pool_cls.requires_mediator:
  119. # just send task directly to pool, skip the mediator.
  120. w.ready_queue.put = w.process_task
  121. else:
  122. w.ready_queue = TaskBucket(task_registry=w.app.tasks)
  123. class Timers(abstract.Component):
  124. """This component initializes the internal timers used by the worker."""
  125. name = "worker.timers"
  126. requires = ("pool", )
  127. def create(self, w):
  128. options = {"on_error": self.on_timer_error,
  129. "on_tick": self.on_timer_tick}
  130. if w.use_eventloop:
  131. # the timers are fired by the hub, so don't use the Timer thread.
  132. w.timer = Schedule(max_interval=10, **options)
  133. else:
  134. if not w.timer_cls:
  135. # Default Timer is set by the pool, as e.g. eventlet
  136. # needs a custom implementation.
  137. w.timer_cls = w.pool.Timer
  138. w.timer = self.instantiate(w.pool.Timer,
  139. max_interval=w.timer_precision,
  140. **options)
  141. def on_timer_error(self, exc):
  142. logger.error("Timer error: %r", exc, exc_info=True)
  143. def on_timer_tick(self, delay):
  144. logger.debug("Timer wake-up! Next eta %s secs.", delay)
  145. class StateDB(abstract.Component):
  146. """This component sets up the workers state db if enabled."""
  147. name = "worker.state-db"
  148. def __init__(self, w, **kwargs):
  149. self.enabled = w.state_db
  150. w._persistence = None
  151. def create(self, w):
  152. w._persistence = state.Persistent(w.state_db)
  153. atexit.register(w._persistence.save)
  154. class WorkController(configurated):
  155. """Unmanaged worker instance."""
  156. RUN = RUN
  157. CLOSE = CLOSE
  158. TERMINATE = TERMINATE
  159. app = None
  160. concurrency = from_config()
  161. loglevel = logging.ERROR
  162. logfile = from_config("log_file")
  163. send_events = from_config()
  164. pool_cls = from_config("pool")
  165. consumer_cls = from_config("consumer")
  166. mediator_cls = from_config("mediator")
  167. timer_cls = from_config("timer")
  168. timer_precision = from_config("timer_precision")
  169. autoscaler_cls = from_config("autoscaler")
  170. autoreloader_cls = from_config("autoreloader")
  171. schedule_filename = from_config()
  172. scheduler_cls = from_config("celerybeat_scheduler")
  173. task_time_limit = from_config()
  174. task_soft_time_limit = from_config()
  175. max_tasks_per_child = from_config()
  176. pool_putlocks = from_config()
  177. force_execv = from_config()
  178. prefetch_multiplier = from_config()
  179. state_db = from_config()
  180. disable_rate_limits = from_config()
  181. worker_lost_wait = from_config()
  182. _state = None
  183. _running = 0
  184. def __init__(self, loglevel=None, hostname=None, ready_callback=noop,
  185. queues=None, app=None, pidfile=None, **kwargs):
  186. self.app = app_or_default(app or self.app)
  187. # all new threads start without a current app, so if an app is not
  188. # passed on to the thread it will fall back to the "default app",
  189. # which then could be the wrong app. So for the worker
  190. # we set this to always return our app. This is a hack,
  191. # and means that only a single app can be used for workers
  192. # running in the same process.
  193. set_default_app(self.app)
  194. self._shutdown_complete = Event()
  195. self.setup_defaults(kwargs, namespace="celeryd")
  196. self.app.select_queues(queues) # select queues subset.
  197. # Options
  198. self.loglevel = loglevel or self.loglevel
  199. self.hostname = hostname or socket.gethostname()
  200. self.ready_callback = ready_callback
  201. self._finalize = Finalize(self, self.stop, exitpriority=1)
  202. self.pidfile = pidfile
  203. self.pidlock = None
  204. self.use_eventloop = (detect_environment() == "default" and
  205. self.app.broker_connection().is_evented)
  206. # Initialize boot steps
  207. self.pool_cls = _concurrency.get_implementation(self.pool_cls)
  208. self.components = []
  209. self.namespace = Namespace(app=self.app).apply(self, **kwargs)
  210. def start(self):
  211. """Starts the workers main loop."""
  212. self._state = self.RUN
  213. if self.pidfile:
  214. self.pidlock = platforms.create_pidlock(self.pidfile)
  215. try:
  216. for i, component in enumerate(self.components):
  217. logger.debug("Starting %s...", qualname(component))
  218. self._running = i + 1
  219. component.start()
  220. logger.debug("%s OK!", qualname(component))
  221. except SystemTerminate:
  222. self.terminate()
  223. except Exception, exc:
  224. logger.error("Unrecoverable error: %r", exc,
  225. exc_info=True)
  226. self.stop()
  227. except (KeyboardInterrupt, SystemExit):
  228. self.stop()
  229. # Will only get here if running green,
  230. # makes sure all greenthreads have exited.
  231. self._shutdown_complete.wait()
  232. def process_task_sem(self, req):
  233. return self.semaphore.acquire(self.process_task, req)
  234. def process_task(self, req):
  235. """Process task by sending it to the pool of workers."""
  236. try:
  237. req.task.execute(req, self.pool, self.loglevel, self.logfile)
  238. except Exception, exc:
  239. logger.critical("Internal error: %r\n%s",
  240. exc, traceback.format_exc(), exc_info=True)
  241. except SystemTerminate:
  242. self.terminate()
  243. raise
  244. except BaseException, exc:
  245. self.stop()
  246. raise exc
  247. def signal_consumer_close(self):
  248. try:
  249. self.consumer.close()
  250. except AttributeError:
  251. pass
  252. def stop(self, in_sighandler=False):
  253. """Graceful shutdown of the worker server."""
  254. self.signal_consumer_close()
  255. if not in_sighandler or self.pool.signal_safe:
  256. self._shutdown(warm=True)
  257. def terminate(self, in_sighandler=False):
  258. """Not so graceful shutdown of the worker server."""
  259. self.signal_consumer_close()
  260. if not in_sighandler or self.pool.signal_safe:
  261. self._shutdown(warm=False)
  262. def _shutdown(self, warm=True):
  263. what = "Stopping" if warm else "Terminating"
  264. if self._state in (self.CLOSE, self.TERMINATE):
  265. return
  266. if self.pool:
  267. self.pool.close()
  268. if self._state != self.RUN or self._running != len(self.components):
  269. # Not fully started, can safely exit.
  270. self._state = self.TERMINATE
  271. self._shutdown_complete.set()
  272. return
  273. self._state = self.CLOSE
  274. for component in reversed(self.components):
  275. logger.debug("%s %s...", what, qualname(component))
  276. stop = component.stop
  277. if not warm:
  278. stop = getattr(component, "terminate", None) or stop
  279. stop()
  280. self.timer.stop()
  281. self.consumer.close_connection()
  282. if self.pidlock:
  283. self.pidlock.release()
  284. self._state = self.TERMINATE
  285. self._shutdown_complete.set()
  286. def reload(self, modules=None, reload=False, reloader=None):
  287. modules = self.app.loader.task_modules if modules is None else modules
  288. imp = self.app.loader.import_from_cwd
  289. for module in set(modules or ()):
  290. if module not in sys.modules:
  291. logger.debug("importing module %s", module)
  292. imp(module)
  293. elif reload:
  294. logger.debug("reloading module %s", module)
  295. reload_from_cwd(sys.modules[module], reloader)
  296. self.pool.restart()
  297. @property
  298. def state(self):
  299. return state