optimizing.rst 8.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260
  1. .. _guide-optimizing:
  2. ============
  3. Optimizing
  4. ============
  5. Introduction
  6. ============
  7. The default configuration makes a lot of compromises. It's not optimal for
  8. any single case, but works well enough for most situations.
  9. There are optimizations that can be applied based on specific use cases.
  10. Optimizations can apply to different properties of the running environment,
  11. be it the time tasks take to execute, the amount of memory used, or
  12. responsiveness at times of high load.
  13. Ensuring Operations
  14. ===================
  15. In the book `Programming Pearls`_, Jon Bentley presents the concept of
  16. back-of-the-envelope calculations by asking the question;
  17. ❝ How much water flows out of the Mississippi River in a day? ❞
  18. The point of this exercise [*]_ is to show that there is a limit
  19. to how much data a system can process in a timely manner.
  20. Back of the envelope calculations can be used as a means to plan for this
  21. ahead of time.
  22. In Celery; If a task takes 10 minutes to complete,
  23. and there are 10 new tasks coming in every minute, the queue will never
  24. be empty. This is why it's very important
  25. that you monitor queue lengths!
  26. A way to do this is by :ref:`using Munin <monitoring-munin>`.
  27. You should set up alerts, that will notify you as soon as any queue has
  28. reached an unacceptable size. This way you can take appropriate action
  29. like adding new worker nodes, or revoking unnecessary tasks.
  30. .. [*] The chapter is available to read for free here:
  31. `The back of the envelope`_. The book is a classic text. Highly
  32. recommended.
  33. .. _`Programming Pearls`: http://www.cs.bell-labs.com/cm/cs/pearls/
  34. .. _`The back of the envelope`:
  35. http://books.google.com/books?id=kse_7qbWbjsC&pg=PA67
  36. .. _optimizing-general-settings:
  37. General Settings
  38. ================
  39. .. _optimizing-librabbitmq:
  40. librabbitmq
  41. -----------
  42. If you're using RabbitMQ (AMQP) as the broker then you can install the
  43. :mod:`librabbitmq` module to use an optimized client written in C:
  44. .. code-block:: console
  45. $ pip install librabbitmq
  46. The 'amqp' transport will automatically use the librabbitmq module if it's
  47. installed, or you can also specify the transport you want directly by using
  48. the ``pyamqp://`` or ``librabbitmq://`` prefixes.
  49. .. _optimizing-connection-pools:
  50. Broker Connection Pools
  51. -----------------------
  52. The broker connection pool is enabled by default since version 2.5.
  53. You can tweak the :setting:`broker_pool_limit` setting to minimize
  54. contention, and the value should be based on the number of
  55. active threads/greenthreads using broker connections.
  56. .. _optimizing-transient-queues:
  57. Using Transient Queues
  58. ----------------------
  59. Queues created by Celery are persistent by default. This means that
  60. the broker will write messages to disk to ensure that the tasks will
  61. be executed even if the broker is restarted.
  62. But in some cases it's fine that the message is lost, so not all tasks
  63. require durability. You can create a *transient* queue for these tasks
  64. to improve performance:
  65. .. code-block:: python
  66. from kombu import Exchange, Queue
  67. task_queues = (
  68. Queue('celery', routing_key='celery'),
  69. Queue('transient', Exchange('transient', delivery_mode=1),
  70. routing_key='transient', durable=False),
  71. )
  72. or by using :setting:`task_routes`:
  73. .. code-block:: python
  74. task_routes = {
  75. 'proj.tasks.add': {'queue': 'celery', 'delivery_mode': 'transient'}
  76. }
  77. The ``delivery_mode`` changes how the messages to this queue are delivered.
  78. A value of 1 means that the message will not be written to disk, and a value
  79. of 2 (default) means that the message can be written to disk.
  80. To direct a task to your new transient queue you can specify the queue
  81. argument (or use the :setting:`task_routes` setting):
  82. .. code-block:: python
  83. task.apply_async(args, queue='transient')
  84. For more information see the :ref:`routing guide <guide-routing>`.
  85. .. _optimizing-worker-settings:
  86. Worker Settings
  87. ===============
  88. .. _optimizing-prefetch-limit:
  89. Prefetch Limits
  90. ---------------
  91. *Prefetch* is a term inherited from AMQP that is often misunderstood
  92. by users.
  93. The prefetch limit is a **limit** for the number of tasks (messages) a worker
  94. can reserve for itself. If it is zero, the worker will keep
  95. consuming messages, not respecting that there may be other
  96. available worker nodes that may be able to process them sooner [*]_,
  97. or that the messages may not even fit in memory.
  98. The workers' default prefetch count is the
  99. :setting:`worker_prefetch_multiplier` setting multiplied by the number
  100. of concurrency slots[*]_ (processes/threads/greenthreads).
  101. If you have many tasks with a long duration you want
  102. the multiplier value to be 1, which means it will only reserve one
  103. task per worker process at a time.
  104. However -- If you have many short-running tasks, and throughput/round trip
  105. latency is important to you, this number should be large. The worker is
  106. able to process more tasks per second if the messages have already been
  107. prefetched, and is available in memory. You may have to experiment to find
  108. the best value that works for you. Values like 50 or 150 might make sense in
  109. these circumstances. Say 64, or 128.
  110. If you have a combination of long- and short-running tasks, the best option
  111. is to use two worker nodes that are configured separately, and route
  112. the tasks according to the run-time. (see :ref:`guide-routing`).
  113. .. [*] RabbitMQ and other brokers deliver messages round-robin,
  114. so this doesn't apply to an active system. If there is no prefetch
  115. limit and you restart the cluster, there will be timing delays between
  116. nodes starting. If there are 3 offline nodes and one active node,
  117. all messages will be delivered to the active node.
  118. .. [*] This is the concurrency setting; :setting:`worker_concurrency` or the
  119. :option:`-c` option to the :program:`celery worker` program.
  120. Reserve one task at a time
  121. --------------------------
  122. The task message is only deleted from the queue after the task is
  123. :term:`acknowledged`, so if the worker crashes before acknowleding the task,
  124. it can be redelivered to another worker (or the same after recovery).
  125. When using the default of early acknowledgement, having a prefetch multiplier setting
  126. of 1, means the worker will reserve at most one extra task for every
  127. worker process: or in other words, if the worker is started with `-c 10`,
  128. the worker may reserve at most 20 tasks (10 unacknowledged tasks executing, and 10
  129. unacknowledged reserved tasks) at any time.
  130. Often users ask if disabling "prefetching of tasks" is possible, but what
  131. they really mean by that is to have a worker only reserve as many tasks as
  132. there are worker processes (10 unacknowledged tasks for `-c 10`)
  133. That is possible, but not without also enabling
  134. :term:`late acknowledgments`. Using this option over the
  135. default beahvior means a task that has already started executing will be
  136. retried in the event of a power failure or the worker instance being killed
  137. abruptly, so this also means the task must be :term:`idempotent`
  138. .. seealso::
  139. Notes at :ref:`faq-acks_late-vs-retry`.
  140. You can enable this behavior by using the following configuration options:
  141. .. code-block:: python
  142. task_acks_late = True
  143. worker_prefetch_multiplier = 1
  144. .. _prefork-pool-prefetch:
  145. Prefork pool prefetch settings
  146. ------------------------------
  147. The prefork pool will asynchronously send as many tasks to the processes
  148. as it can and this means that the processes are, in effect, prefetching
  149. tasks.
  150. This benefits performance but it also means that tasks may be stuck
  151. waiting for long running tasks to complete::
  152. -> send task T1 to process A
  153. # A executes T1
  154. -> send task T2 to process B
  155. # B executes T2
  156. <- T2 complete sent by process B
  157. -> send task T3 to process A
  158. # A still executing T1, T3 stuck in local buffer and will not start until
  159. # T1 returns, and other queued tasks will not be sent to idle processes
  160. <- T1 complete sent by process A
  161. # A executes T3
  162. The worker will send tasks to the process as long as the pipe buffer is
  163. writable. The pipe buffer size varies based on the operating system: some may
  164. have a buffer as small as 64kb but on recent Linux versions the buffer
  165. size is 1MB (can only be changed system wide).
  166. You can disable this prefetching behavior by enabling the :option:`-Ofair`
  167. worker option:
  168. .. code-block:: console
  169. $ celery -A proj worker -l info -Ofair
  170. With this option enabled the worker will only write to processes that are
  171. available for work, disabling the prefetch behavior::
  172. -> send task T1 to process A
  173. # A executes T1
  174. -> send task T2 to process B
  175. # B executes T2
  176. <- T2 complete sent by process B
  177. -> send T3 to process B
  178. # B executes T3
  179. <- T3 complete sent by process B
  180. <- T1 complete sent by process A