Sfoglia il codice sorgente

Renames celery.bin.celeryd to celery.bin.worker

Ask Solem 12 anni fa
parent
commit
43c4406910
47 ha cambiato i file con 296 aggiunte e 330 eliminazioni
  1. 1 1
      celery/__main__.py
  2. 1 1
      celery/app/base.py
  3. 6 5
      celery/apps/worker.py
  4. 1 1
      celery/bin/celery.py
  5. 0 16
      celery/bin/celeryctl.py
  6. 3 3
      celery/bin/celeryd_detach.py
  7. 0 0
      celery/bin/worker.py
  8. 5 4
      celery/tests/app/test_app.py
  9. 1 1
      celery/tests/bin/test_celeryd_detach.py
  10. 3 3
      celery/tests/bin/test_events.py
  11. 8 8
      celery/tests/bin/test_multi.py
  12. 5 5
      celery/tests/bin/test_worker.py
  13. 3 3
      celery/tests/functional/case.py
  14. 1 1
      celery/worker/consumer.py
  15. 6 6
      docs/configuration.rst
  16. 23 23
      docs/faq.rst
  17. 2 1
      docs/getting-started/brokers/beanstalk.rst
  18. 2 1
      docs/getting-started/brokers/couchdb.rst
  19. 1 1
      docs/getting-started/first-steps-with-celery.rst
  20. 5 5
      docs/getting-started/introduction.rst
  21. 2 2
      docs/getting-started/next-steps.rst
  22. 34 33
      docs/history/changelog-1.0.rst
  23. 26 26
      docs/history/changelog-2.0.rst
  24. 30 30
      docs/history/changelog-2.1.rst
  25. 31 31
      docs/history/changelog-2.2.rst
  26. 7 7
      docs/history/changelog-2.3.rst
  27. 15 15
      docs/history/changelog-2.4.rst
  28. 8 8
      docs/history/changelog-2.5.rst
  29. 3 3
      docs/reference/celery.bin.events.rst
  30. 3 3
      docs/reference/celery.bin.worker.rst
  31. 1 1
      docs/reference/celery.rst
  32. 1 1
      docs/reference/index.rst
  33. 24 25
      docs/tutorials/daemonizing.rst
  34. 6 4
      docs/userguide/monitoring.rst
  35. 1 1
      docs/userguide/optimizing.rst
  36. 1 1
      docs/userguide/signals.rst
  37. 2 2
      docs/userguide/workers.rst
  38. 1 1
      examples/app/myapp.py
  39. 3 2
      examples/eventlet/README.rst
  40. 2 2
      extra/centos/celeryd.init
  41. 2 2
      extra/centos/celeryd.sysconfig
  42. 4 3
      extra/mac/org.celeryq.beat.plist
  43. 0 28
      extra/mac/org.celeryq.celerymon.plist
  44. 4 3
      extra/mac/org.celeryq.worker.plist
  45. 1 1
      extra/release/core-modules.txt
  46. 0 1
      extra/release/doc4allmods
  47. 7 5
      extra/supervisord/celeryd.conf

+ 1 - 1
celery/__main__.py

@@ -31,7 +31,7 @@ def main():
 def _compat_worker():
     maybe_patch_concurrency()
     _warn_deprecated('celery worker')
-    from celery.bin.celeryd import main
+    from celery.bin.worker import main
     main()
 
 

+ 1 - 1
celery/app/base.py

@@ -143,7 +143,7 @@ class Celery(object):
                     .execute_from_commandline(argv)
 
     def worker_main(self, argv=None):
-        return instantiate('celery.bin.celeryd:WorkerCommand', app=self) \
+        return instantiate('celery.bin.worker:WorkerCommand', app=self) \
                     .execute_from_commandline(argv)
 
     def task(self, *args, **opts):

+ 6 - 5
celery/apps/worker.py

@@ -118,7 +118,8 @@ class Worker(WorkController):
 
         if getattr(os, 'getuid', None) and os.getuid() == 0:
             warnings.warn(RuntimeWarning(
-                'Running celeryd with superuser privileges is discouraged!'))
+                'Running the worker with superuser privileges is discouraged!',
+            ))
 
         if self.purge:
             self.purge_messages()
@@ -206,7 +207,7 @@ class Worker(WorkController):
         # Install signal handler so SIGHUP restarts the worker.
         if not self._isatty:
             # only install HUP handler if detached from terminal,
-            # so closing the terminal window doesn't restart celeryd
+            # so closing the terminal window doesn't restart the worker
             # into the background.
             if self.app.IS_OSX:
                 # OS X can't exec from a process using threads.
@@ -239,7 +240,7 @@ def _shutdown_handler(worker, sig='TERM', how='Warm', exc=SystemExit,
             if current_process()._name == 'MainProcess':
                 if callback:
                     callback(worker)
-                safe_say('celeryd: {0} shutdown (MainProcess)'.format(how))
+                safe_say('worker: {0} shutdown (MainProcess)'.format(how))
             if active_thread_count() > 1:
                 setattr(state, {'Warm': 'should_stop',
                                 'Cold': 'should_terminate'}[how], True)
@@ -259,7 +260,7 @@ else:
 
 
 def on_SIGINT(worker):
-    safe_say('celeryd: Hitting Ctrl+C again will terminate all running tasks!')
+    safe_say('worker: Hitting Ctrl+C again will terminate all running tasks!')
     install_worker_term_hard_handler(worker, sig='SIGINT')
 if not is_jython:
     install_worker_int_handler = partial(
@@ -279,7 +280,7 @@ def install_worker_restart_handler(worker, sig='SIGHUP'):
     def restart_worker_sig_handler(*args):
         """Signal handler restarting the current python program."""
         set_in_sighandler(True)
-        safe_say('Restarting celeryd ({0})'.format(' '.join(sys.argv)))
+        safe_say('Restarting celery worker ({0})'.format(' '.join(sys.argv)))
         import atexit
         atexit.register(_clone_current_worker)
         from celery.worker import state

+ 1 - 1
celery/bin/celery.py

@@ -279,7 +279,7 @@ class worker(Delegate):
 
         celery worker --autoscale=10,0
     """
-    Command = 'celery.bin.celeryd:WorkerCommand'
+    Command = 'celery.bin.worker:WorkerCommand'
 
     def run_from_argv(self, prog_name, argv):
         self.target.maybe_detach(argv)

+ 0 - 16
celery/bin/celeryctl.py

@@ -1,16 +0,0 @@
-# -*- coding: utf-8 -*-
-"""
-    celery.bin.celeryctl
-    ~~~~~~~~~~~~~~~~~~~~
-
-    Now replaced by the :program:`celery` command.
-
-"""
-from __future__ import absolute_import
-
-from celery.bin.celery import (  # noqa
-    CeleryCommand as celeryctl, Command, main,
-)
-
-if __name__ == '__main__':  # pragma: no cover
-    main()

+ 3 - 3
celery/bin/celeryd_detach.py

@@ -3,7 +3,7 @@
     celery.bin.celeryd_detach
     ~~~~~~~~~~~~~~~~~~~~~~~~~
 
-    Program used to daemonize celeryd.
+    Program used to daemonize the worker
 
     Using :func:`os.execv` because forking and multiprocessing
     leads to weird issues (it was a long time ago now, but it
@@ -104,11 +104,11 @@ class detached_celeryd(object):
     option_list = OPTION_LIST
     usage = '%prog [options] [celeryd options]'
     version = celery.VERSION_BANNER
-    description = ('Detaches Celery worker nodes.  See `celeryd --help` '
+    description = ('Detaches Celery worker nodes.  See `celery worker --help` '
                    'for the list of supported worker arguments.')
     command = sys.executable
     execv_path = sys.executable
-    execv_argv = ['-m', 'celery.bin.celeryd']
+    execv_argv = ['-m', 'celery', 'worker']
 
     def Parser(self, prog_name):
         return PartialOptionParser(prog=prog_name,

+ 0 - 0
celery/bin/celeryd.py → celery/bin/worker.py


+ 5 - 4
celery/tests/app/test_app.py

@@ -268,19 +268,20 @@ class test_App(Case):
         self.assertDictContainsSubset(changes, restored.conf)
 
     def test_worker_main(self):
-        from celery.bin import celeryd
+        from celery.bin import worker as worker_bin
 
-        class WorkerCommand(celeryd.WorkerCommand):
+        class WorkerCommand(worker_bin.WorkerCommand):
 
             def execute_from_commandline(self, argv):
                 return argv
 
-        prev, celeryd.WorkerCommand = celeryd.WorkerCommand, WorkerCommand
+        prev, worker_bin.WorkerCommand = \
+                worker_bin.WorkerCommand, WorkerCommand
         try:
             ret = self.app.worker_main(argv=['--version'])
             self.assertListEqual(ret, ['--version'])
         finally:
-            celeryd.WorkerCommand = prev
+            worker_bin.WorkerCommand = prev
 
     def test_config_from_envvar(self):
         os.environ['CELERYTEST_CONFIG_OBJECT'] = 'celery.tests.app.test_app'

+ 1 - 1
celery/tests/bin/test_celeryd_detach.py

@@ -87,7 +87,7 @@ class test_Command(Case):
         detach.assert_called_with(path=x.execv_path, uid=None, gid=None,
             umask=0, fake=False,
             logfile='/var/log', pidfile='celeryd.pid',
-            argv=['-m', 'celery.bin.celeryd', '-c', '1', '-lDEBUG',
+            argv=['-m', 'celery', 'worker', '-c', '1', '-lDEBUG',
                   '--logfile=/var/log', '--pidfile=celeryd.pid',
                   '--', '.disable_rate_limits=1'],
         )

+ 3 - 3
celery/tests/bin/test_events.py

@@ -31,7 +31,7 @@ class test_EvCommand(Case):
     @patch('celery.bin.events', 'set_process_title', proctitle)
     def test_run_dump(self):
         self.assertEqual(self.ev.run(dump=True), 'me dumper, you?')
-        self.assertIn('celeryev:dump', proctitle.last[0])
+        self.assertIn('celery events:dump', proctitle.last[0])
 
     def test_run_top(self):
         try:
@@ -43,7 +43,7 @@ class test_EvCommand(Case):
         @patch('celery.bin.events', 'set_process_title', proctitle)
         def _inner():
             self.assertEqual(self.ev.run(), 'me top, you?')
-            self.assertIn('celeryev:top', proctitle.last[0])
+            self.assertIn('celery events:top', proctitle.last[0])
         return _inner()
 
     @patch('celery.events.snapshot', 'evcam', lambda *a, **k: (a, k))
@@ -55,7 +55,7 @@ class test_EvCommand(Case):
         self.assertIsNone(kw['maxrate'])
         self.assertEqual(kw['loglevel'], 'INFO')
         self.assertEqual(kw['logfile'], 'logfile')
-        self.assertIn('celeryev:cam', proctitle.last[0])
+        self.assertIn('celery events:cam', proctitle.last[0])
 
     @mpatch('celery.events.snapshot.evcam')
     @mpatch('celery.bin.events.detached')

+ 8 - 8
celery/tests/bin/test_multi.py

@@ -332,7 +332,7 @@ class test_MultiTool(Case):
         self.t.node_alive.return_value = False
 
         callback = Mock()
-        self.t.stop(['foo', 'bar', 'baz'], 'celeryd', callback=callback)
+        self.t.stop(['foo', 'bar', 'baz'], 'celery worker', callback=callback)
         sigs = sorted(self.t.signal_node.call_args_list)
         self.assertEqual(len(sigs), 2)
         self.assertIn(('foo@e.com', 10, signal.SIGTERM),
@@ -341,7 +341,7 @@ class test_MultiTool(Case):
                 [tup[0] for tup in sigs])
         self.t.signal_node.return_value = False
         self.assertTrue(callback.called)
-        self.t.stop(['foo', 'bar', 'baz'], 'celeryd', callback=None)
+        self.t.stop(['foo', 'bar', 'baz'], 'celery worker', callback=None)
 
         def on_node_alive(pid):
             if node_alive.call_count > 4:
@@ -349,7 +349,7 @@ class test_MultiTool(Case):
             return False
         self.t.signal_node.return_value = True
         self.t.node_alive.side_effect = on_node_alive
-        self.t.stop(['foo', 'bar', 'baz'], 'celeryd', retry=True)
+        self.t.stop(['foo', 'bar', 'baz'], 'celery worker', retry=True)
 
     @patch('os.kill')
     def test_node_alive(self, kill):
@@ -387,13 +387,13 @@ class test_MultiTool(Case):
     def test_start(self):
         self.t.waitexec = Mock()
         self.t.waitexec.return_value = 0
-        self.assertFalse(self.t.start(['foo', 'bar', 'baz'], 'celeryd'))
+        self.assertFalse(self.t.start(['foo', 'bar', 'baz'], 'celery worker'))
 
         self.t.waitexec.return_value = 1
-        self.assertFalse(self.t.start(['foo', 'bar', 'baz'], 'celeryd'))
+        self.assertFalse(self.t.start(['foo', 'bar', 'baz'], 'celery worker'))
 
     def test_show(self):
-        self.t.show(['foo', 'bar', 'baz'], 'celeryd')
+        self.t.show(['foo', 'bar', 'baz'], 'celery worker')
         self.assertTrue(self.fh.getvalue())
 
     @patch('socket.gethostname')
@@ -407,7 +407,7 @@ class test_MultiTool(Case):
     @patch('socket.gethostname')
     def test_names(self, gethostname):
         gethostname.return_value = 'e.com'
-        self.t.names(['foo', 'bar', 'baz'], 'celeryd')
+        self.t.names(['foo', 'bar', 'baz'], 'celery worker')
         self.assertIn('foo@e.com\nbar@e.com\nbaz@e.com', self.fh.getvalue())
 
     def test_execute_from_commandline(self):
@@ -438,7 +438,7 @@ class test_MultiTool(Case):
 
     def test_stopwait(self):
         self.t._stop_nodes = Mock()
-        self.t.stopwait(['foo', 'bar', 'baz'], 'celeryd')
+        self.t.stopwait(['foo', 'bar', 'baz'], 'celery worker')
         self.assertEqual(self.t._stop_nodes.call_args[1]['retry'], 2)
 
     @patch('celery.bin.multi.MultiTool')

+ 5 - 5
celery/tests/bin/test_celeryd.py → celery/tests/bin/test_worker.py

@@ -17,7 +17,7 @@ from celery import platforms
 from celery import signals
 from celery import current_app
 from celery.apps import worker as cd
-from celery.bin.celeryd import WorkerCommand, main as celeryd_main
+from celery.bin.worker import WorkerCommand, main as worker_main
 from celery.exceptions import ImproperlyConfigured, SystemTerminate
 from celery.task import trace
 from celery.utils.log import ensure_process_aware_logger
@@ -101,7 +101,7 @@ class test_Worker(WorkerAppCase):
         x = WorkerCommand()
         x.run = Mock()
         with self.assertRaises(ImportError):
-            x.execute_from_commandline(['celeryd', '-P', 'xyzybox'])
+            x.execute_from_commandline(['worker', '-P', 'xyzybox'])
 
     @disable_stdouts
     def test_invalid_loglevel_gives_error(self):
@@ -405,15 +405,15 @@ class test_funs(WorkerAppCase):
     def test_parse_options(self):
         cmd = WorkerCommand()
         cmd.app = current_app
-        opts, args = cmd.parse_options('celeryd', ['--concurrency=512'])
+        opts, args = cmd.parse_options('worker', ['--concurrency=512'])
         self.assertEqual(opts.concurrency, 512)
 
     @disable_stdouts
     def test_main(self):
         p, cd.Worker = cd.Worker, Worker
-        s, sys.argv = sys.argv, ['celeryd', '--discard']
+        s, sys.argv = sys.argv, ['worker', '--discard']
         try:
-            celeryd_main()
+            worker_main()
         finally:
             cd.Worker = p
             sys.argv = s

+ 3 - 3
celery/tests/functional/case.py

@@ -52,9 +52,9 @@ class Worker(object):
         pid = os.fork()
         if pid == 0:
             from celery import current_app
-            current_app.worker_main(['celeryd', '--loglevel=INFO',
-                                                '-n', self.hostname,
-                                                '-P', 'solo'])
+            current_app.worker_main(['worker', '--loglevel=INFO',
+                                               '-n', self.hostname,
+                                               '-P', 'solo'])
             os._exit(0)
         self.pid = pid
 

+ 1 - 1
celery/worker/consumer.py

@@ -301,7 +301,7 @@ class Consumer(object):
         as a dict.
 
         This is also the consumer related info returned by
-        ``celeryctl stats``.
+        :program:`celery inspect stats`.
 
         """
         conninfo = {}

+ 6 - 6
docs/configuration.rst

@@ -1017,17 +1017,17 @@ has been executed, not *just before*, which is the default behavior.
 
     FAQ: :ref:`faq-acks_late-vs-retry`.
 
-.. _conf-celeryd:
+.. _conf-worker:
 
-Worker: celeryd
----------------
+Worker
+------
 
 .. setting:: CELERY_IMPORTS
 
 CELERY_IMPORTS
 ~~~~~~~~~~~~~~
 
-A sequence of modules to import when the celery daemon starts.
+A sequence of modules to import when the worker starts.
 
 This is used to specify the task modules to import, but also
 to import signal handlers and additional remote control commands, etc.
@@ -1128,7 +1128,7 @@ Can be a relative or absolute path, but be aware that the suffix `.db`
 may be appended to the file name (depending on Python version).
 
 Can also be set via the :option:`--statedb` argument to
-:mod:`~celery.bin.celeryd`.
+:mod:`~celery.bin.worker`.
 
 Not enabled by default.
 
@@ -1373,7 +1373,7 @@ CELERYD_TASK_LOG_FORMAT
 ~~~~~~~~~~~~~~~~~~~~~~~
 
 The format to use for log messages logged in tasks.  Can be overridden using
-the :option:`--loglevel` option to :mod:`~celery.bin.celeryd`.
+the :option:`--loglevel` option to :mod:`~celery.bin.worker`.
 
 Default is::
 

+ 23 - 23
docs/faq.rst

@@ -223,7 +223,7 @@ Is Celery multilingual?
 
 **Answer:** Yes.
 
-:mod:`~celery.bin.celeryd` is an implementation of Celery in Python. If the
+:mod:`~celery.bin.worker` is an implementation of Celery in Python. If the
 language has an AMQP client, there shouldn't be much work to create a worker
 in your language.  A Celery worker is just a program connecting to the broker
 to process messages.
@@ -259,8 +259,8 @@ Transaction Model and Locking`_ in the MySQL user manual.
 
 .. _faq-worker-hanging:
 
-celeryd is not doing anything, just hanging
---------------------------------------------
+The worker is not doing anything, just hanging
+----------------------------------------------
 
 **Answer:** See `MySQL is throwing deadlock errors, what can I do?`_.
             or `Why is Task.delay/apply\* just hanging?`.
@@ -275,8 +275,8 @@ using MySQL, see `MySQL is throwing deadlock errors, what can I do?`_.
 
 .. _faq-publish-hanging:
 
-Why is Task.delay/apply\*/celeryd just hanging?
------------------------------------------------
+Why is Task.delay/apply\*/the worker just hanging?
+--------------------------------------------------
 
 **Answer:** There is a bug in some AMQP clients that will make it hang if
 it's not able to authenticate the current user, the password doesn't match or
@@ -284,7 +284,7 @@ the user does not have access to the virtual host specified. Be sure to check
 your broker logs (for RabbitMQ that is :file:`/var/log/rabbitmq/rabbit.log` on
 most systems), it usually contains a message describing the reason.
 
-.. _faq-celeryd-on-freebsd:
+.. _faq-worker-on-freebsd:
 
 Does it work on FreeBSD?
 ------------------------
@@ -323,7 +323,7 @@ This shows that there's 2891 messages waiting to be processed in the task
 queue, and there are two consumers processing them.
 
 One reason that the queue is never emptied could be that you have a stale
-worker process taking the messages hostage. This could happen if celeryd
+worker process taking the messages hostage. This could happen if the worker
 wasn't properly shut down.
 
 When a message is received by a worker the broker waits for it to be
@@ -355,7 +355,7 @@ task manually:
     >>> from myapp.tasks import MyPeriodicTask
     >>> MyPeriodicTask.delay()
 
-Watch celeryd`s log file to see if it's able to find the task, or if some
+Watch the workers log file to see if it's able to find the task, or if some
 other error is happening.
 
 .. _faq-periodic-task-does-not-run:
@@ -394,7 +394,7 @@ you have to use the AMQP API or the :program:`celery amqp` utility:
 
 The number 1753 is the number of messages deleted.
 
-You can also start :mod:`~celery.bin.celeryd` with the
+You can also start :mod:`~celery.bin.worker` with the
 :option:`--purge` argument, to purge messages when the worker starts.
 
 .. _faq-messages-left-after-purge:
@@ -470,8 +470,8 @@ You can enable this using the :setting:`BROKER_USE_SSL` setting.
 It is also possible to add additional encryption and security to messages,
 if you have a need for this then you should contact the :ref:`mailing-list`.
 
-Is it safe to run :program:`celeryd` as root?
----------------------------------------------
+Is it safe to run :program:`celery worker` as root?
+---------------------------------------------------
 
 **Answer**: No!
 
@@ -508,7 +508,7 @@ important that you are aware of the common pitfalls.
 
 * Events.
 
-Running :mod:`~celery.bin.celeryd` with the :option:`-E`/:option:`--events`
+Running :mod:`~celery.bin.worker` with the :option:`-E`/:option:`--events`
 option will send messages for events happening inside of the worker.
 
 Events should only be enabled if you have an active monitor consuming them,
@@ -583,7 +583,7 @@ without a tty to run sudo::
     Defaults requiretty
 
 If you have this configuration in your :file:`/etc/sudoers` file then
-tasks will not be able to call sudo when celeryd is running as a daemon.
+tasks will not be able to call sudo when the worker is running as a daemon.
 If you want to enable that, then you need to remove the line from sudoers.
 
 See: http://timelordz.com/wiki/Apache_Sudo_Commands
@@ -707,15 +707,15 @@ uses its host name to create a unique queue name to listen to,
 so if you have more than one worker with the same host name, the
 control commands will be received in round-robin between them.
 
-To work around this you can explicitly set the host name for every worker
-using the :option:`--hostname` argument to :mod:`~celery.bin.celeryd`:
+To work around this you can explicitly set the nodename for every worker
+using the :option:`-n` argument to :mod:`~celery.bin.worker`:
 
 .. code-block:: bash
 
-    $ celeryd --hostname=$(hostname).1
-    $ celeryd --hostname=$(hostname).2
+    $ celery worker -n worker1@%h
+    $ celery worker -n worker2@%h
 
-etc., etc...
+where ``%h`` is automatically expanded into the current hostname.
 
 .. _faq-task-routing:
 
@@ -829,13 +829,13 @@ Or to schedule a periodic task at a specific time, use the
 
 .. _faq-safe-worker-shutdown:
 
-How do I shut down `celeryd` safely?
+How can I safely shut down the worker?
 --------------------------------------
 
 **Answer**: Use the :sig:`TERM` signal, and the worker will finish all currently
 executing jobs and shut down as soon as possible. No tasks should be lost.
 
-You should never stop :mod:`~celery.bin.celeryd` with the :sig:`KILL` signal
+You should never stop :mod:`~celery.bin.worker` with the :sig:`KILL` signal
 (:option:`-9`), unless you've tried :sig:`TERM` a few times and waited a few
 minutes to let it get a chance to shut down.  As if you do tasks may be
 terminated mid-execution, and they will not be re-run unless you have the
@@ -847,8 +847,8 @@ terminated mid-execution, and they will not be re-run unless you have the
 
 .. _faq-daemonizing:
 
-How do I run celeryd in the background on [platform]?
------------------------------------------------------
+How do I run the worker in the background on [platform]?
+--------------------------------------------------------
 **Answer**: Please see :ref:`daemonizing`.
 
 .. _faq-django:
@@ -890,7 +890,7 @@ Windows
 
 .. _faq-windows-worker-embedded-beat:
 
-The `-B` / `--beat` option to celeryd doesn't work?
+The `-B` / `--beat` option to worker doesn't work?
 ----------------------------------------------------------------
 **Answer**: That's right. Run `celery beat` and `celery worker` as separate
 services instead.

+ 2 - 1
docs/getting-started/brokers/beanstalk.rst

@@ -54,6 +54,7 @@ Limitations
 
 The Beanstalk message transport does not currently support:
 
-    * Remote control commands (celeryctl, broadcast)
+    * Remote control commands (:program:`celery control`,
+      :program:`celery inspect`, broadcast)
     * Authentication
 

+ 2 - 1
docs/getting-started/brokers/couchdb.rst

@@ -54,4 +54,5 @@ Limitations
 
 The CouchDB message transport does not currently support:
 
-    * Remote control commands (celeryctl, broadcast)
+    * Remote control commands (:program:`celery inspect`,
+      :program:`celery control`, broadcast)

+ 1 - 1
docs/getting-started/first-steps-with-celery.rst

@@ -153,7 +153,7 @@ e.g. for Redis you can use ``redis://localhost``, or MongoDB:
 
 You defined a single task, called ``add``, which returns the sum of two numbers.
 
-.. _celerytut-running-celeryd:
+.. _celerytut-running-the-worker:
 
 Running the celery worker server
 ================================

+ 5 - 5
docs/getting-started/introduction.rst

@@ -277,11 +277,11 @@ Quickjump
         - :ref:`retry a task when it fails <task-retry>`
         - :ref:`get the id of the current task <task-request-info>`
         - :ref:`know what queue a task was delivered to <task-request-info>`
-        - :ref:`see a list of running workers <monitoring-celeryctl>`
-        - :ref:`purge all messages <monitoring-celeryctl>`
-        - :ref:`inspect what the workers are doing <monitoring-celeryctl>`
-        - :ref:`see what tasks a worker has registerd <monitoring-celeryctl>`
-        - :ref:`migrate tasks to a new broker <monitoring-celeryctl>`
+        - :ref:`see a list of running workers <monitoring-control>`
+        - :ref:`purge all messages <monitoring-control>`
+        - :ref:`inspect what the workers are doing <monitoring-control>`
+        - :ref:`see what tasks a worker has registerd <monitoring-control>`
+        - :ref:`migrate tasks to a new broker <monitoring-control>`
         - :ref:`see a list of event message types <event-reference>`
         - :ref:`contribute to Celery <contributing>`
         - :ref:`learn about available configuration settings <configuration>`

+ 2 - 2
docs/getting-started/next-steps.rst

@@ -167,7 +167,7 @@ You can restart it too:
     > Waiting for 1 node.....
         > w1.halcyon.local: OK
     > Restarting node w1.halcyon.local: OK
-    celeryd multi v3.1.0 (Cipater)
+    celery multi v3.1.0 (Cipater)
     > Stopping nodes...
         > w1.halcyon.local: TERM -> 64052
 
@@ -209,7 +209,7 @@ e.g:
 
 .. code-block:: bash
 
-    $ celeryd multi start 10 -A proj -l info -Q:1-3 images,video -Q:4,5 data \
+    $ celery multi start 10 -A proj -l info -Q:1-3 images,video -Q:4,5 data \
         -Q default -L:4,5 debug
 
 For more examples see the :mod:`~celery.bin.multi` module in the API

+ 34 - 33
docs/history/changelog-1.0.rst

@@ -52,11 +52,11 @@ Critical
 
 * Now depends on :mod:`billiard` >= 0.3.1
 
-* celeryd: Previously exceptions raised by worker components could stall startup,
+* worker: Previously exceptions raised by worker components could stall startup,
   now it correctly logs the exceptions and shuts down.
 
-* celeryd: Prefetch counts was set too late. QoS is now set as early as possible,
-  so celeryd can't slurp in all the messages at start-up.
+* worker: Prefetch counts was set too late. QoS is now set as early as possible,
+  so the worker: can't slurp in all the messages at start-up.
 
 .. _v105-changes:
 
@@ -397,7 +397,7 @@ Fixes
 
     See: http://bit.ly/94fwdd
 
-* celeryd: The worker components are now configurable: :setting:`CELERYD_POOL`,
+* worker: The worker components are now configurable: :setting:`CELERYD_POOL`,
   :setting:`CELERYD_CONSUMER`, :setting:`CELERYD_MEDIATOR`, and
   :setting:`CELERYD_ETA_SCHEDULER`.
 
@@ -442,7 +442,7 @@ Fixes
 * Fixed a potential infinite loop in `BaseAsyncResult.__eq__`, although
   there is no evidence that it has ever been triggered.
 
-* celeryd: Now handles messages with encoding problems by acking them and
+* worker: Now handles messages with encoding problems by acking them and
   emitting an error message.
 
 .. _version-1.0.1:
@@ -474,9 +474,10 @@ Fixes
         worked on, this patch would enable us to use a better solution, and is
         scheduled for inclusion in the `2.0.0` release.
 
-* celeryd now shutdowns cleanly when receiving the :sig:`SIGTERM` signal.
+* The worker now shutdowns cleanly when receiving the :sig:`SIGTERM` signal.
 
-* celeryd now does a cold shutdown if the :sig:`SIGINT` signal is received (Ctrl+C),
+* The worker now does a cold shutdown if the :sig:`SIGINT` signal
+  is received (Ctrl+C),
   this means it tries to terminate as soon as possible.
 
 * Caching of results now moved to the base backend classes, so no need
@@ -492,7 +493,7 @@ Fixes
     `backend.reload_taskset_result` (that's for those who want to send
     results incrementally).
 
-* `celeryd` now works on Windows again.
+* The worker now works on Windows again.
 
     .. warning::
 
@@ -555,7 +556,7 @@ Fixes
 * The redis result backend no longer calls `SAVE` when disconnecting,
   as this is apparently better handled by Redis itself.
 
-* If `settings.DEBUG` is on, celeryd now warns about the possible
+* If `settings.DEBUG` is on, the worker now warns about the possible
   memory leak it can result in.
 
 * The ETA scheduler now sleeps at most two seconds between iterations.
@@ -579,7 +580,7 @@ Fixes
 * Events: Fields was not passed by `.send()` (fixes the UUID key errors
   in celerymon)
 
-* Added `--schedule`/`-s` option to celeryd, so it is possible to
+* Added `--schedule`/`-s` option to the worker, so it is possible to
   specify a custom schedule filename when using an embedded celerybeat
   server (the `-B`/`--beat`) option.
 
@@ -624,7 +625,7 @@ Backward incompatible changes
   available on your platform, or something like Supervisord to make
   celeryd/celerybeat/celerymon into background processes.
 
-    We've had too many problems with celeryd daemonizing itself, so it was
+    We've had too many problems with the worker daemonizing itself, so it was
     decided it has to be removed. Example startup scripts has been added to
     the `extra/` directory:
 
@@ -651,8 +652,8 @@ Backward incompatible changes
     Also the following configuration keys has been removed:
     `CELERYD_PID_FILE`, `CELERYBEAT_PID_FILE`, `CELERYMON_PID_FILE`.
 
-* Default celeryd loglevel is now `WARN`, to enable the previous log level
-  start celeryd with `--loglevel=INFO`.
+* Default worker loglevel is now `WARN`, to enable the previous log level
+  start the worker with `--loglevel=INFO`.
 
 * Tasks are automatically registered.
 
@@ -698,7 +699,7 @@ Backward incompatible changes
 
 * The periodic task system has been rewritten to a centralized solution.
 
-    This means `celeryd` no longer schedules periodic tasks by default,
+    This means the worker no longer schedules periodic tasks by default,
     but a new daemon has been introduced: `celerybeat`.
 
     To launch the periodic task scheduler you have to run celerybeat:
@@ -710,7 +711,7 @@ Backward incompatible changes
     Make sure this is running on one server only, if you run it twice, all
     periodic tasks will also be executed twice.
 
-    If you only have one worker server you can embed it into celeryd like this:
+    If you only have one worker server you can embed it into the worker like this:
 
     .. code-block:: bash
 
@@ -808,7 +809,7 @@ News
 
 * New cool task decorator syntax.
 
-* celeryd now sends events if enabled with the `-E` argument.
+* worker: now sends events if enabled with the `-E` argument.
 
     Excellent for monitoring tools, one is already in the making
     (http://github.com/celery/celerymon).
@@ -819,7 +820,7 @@ News
 
 * You can now delete (revoke) tasks that has already been applied.
 
-* You can now set the hostname celeryd identifies as using the `--hostname`
+* You can now set the hostname the worker identifies as using the `--hostname`
   argument.
 
 * Cache backend now respects the :setting:`CELERY_TASK_RESULT_EXPIRES` setting.
@@ -827,7 +828,7 @@ News
 * Message format has been standardized and now uses ISO-8601 format
   for dates instead of datetime.
 
-* `celeryd` now responds to the :sig:`SIGHUP` signal by restarting itself.
+* worker now responds to the :sig:`SIGHUP` signal by restarting itself.
 
 * Periodic tasks are now scheduled on the clock.
 
@@ -924,7 +925,7 @@ Changes
 * You can now set :setting:`CELERY_IGNORE_RESULT` to ignore task results by
   default (if enabled, tasks doesn't save results or errors to the backend used).
 
-* celeryd now correctly handles malformed messages by throwing away and
+* The worker now correctly handles malformed messages by throwing away and
   acknowledging the message, instead of crashing.
 
 .. _v100-bugs:
@@ -1134,7 +1135,7 @@ Important changes
 
     This to not receive more messages than we can handle.
 
-* Now redirects stdout/stderr to the celeryd log file when detached
+* Now redirects stdout/stderr to the workers log file when detached
 
 * Now uses `inspect.getargspec` to only pass default arguments
     the task supports.
@@ -1156,7 +1157,7 @@ Important changes
     This feature misses documentation and tests, so anyone interested
     is encouraged to improve this situation.
 
-* celeryd now survives a restart of the AMQP server!
+* The worker now survives a restart of the AMQP server!
 
   Automatically re-establish AMQP broker connection if it's lost.
 
@@ -1232,7 +1233,7 @@ Important changes
 
 * Fixed a bug where tasks raising unpickleable exceptions crashed pool
     workers. So if you've had pool workers mysteriously disappearing, or
-    problems with celeryd stopping working, this has been fixed in this
+    problems with the worker stopping working, this has been fixed in this
     version.
 
 * Fixed a race condition with periodic tasks.
@@ -1255,7 +1256,7 @@ News
 
 * New Tutorial: Creating a click counter using carrot and celery
 
-* Database entries for periodic tasks are now created at `celeryd`
+* Database entries for periodic tasks are now created at the workers
     startup instead of for each check (which has been a forgotten TODO/XXX
     in the code for a long time)
 
@@ -1264,7 +1265,7 @@ News
     stored task results are deleted. For the moment this only works for the
     database backend.
 
-* `celeryd` now emits a debug log message for which periodic tasks
+* The worker now emits a debug log message for which periodic tasks
     has been launched.
 
 * The periodic task table is now locked for reading while getting
@@ -1369,13 +1370,13 @@ News
   being raised.)
 
 * Added support for statistics for profiling and monitoring.
-  To start sending statistics start `celeryd` with the
+  To start sending statistics start the worker with the
   `--statistics option. Then after a while you can dump the results
   by running `python manage.py celerystats`. See
   `celery.monitoring` for more information.
 
 * The celery daemon can now be supervised (i.e. it is automatically
-  restarted if it crashes). To use this start celeryd with the
+  restarted if it crashes). To use this start the worker with the
   --supervised` option (or alternatively `-S`).
 
 * views.apply: View calling a task. Example
@@ -1457,7 +1458,7 @@ News
 * Now works with PostgreSQL (psycopg2) again by registering the
   `PickledObject` field.
 
-* `celeryd`: Added `--detach` option as an alias to `--daemon`, and
+* Worker: Added `--detach` option as an alias to `--daemon`, and
   it's the term used in the documentation from now on.
 
 * Make sure the pool and periodic task worker thread is terminated
@@ -1488,10 +1489,10 @@ News
 =====
 :release-date: 2009-06-08 01:07 P.M CET
 
-* celeryd: Added option `--discard`: Discard (delete!) all waiting
+* worker: Added option `--discard`: Discard (delete!) all waiting
   messages in the queue.
 
-* celeryd: The `--wakeup-after` option was not handled as a float.
+* Worker: The `--wakeup-after` option was not handled as a float.
 
 .. _version-0.3.1:
 
@@ -1741,7 +1742,7 @@ arguments, so be sure to flush your task queue before you upgrade.
 
 * Better test coverage
 * More documentation
-* celeryd doesn't emit `Queue is empty` message if
+* The worker doesn't emit `Queue is empty` message if
   `settings.CELERYD_EMPTY_MSG_EMIT_EVERY` is 0.
 
 .. _version-0.1.7:
@@ -1778,14 +1779,14 @@ arguments, so be sure to flush your task queue before you upgrade.
 
 * `autodiscover()` now works with zipped eggs.
 
-* celeryd: Now adds current working directory to `sys.path` for
+* Worker: Now adds current working directory to `sys.path` for
   convenience.
 
 * The `run_every` attribute of `PeriodicTask` classes can now be a
   `datetime.timedelta()` object.
 
-* celeryd: You can now set the `DJANGO_PROJECT_DIR` variable
-  for `celeryd` and it will add that to `sys.path` for easy launching.
+* Worker: You can now set the `DJANGO_PROJECT_DIR` variable
+  for the worker and it will add that to `sys.path` for easy launching.
 
 * Can now check if a task has been executed or not via HTTP.
 

+ 26 - 26
docs/history/changelog-2.0.rst

@@ -18,10 +18,10 @@
 Fixes
 -----
 
-* celeryd: Properly handle connection errors happening while
+* Worker: Properly handle connection errors happening while
   closing consumers.
 
-* celeryd: Events are now buffered if the connection is down,
+* Worker: Events are now buffered if the connection is down,
   then sent when the connection is re-established.
 
 * No longer depends on the :mod:`mailer` package.
@@ -73,7 +73,7 @@ Fixes
 
 * Can now store result/metadata for custom states.
 
-* celeryd: A warning is now emitted if the sending of task error
+* Worker: A warning is now emitted if the sending of task error
   emails fails.
 
 * celeryev: Curses monitor no longer crashes if the terminal window
@@ -81,7 +81,7 @@ Fixes
 
     See issue #160.
 
-* celeryd: On OS X it is not possible to run `os.exec*` in a process
+* Worker: On OS X it is not possible to run `os.exec*` in a process
   that is threaded.
 
       This breaks the SIGHUP restart handler,
@@ -104,7 +104,7 @@ Fixes
 
     See issue #163.
 
-* Debian init scripts: Use the absolute path of celeryd to allow stat
+* Debian init scripts: Use the absolute path of celeryd program to allow stat
 
     See issue #162.
 
@@ -173,18 +173,18 @@ Documentation
 
 * Experimental Cassandra backend added.
 
-* celeryd: SIGHUP handler accidentally propagated to worker pool processes.
+* Worker: SIGHUP handler accidentally propagated to worker pool processes.
 
     In combination with 7a7c44e39344789f11b5346e9cc8340f5fe4846c
-    this would make each child process start a new celeryd when
+    this would make each child process start a new worker instance when
     the terminal window was closed :/
 
-* celeryd: Do not install SIGHUP handler if running from a terminal.
+* Worker: Do not install SIGHUP handler if running from a terminal.
 
-    This fixes the problem where celeryd is launched in the background
+    This fixes the problem where the worker is launched in the background
     when closing the terminal.
 
-* celeryd: Now joins threads at shutdown.
+* Worker: Now joins threads at shutdown.
 
     See issue #152.
 
@@ -193,7 +193,7 @@ Documentation
 
     See issue #154.
 
-* Debian init script for celeryd: Stop now works correctly.
+* Debian worker init script: Stop now works correctly.
 
 * Task logger: `warn` method added (synonym for `warning`)
 
@@ -205,7 +205,7 @@ Documentation
 
     See issue #153.
 
-* celeryd: Now handles overflow exceptions in `time.mktime` while parsing
+* Worker: Now handles overflow exceptions in `time.mktime` while parsing
   the ETA field.
 
 * LoggerWrapper: Try to detect loggers logging back to stderr/stdout making
@@ -258,7 +258,7 @@ Documentation
 * Functional test suite added
 
     :mod:`celery.tests.functional.case` contains utilities to start
-    and stop an embedded celeryd process, for use in functional testing.
+    and stop an embedded worker process, for use in functional testing.
 
 .. _version-2.0.1:
 
@@ -382,7 +382,7 @@ Documentation
 
 * Added experimental support for persistent revokes.
 
-    Use the `-S|--statedb` argument to celeryd to enable it:
+    Use the `-S|--statedb` argument to the worker to enable it:
 
     .. code-block:: bash
 
@@ -539,7 +539,7 @@ Backward incompatible changes
 * Default (python) loader now prints warning on missing `celeryconfig.py`
   instead of raising :exc:`ImportError`.
 
-    celeryd raises :exc:`~@ImproperlyConfigured` if the configuration
+    The worker raises :exc:`~@ImproperlyConfigured` if the configuration
     is not set up. This makes it possible to use `--help` etc., without having a
     working configuration.
 
@@ -660,7 +660,7 @@ News
     * :ref:`guide-canvas`
     * :ref:`guide-routing`
 
-* celeryd: Standard out/error is now being redirected to the log file.
+* Worker: Standard out/error is now being redirected to the log file.
 
 * :mod:`billiard` has been moved back to the celery repository.
 
@@ -678,9 +678,9 @@ News
 
 * now depends on :mod:`pyparsing`
 
-* celeryd: Added `--purge` as an alias to `--discard`.
+* Worker: Added `--purge` as an alias to `--discard`.
 
-* celeryd: Ctrl+C (SIGINT) once does warm shutdown, hitting Ctrl+C twice
+* Worker: Ctrl+C (SIGINT) once does warm shutdown, hitting Ctrl+C twice
   forces termination.
 
 * Added support for using complex crontab-expressions in periodic tasks. For
@@ -694,7 +694,7 @@ News
 
   See :ref:`guide-beat`.
 
-* celeryd: Now waits for available pool processes before applying new
+* Worker: Now waits for available pool processes before applying new
   tasks to the pool.
 
     This means it doesn't have to wait for dozens of tasks to finish at shutdown
@@ -736,7 +736,7 @@ News
                                "routing_key": "name}
 
    This feature is added for easily setting up routing using the `-Q`
-   option to `celeryd`:
+   option to the worker:
 
    .. code-block:: bash
 
@@ -858,7 +858,7 @@ News
    :meth:`~celery.task.base.Task.on_retry`/
    :meth:`~celery.task.base.Task.on_failure` as einfo keyword argument.
 
-* celeryd: Added :setting:`CELERYD_MAX_TASKS_PER_CHILD` /
+* Worker: Added :setting:`CELERYD_MAX_TASKS_PER_CHILD` /
   :option:`--maxtasksperchild`
 
     Defines the maximum number of tasks a pool worker can process before
@@ -875,19 +875,19 @@ News
 * New signal: :signal:`~celery.signals.worker_process_init`: Sent inside the
   pool worker process at init.
 
-* celeryd :option:`-Q` option: Ability to specify list of queues to use,
+* Worker: :option:`-Q` option: Ability to specify list of queues to use,
   disabling other configured queues.
 
     For example, if :setting:`CELERY_QUEUES` defines four
     queues: `image`, `video`, `data` and `default`, the following
-    command would make celeryd only consume from the `image` and `video`
+    command would make the worker only consume from the `image` and `video`
     queues:
 
     .. code-block:: bash
 
         $ celeryd -Q image,video
 
-* celeryd: New return value for the `revoke` control command:
+* Worker: New return value for the `revoke` control command:
 
     Now returns::
 
@@ -895,7 +895,7 @@ News
 
     instead of `True`.
 
-* celeryd: Can now enable/disable events using remote control
+* Worker: Can now enable/disable events using remote control
 
     Example usage:
 
@@ -936,7 +936,7 @@ News
 
     The coverage output is then located in `celery/tests/cover/index.html`.
 
-* celeryd: New option `--version`: Dump version info and exit.
+* Worker: New option `--version`: Dump version info and exit.
 
 * :mod:`celeryd-multi <celeryd.bin.multi>`: Tool for shell scripts
   to start multiple workers.

+ 30 - 30
docs/history/changelog-2.1.rst

@@ -22,10 +22,10 @@ Fixes
   returned by active routers.  This was a regression introduced recently
   (Issue #244).
 
-* `celeryev` curses monitor: Long arguments are now truncated so curses
+* curses monitor: Long arguments are now truncated so curses
   doesn't crash with out of bounds errors.  (Issue #235).
 
-* `celeryd`: Channel errors occurring while handling control commands no
+* multi: Channel errors occurring while handling control commands no
   longer crash the worker but are instead logged with severity error.
 
 * SQLAlchemy database backend: Fixed a race condition occurring when
@@ -43,9 +43,9 @@ Fixes
 
 * Unit test output no longer emits non-standard characters.
 
-* `celeryd`: The broadcast consumer is now closed if the connection is reset.
+* worker: The broadcast consumer is now closed if the connection is reset.
 
-* `celeryd`: Now properly handles errors occurring while trying to acknowledge
+* worker: Now properly handles errors occurring while trying to acknowledge
   the message.
 
 * `TaskRequest.on_failure` now encodes traceback using the current filesystem
@@ -86,7 +86,7 @@ Documentation
 * Fixed pickling errors when pickling :class:`AsyncResult` on older Python
   versions.
 
-* celeryd: prefetch count was decremented by eta tasks even if there
+* worker: prefetch count was decremented by eta tasks even if there
   were no active prefetch limits.
 
 
@@ -101,15 +101,15 @@ Documentation
 Fixes
 -----
 
-* celeryd: Now sends the :event:`task-retried` event for retried tasks.
+* worker: Now sends the :event:`task-retried` event for retried tasks.
 
-* celeryd: Now honors ignore result for
+* worker: Now honors ignore result for
   :exc:`~@WorkerLostError` and timeout errors.
 
 * celerybeat: Fixed :exc:`UnboundLocalError` in celerybeat logging
   when using logging setup signals.
 
-* celeryd: All log messages now includes `exc_info`.
+* worker: All log messages now includes `exc_info`.
 
 .. _version-2.1.1:
 
@@ -128,7 +128,7 @@ Fixes
 
 * snapshots: Fixed race condition leading to loss of events.
 
-* celeryd: Reject tasks with an eta that cannot be converted to a time stamp.
+* worker: Reject tasks with an eta that cannot be converted to a time stamp.
 
     See issue #209
 
@@ -144,10 +144,10 @@ Fixes
 
 * control command `dump_scheduled`: was using old .info attribute
 
-* :program:`celeryd-multi`: Fixed `set changed size during iteration` bug
+* multi: Fixed `set changed size during iteration` bug
     occurring in the restart command.
 
-* celeryd: Accidentally tried to use additional command-line arguments.
+* worker: Accidentally tried to use additional command-line arguments.
 
    This would lead to an error like:
 
@@ -178,8 +178,8 @@ News
 * Added :setting:`CELERY_REDIRECT_STDOUTS`, and
   :setting:`CELERYD_REDIRECT_STDOUTS_LEVEL` settings.
 
-    :setting:`CELERY_REDIRECT_STDOUTS` is used by :program:`celeryd` and
-    :program:`celerybeat`.  All output to `stdout` and `stderr` will be
+    :setting:`CELERY_REDIRECT_STDOUTS` is used by the worker and
+    beat.  All output to `stdout` and `stderr` will be
     redirected to the current logger if enabled.
 
     :setting:`CELERY_REDIRECT_STDOUTS_LEVEL` decides the log level used and is
@@ -323,7 +323,7 @@ News
 
 * celeryev: Event Snapshots
 
-    If enabled, :program:`celeryd` sends messages about what the worker is doing.
+    If enabled, the worker sends messages about what the worker is doing.
     These messages are called "events".
     The events are used by real-time monitors to show what the
     cluster is doing, but they are not very useful for monitoring
@@ -357,7 +357,7 @@ News
     lets you perform some actions, like revoking and rate limiting tasks,
     and shutting down worker nodes.
 
-    There's also a Debian init.d script for :mod:`~celery.bin.celeryev` available,
+    There's also a Debian init.d script for :mod:`~celery.bin.events` available,
     see :ref:`daemonizing` for more information.
 
     New command-line arguments to celeryev:
@@ -457,7 +457,7 @@ News
     will be configured using the :option:`--loglevel`/:option:`--logfile`
     argument, this will be used for *all defined loggers*.
 
-    Remember that celeryd also redirects stdout and stderr
+    Remember that the worker also redirects stdout and stderr
     to the celery logger, if manually configure logging
     you also need to redirect the stdouts manually:
 
@@ -472,7 +472,7 @@ News
             stdouts = logging.getLogger("mystdoutslogger")
             log.redirect_stdouts_to_logger(stdouts, loglevel=logging.WARNING)
 
-* celeryd: Added command-line option :option:`-I`/:option:`--include`:
+* worker: Added command-line option :option:`-I`/:option:`--include`:
 
     A comma separated list of (task) modules to be imported.
 
@@ -482,15 +482,15 @@ News
 
         $ celeryd -I app1.tasks,app2.tasks
 
-* celeryd: now emits a warning if running as the root user (euid is 0).
+* worker: now emits a warning if running as the root user (euid is 0).
 
 * :func:`celery.messaging.establish_connection`: Ability to override defaults
   used using keyword argument "defaults".
 
-* celeryd: Now uses `multiprocessing.freeze_support()` so that it should work
+* worker: Now uses `multiprocessing.freeze_support()` so that it should work
   with **py2exe**, **PyInstaller**, **cx_Freeze**, etc.
 
-* celeryd: Now includes more metadata for the :state:`STARTED` state: PID and
+* worker: Now includes more metadata for the :state:`STARTED` state: PID and
   host name of the worker that started the task.
 
     See issue #181
@@ -508,7 +508,7 @@ News
 
     See issue #182.
 
-* celeryd: Now emits a warning if there is already a worker node using the same
+* worker: Now emits a warning if there is already a worker node using the same
   name running on the same virtual host.
 
 * AMQP result backend: Sending of results are now retried if the connection
@@ -533,11 +533,11 @@ News
 * The crontab scheduler no longer wakes up every second, but implements
   `remaining_estimate` (*Optimization*).
 
-* celeryd:  Store :state:`FAILURE` result if the
+* worker:  Store :state:`FAILURE` result if the
    :exc:`~@WorkerLostError` exception occurs (worker process
    disappeared).
 
-* celeryd: Store :state:`FAILURE` result if one of the `*TimeLimitExceeded`
+* worker: Store :state:`FAILURE` result if one of the `*TimeLimitExceeded`
   exceptions occurs.
 
 * Refactored the periodic task responsible for cleaning up results.
@@ -596,7 +596,7 @@ News
 
 * unit tests: Don't leave threads running at tear down.
 
-* celeryd: Task results shown in logs are now truncated to 46 chars.
+* worker: Task results shown in logs are now truncated to 46 chars.
 
 * `Task.__name__` is now an alias to `self.__class__.__name__`.
    This way tasks introspects more like regular functions.
@@ -634,7 +634,7 @@ News
 * :meth:`EventReceiver.capture <celery.events.EventReceiver.capture>`
   Now supports a timeout keyword argument.
 
-* celeryd: The mediator thread is now disabled if
+* worker: The mediator thread is now disabled if
   :setting:`CELERY_RATE_LIMITS` is enabled, and tasks are directly sent to the
   pool without going through the ready queue (*Optimization*).
 
@@ -653,7 +653,7 @@ Fixes
 
     See issue #187.
 
-* celeryd no longer marks tasks as revoked if :setting:`CELERY_IGNORE_RESULT`
+* the worker no longer marks tasks as revoked if :setting:`CELERY_IGNORE_RESULT`
   is enabled.
 
     See issue #207.
@@ -684,9 +684,9 @@ Fixes
 Experimental
 ------------
 
-* celeryd-multi: Added daemonization support.
+* multi: Added daemonization support.
 
-    celeryd-multi can now be used to start, stop and restart worker nodes:
+    multi can now be used to start, stop and restart worker nodes:
 
     .. code-block:: bash
 
@@ -724,12 +724,12 @@ Experimental
 
     See `celeryd-multi help` for help.
 
-* celeryd-multi: `start` command renamed to `show`.
+* multi: `start` command renamed to `show`.
 
     `celeryd-multi start` will now actually start and detach worker nodes.
     To just generate the commands you have to use `celeryd-multi show`.
 
-* celeryd: Added `--pidfile` argument.
+* worker: Added `--pidfile` argument.
 
    The worker will write its pid when it starts.  The worker will
    not be started if this file exists and the pid contained is still alive.

+ 31 - 31
docs/history/changelog-2.2.rst

@@ -21,8 +21,8 @@ Security Fixes
 
 * [Security: `CELERYSA-0001`_] Daemons would set effective id's rather than
   real id's when the :option:`--uid`/:option:`--gid` arguments to
-  :program:`celeryd-multi`, :program:`celeryd_detach`,
-  :program:`celerybeat` and :program:`celeryev` were used.
+  :program:`celery multi`, :program:`celeryd_detach`,
+  :program:`celery beat` and :program:`celery events` were used.
 
   This means privileges weren't properly dropped, and that it would
   be possible to regain supervisor privileges later.
@@ -45,9 +45,9 @@ Security Fixes
 
 * Redis result backend now works with Redis 2.4.4.
 
-* celeryd_multi: The :option:`--gid` option now works correctly.
+* multi: The :option:`--gid` option now works correctly.
 
-* celeryd: Retry wrongfully used the repr of the traceback instead
+* worker: Retry wrongfully used the repr of the traceback instead
   of the string representation.
 
 * App.config_from_object: Now loads module, not attribute of module.
@@ -108,7 +108,7 @@ Fixes
 * ``AsyncResult.get`` did not accept the ``interval`` and ``propagate``
    arguments.
 
-* celeryd: Fixed a bug where celeryd would not shutdown if a
+* worker: Fixed a bug where the worker would not shutdown if a
    :exc:`socket.error` was raised.
 
 .. _version-2.2.5:
@@ -163,7 +163,7 @@ News
 * New :setting:`BROKER_TRANSPORT_OPTIONS` setting can be used to pass
   additional arguments to a particular broker transport.
 
-* celeryd: ``worker_pid`` is now part of the request info as returned by
+* worker: ``worker_pid`` is now part of the request info as returned by
   broadcast commands.
 
 * TaskSet.apply/Taskset.apply_async now accepts an optional ``taskset_id``
@@ -213,13 +213,13 @@ Fixes
 * Internal module ``celery.worker.controllers`` renamed to
   ``celery.worker.mediator``.
 
-* celeryd: Threads now terminates the program by calling ``os._exit``, as it
+* worker: Threads now terminates the program by calling ``os._exit``, as it
   is the only way to ensure exit in the case of syntax errors, or other
   unrecoverable errors.
 
 * Fixed typo in ``maybe_timedelta`` (Issue #352).
 
-* celeryd: Broadcast commands now logs with loglevel debug instead of warning.
+* worker: Broadcast commands now logs with loglevel debug instead of warning.
 
 * AMQP Result Backend: Now resets cached channel if the connection is lost.
 
@@ -242,7 +242,7 @@ Fixes
 * Autoscaler: The "all processes busy" log message is now severity debug
   instead of error.
 
-* celeryd: If the message body can't be decoded, it is now passed through
+* worker: If the message body can't be decoded, it is now passed through
   ``safe_str`` when logging.
 
     This to ensure we don't get additional decoding errors when trying to log
@@ -257,7 +257,7 @@ Fixes
 * :mod:`celery.contrib.batches`: Now sets loglevel and logfile in the task
   request so ``task.get_logger`` works with batch tasks (Issue #357).
 
-* celeryd: An exception was raised if using the amqp transport and the prefetch
+* worker: An exception was raised if using the amqp transport and the prefetch
   count value exceeded 65535 (Issue #359).
 
     The prefetch count is incremented for every received task with an
@@ -271,7 +271,7 @@ Fixes
 * eventlet/gevent is now imported on demand so autodoc can import the modules
   without having eventlet/gevent installed.
 
-* celeryd: Ack callback now properly handles ``AttributeError``.
+* worker: Ack callback now properly handles ``AttributeError``.
 
 * ``Task.after_return`` is now always called *after* the result has been
   written.
@@ -305,7 +305,7 @@ Fixes
 Fixes
 -----
 
-* celeryd: 2.2.3 broke error logging, resulting in tracebacks not being logged.
+* worker: 2.2.3 broke error logging, resulting in tracebacks not being logged.
 
 * AMQP result backend: Polling task states did not work properly if there were
   more than one result message in the queue.
@@ -373,7 +373,7 @@ Fixes
     structure: the exchange key is now a dictionary containing the
     exchange declaration in full.
 
-* The :option:`-Q` option to :program:`celeryd` removed unused queue
+* The :option:`-Q` option to :program:`celery worker` removed unused queue
   declarations, so routing of tasks could fail.
 
     Queues are no longer removed, but rather `app.amqp.queues.consume_from()`
@@ -403,8 +403,8 @@ Fixes
 
     Previously it was overwritten by the countdown argument.
 
-* celeryd-multi/celeryd_detach: Now logs errors occuring when executing
-  the `celeryd` command.
+* celery multi/celeryd_detach: Now logs errors occuring when executing
+  the `celery worker` command.
 
 * daemonizing tutorial: Fixed typo ``--time-limit 300`` ->
   ``--time-limit=300``
@@ -431,7 +431,7 @@ Fixes
 
 * ``BasePool.on_terminate`` stub did not exist
 
-* celeryd detach: Adds readable error messages if user/group name does not
+* celeryd_detach: Adds readable error messages if user/group name does not
    exist.
 
 * Smarter handling of unicode decod errors when logging errors.
@@ -562,7 +562,7 @@ Important Notes
     This is great news for I/O-bound tasks!
 
     To change pool implementations you use the :option:`-P|--pool` argument
-    to :program:`celeryd`, or globally using the
+    to :program:`celery worker`, or globally using the
     :setting:`CELERYD_POOL` setting.  This can be the full name of a class,
     or one of the following aliases: `processes`, `eventlet`, `gevent`.
 
@@ -600,7 +600,7 @@ Important Notes
     to a newer version of Python, you can just continue to use Celery 2.2.
     Important fixes can be backported for as long as there is interest.
 
-* `celeryd`: Now supports Autoscaling of child worker processes.
+* worker: Now supports Autoscaling of child worker processes.
 
     The :option:`--autoscale` option can be used to configure the minimum
     and maximum number of child worker processes::
@@ -645,7 +645,7 @@ Important Notes
     to enable access from the outside you have to set the environment
     variable :envvar:`CELERY_RDB_HOST`.
 
-    When `celeryd` encounters your breakpoint it will log the following
+    When the worker encounters your breakpoint it will log the following
     information::
 
         [INFO/MainProcess] Got task from broker:
@@ -707,7 +707,7 @@ Important Notes
 
             $ camqadm exchange.delete celeryevent
 
-* `celeryd` now starts without configuration, and configuration can be
+* The worker now starts without configuration, and configuration can be
   specified directly on the command-line.
 
   Configuration options must appear after the last argument, separated
@@ -715,7 +715,7 @@ Important Notes
 
   .. code-block:: bash
 
-      $ celeryd -l info -I tasks -- broker.host=localhost broker.vhost=/app
+      $ celery worker -l info -I tasks -- broker.host=localhost broker.vhost=/app
 
 * Configuration is now an alias to the original configuration, so changes
   to the original will reflect Celery at runtime.
@@ -824,7 +824,7 @@ News
     But you are encouraged to use the more flexible
     :setting:`CELERYBEAT_SCHEDULE` setting.
 
-* Built-in daemonization support of celeryd using `celeryd-multi`
+* Built-in daemonization support of the worker using `celery multi`
   is no longer experimental and is considered production quality.
 
      See :ref:`daemon-generic` if you want to use the new generic init
@@ -834,7 +834,7 @@ News
   :setting:`CELERY_MESSAGE_COMPRESSION` setting, or the `compression` argument
   to `apply_async`.  This can also be set using routers.
 
-* `celeryd`: Now logs stacktrace of all threads when receiving the
+* worker: Now logs stacktrace of all threads when receiving the
    `SIGUSR1` signal.  (Does not work on cPython 2.4, Windows or Jython).
 
     Inspired by https://gist.github.com/737056
@@ -885,7 +885,7 @@ News
 
 * The following fields have been added to all events in the worker class:
 
-    * `sw_ident`: Name of worker software (e.g. celeryd).
+    * `sw_ident`: Name of worker software (e.g. py-celery).
     * `sw_ver`: Software version (e.g. 2.2.0).
     * `sw_sys`: Operating System (e.g. Linux, Windows, Darwin).
 
@@ -918,7 +918,7 @@ News
 
     .. code-block:: bash
 
-        $ celeryd --config=celeryconfig.py --loader=myloader.Loader
+        $ celery worker --config=celeryconfig.py --loader=myloader.Loader
 
 * Added signals: `beat_init` and `beat_embedded_init`
 
@@ -936,7 +936,7 @@ News
 * Redis result backend: Removed deprecated settings `REDIS_TIMEOUT` and
   `REDIS_CONNECT_RETRY`.
 
-* CentOS init script for :program:`celeryd` now available in `extra/centos`.
+* CentOS init script for :program:`celery worker` now available in `extra/centos`.
 
 * Now depends on `pyparsing` version 1.5.0 or higher.
 
@@ -956,7 +956,7 @@ Fixes
 * AMQP Backend: Exceptions occurring while sending task results are now
   propagated instead of silenced.
 
-    `celeryd` will then show the full traceback of these errors in the log.
+    the worker will then show the full traceback of these errors in the log.
 
 * AMQP Backend: No longer deletes the result queue after successful
   poll, as this should be handled by the
@@ -964,7 +964,7 @@ Fixes
 
 * AMQP Backend: Now ensures queues are declared before polling results.
 
-* Windows: celeryd: Show error if running with `-B` option.
+* Windows: worker: Show error if running with `-B` option.
 
     Running celerybeat embedded is known not to work on Windows, so
     users are encouraged to run celerybeat as a separate service instead.
@@ -985,14 +985,14 @@ Fixes
 Experimental
 ------------
 
-* Jython: celeryd now runs on Jython using the threaded pool.
+* Jython: worker now runs on Jython using the threaded pool.
 
     All tests pass, but there may still be bugs lurking around the corners.
 
-* PyPy: celeryd now runs on PyPy.
+* PyPy: worker now runs on PyPy.
 
     It runs without any pool, so to get parallel execution you must start
-    multiple instances (e.g. using :program:`celeryd-multi`).
+    multiple instances (e.g. using :program:`multi`).
 
     Sadly an initial benchmark seems to show a 30% performance decrease on
     pypy-1.4.1 + JIT.  We would like to find out why this is, so stay tuned.

+ 7 - 7
docs/history/changelog-2.3.rst

@@ -21,8 +21,8 @@ Security Fixes
 
 * [Security: `CELERYSA-0001`_] Daemons would set effective id's rather than
   real id's when the :option:`--uid`/:option:`--gid` arguments to
-  :program:`celeryd-multi`, :program:`celeryd_detach`,
-  :program:`celerybeat` and :program:`celeryev` were used.
+  :program:`celery multi`, :program:`celeryd_detach`,
+  :program:`celery beat` and :program:`celery events` were used.
 
   This means privileges weren't properly dropped, and that it would
   be possible to regain supervisor privileges later.
@@ -276,7 +276,7 @@ News
 * PyPy: The default pool implementation used is now multiprocessing
   if running on PyPy 1.5.
 
-* celeryd-multi: now supports "pass through" options.
+* multi: now supports "pass through" options.
 
     Pass through options makes it easier to use celery without a
     configuration file, or just add last-minute options on the command
@@ -286,9 +286,9 @@ News
 
     .. code-block:: bash
 
-        $ celeryd-multi start 4  -c 2  -- broker.host=amqp.example.com \
-                                          broker.vhost=/               \
-                                          celery.disable_rate_limits=yes
+        $ celery multi start 4  -c 2  -- broker.host=amqp.example.com \
+                                         broker.vhost=/               \
+                                         celery.disable_rate_limits=yes
 
 * celerybeat: Now retries establishing the connection (Issue #419).
 
@@ -336,7 +336,7 @@ News
 * Terminating a task on Windows now also terminates all of the tasks child
   processes (Issue #384).
 
-* celeryd: ``-I|--include`` option now always searches the current directory
+* worker: ``-I|--include`` option now always searches the current directory
   to import the specified modules.
 
 * Cassandra backend: Now expires results by using TTLs.

+ 15 - 15
docs/history/changelog-2.4.rst

@@ -17,7 +17,7 @@
 * Periodic task interval schedules were accidentally rounded down,
   resulting in some periodic tasks being executed early.
 
-* Logging of humanized times in the celerybeat log is now more detailed.
+* Logging of humanized times in the beat log is now more detailed.
 
 * New :ref:`brokers` section in the Getting Started part of the Documentation
 
@@ -38,8 +38,8 @@ Security Fixes
 
 * [Security: `CELERYSA-0001`_] Daemons would set effective id's rather than
   real id's when the :option:`--uid`/:option:`--gid` arguments to
-  :program:`celeryd-multi`, :program:`celeryd_detach`,
-  :program:`celerybeat` and :program:`celeryev` were used.
+  :program:`celery multi`, :program:`celeryd_detach`,
+  :program:`celery beat` and :program:`celery events` were used.
 
   This means privileges weren't properly dropped, and that it would
   be possible to regain supervisor privileges later.
@@ -110,7 +110,7 @@ Fixes
 * processes pool: MaybeEncodingError was not wrapped in ExceptionInfo
   (Issue #524).
 
-* celeryd: would silence errors occuring after task consumer started.
+* worker: would silence errors occuring after task consumer started.
 
 * logging: Fixed a bug where unicode in stdout redirected log messages
   couldn't be written (Issue #522).
@@ -207,9 +207,9 @@ Important Notes
 
     .. code-block:: bash
 
-        $ celeryd -b redis://localhost
+        $ celery worker -b redis://localhost
 
-        $ celeryctl -b amqp://guest:guest@localhost//e
+        $ celery inspect -b amqp://guest:guest@localhost//e
 
     The environment variable :envvar:`CELERY_BROKER_URL` can also be used to
     easily override the default broker used.
@@ -250,10 +250,10 @@ Important Notes
     =====================================  ===================================
     **Old setting**                        **Alternative**
     =====================================  ===================================
-    `CELERYD_LOG_LEVEL`                    ``celeryd --loglevel=``
-    `CELERYD_LOG_FILE`                     ``celeryd --logfile=``
-    `CELERYBEAT_LOG_LEVEL`                 ``celerybeat --loglevel=``
-    `CELERYBEAT_LOG_FILE`                  ``celerybeat --logfile=``
+    `CELERYD_LOG_LEVEL`                    ``celery worker --loglevel=``
+    `CELERYD_LOG_FILE`                     ``celery worker --logfile=``
+    `CELERYBEAT_LOG_LEVEL`                 ``celery beat --loglevel=``
+    `CELERYBEAT_LOG_FILE`                  ``celery beat --logfile=``
     `CELERYMON_LOG_LEVEL`                  ``celerymon --loglevel=``
     `CELERYMON_LOG_FILE`                   ``celerymon --logfile=``
     =====================================  ===================================
@@ -318,7 +318,7 @@ News
 
     Contributed by Chris Adams.
 
-* ``celeryd_multi`` now supports a ``stop_verify`` command to wait for
+* ``multi`` now supports a ``stop_verify`` command to wait for
   processes to shutdown.
 
 * Cache backend did not work if the cache key was unicode (Issue #504).
@@ -336,14 +336,14 @@ News
 
     Fix contributed by Remy Noel
 
-* celeryd-multi did not work on Windows (Issue #472).
+* multi did not work on Windows (Issue #472).
 
 * New-style ``CELERY_REDIS_*`` settings now takes precedence over
   the old ``REDIS_*`` configuration keys (Issue #508).
 
     Fix contributed by Joshua Ginsberg
 
-* Generic celerybeat init script no longer sets `bash -e` (Issue #510).
+* Generic beat init script no longer sets `bash -e` (Issue #510).
 
     Fix contributed by Roger Hu.
 
@@ -364,7 +364,7 @@ News
 * Worker logged the string representation of args and kwargs
   without safe guards (Issue #480).
 
-* RHEL init script: Changed celeryd startup priority.
+* RHEL init script: Changed worker startup priority.
 
     The default start / stop priorities for MySQL on RHEL are
 
@@ -385,7 +385,7 @@ News
 * KeyValueStoreBackend.get_many did not respect the ``timeout`` argument
   (Issue #512).
 
-* celerybeat/celeryev's --workdir option did not chdir before after
+* beat/events's --workdir option did not chdir before after
   configuration was attempted (Issue #506).
 
 * After deprecating 2.4 support we can now name modules correctly, since we

+ 8 - 8
docs/history/changelog-2.5.rst

@@ -116,7 +116,7 @@ Fixes
 
         echo -n "1" > celeryd.pid
 
-    would cause celeryd to think that an existing instance was already
+    would cause the worker to think that an existing instance was already
     running (init has pid 1 after all).
 
 - Fixed 2.5 compatibility issue with use of print_exception.
@@ -133,14 +133,14 @@ Fixes
 
     Fix contributed by Martin Melin.
 
-- celeryctl can now be configured on the command-line.
+- [celery control|inspect] can now be configured on the command-line.
 
-    Like with celeryd it is now possible to configure celery settings
-    on the command-line for celeryctl:
+    Like with the worker it is now possible to configure celery settings
+    on the command-line for celery control|inspect
 
     .. code-block:: bash
 
-        $ celeryctl -- broker.pool_limit=30
+        $ celery inspect -- broker.pool_limit=30
 
 - Version dependency for python-dateutil fixed to be strict.
 
@@ -171,12 +171,12 @@ Fixes
 Fixes
 -----
 
-* Eventlet/Gevent: A small typo caused celeryd to hang when eventlet/gevent
+* Eventlet/Gevent: A small typo caused the worker to hang when eventlet/gevent
   was used, this was because the environment was not monkey patched
   early enough.
 
 * Eventlet/Gevent: Another small typo caused the mediator to be started
-  with eventlet/gevent, which would make celeryd sometimes hang at shutdown.
+  with eventlet/gevent, which would make the worker sometimes hang at shutdown.
 
 * Mulitprocessing: Fixed an error occurring if the pool was stopped
   before it was properly started.
@@ -187,7 +187,7 @@ Fixes
 * Internal timer (timer2) now logs exceptions instead of swallowing them
   (Issue #626).
 
-* celeryctl shell: can now be started with :option:`--eventlet` or
+* celery shell: can now be started with :option:`--eventlet` or
   :option:`--gevent` options to apply their monkey patches.
 
 .. _version-2.5.0:

+ 3 - 3
docs/reference/celery.bin.events.rst

@@ -1,11 +1,11 @@
 =====================================================
- celery.bin.celeryev
+ celery.bin.events
 =====================================================
 
 .. contents::
     :local:
-.. currentmodule:: celery.bin.celeryev
+.. currentmodule:: celery.bin.events
 
-.. automodule:: celery.bin.celeryev
+.. automodule:: celery.bin.events
     :members:
     :undoc-members:

+ 3 - 3
docs/reference/celery.bin.celeryd.rst → docs/reference/celery.bin.worker.rst

@@ -1,11 +1,11 @@
 ==========================================
- celery.bin.celeryd
+ celery.bin.worker
 ==========================================
 
 .. contents::
     :local:
-.. currentmodule:: celery.bin.celeryd
+.. currentmodule:: celery.bin.worker
 
-.. automodule:: celery.bin.celeryd
+.. automodule:: celery.bin.worker
     :members:
     :undoc-members:

+ 1 - 1
docs/reference/celery.rst

@@ -230,7 +230,7 @@ Application
 
     .. method:: Celery.worker_main(argv=None)
 
-        Run :program:`celeryd` using `argv`.
+        Run :program:`celery worker` using `argv`.
 
         Uses :data:`sys.argv` if `argv` is not specified."""
 

+ 1 - 1
docs/reference/index.rst

@@ -46,7 +46,7 @@
     celery.apps.worker
     celery.apps.beat
     celery.bin.base
-    celery.bin.celeryd
+    celery.bin.worker
     celery.bin.beat
     celery.bin.events
     celery.bin.celery

+ 24 - 25
docs/tutorials/daemonizing.rst

@@ -18,8 +18,9 @@ Generic init scripts
 
 See the `extra/generic-init.d/`_ directory Celery distribution.
 
-This directory contains generic bash init scripts for :program:`celeryd`,
-that should run on Linux, FreeBSD, OpenBSD, and other Unix platforms.
+This directory contains generic bash init scripts for the
+:program:`celery worker` program,
+these should run on Linux, FreeBSD, OpenBSD, and other Unix-like platforms.
 
 .. _`extra/generic-init.d/`:
     http://github.com/celery/celery/tree/3.0/extra/generic-init.d/
@@ -32,8 +33,9 @@ Init script: celeryd
 :Usage: `/etc/init.d/celeryd {start|stop|restart|status}`
 :Configuration file: /etc/default/celeryd
 
-To configure celeryd you probably need to at least tell it where to change
-directory to when it starts (to find your `celeryconfig`).
+To configure this script to run the worker properly you probably need to at least tell it where to change
+directory to when it starts (to find the module containing your app, or your
+configuration module).
 
 .. _generic-initd-celeryd-example:
 
@@ -65,15 +67,15 @@ This is an example configuration for a Python project.
     # Where to chdir at start.
     CELERYD_CHDIR="/opt/Myproject/"
 
-    # Extra arguments to celeryd
+    # Extra command-line arguments to the worker
     CELERYD_OPTS="--time-limit=300 --concurrency=8"
 
     # Name of the celery config module.
     CELERY_CONFIG_MODULE="celeryconfig"
 
-    # %n will be replaced with the nodename.
-    CELERYD_LOG_FILE="/var/log/celery/%n.log"
-    CELERYD_PID_FILE="/var/run/celery/%n.pid"
+    # %N will be replaced with the first part of the nodename.
+    CELERYD_LOG_FILE="/var/log/celery/%N.log"
+    CELERYD_PID_FILE="/var/run/celery/%N.pid"
 
     # Workers should run as an unprivileged user.
     CELERYD_USER="celery"
@@ -99,7 +101,7 @@ This is an example configuration for those using `django-celery`:
     # How to call "manage.py celery"
     CELERY_BIN="$CELERYD_CHDIR/manage.py celery"
 
-    # Extra arguments to celeryd
+    # Extra command-line arguments for the worker (see celery worker --help).
     CELERYD_OPTS="--time-limit=300 --concurrency=8"
 
     # %n will be replaced with the nodename.
@@ -137,7 +139,7 @@ environment's python interpreter:
     # How to call "manage.py celery"
     CELERY_BIN="$ENV_PYTHON $CELERYD_CHDIR/manage.py celery"
 
-    # Extra arguments to celeryd
+    # Extra command-line arguments to the worker (see celery worker --help)
     CELERYD_OPTS="--time-limit=300 --concurrency=8"
 
     # Name of the celery config module.
@@ -175,26 +177,27 @@ Available options
     Node names to start.
 
 * CELERYD_OPTS
-    Additional arguments to celeryd, see `celeryd --help` for a list.
+    Additional command-line arguments for the worker, see
+    `celery worker --help` for a list.
 
 * CELERYD_CHDIR
     Path to change directory to at start. Default is to stay in the current
     directory.
 
 * CELERYD_PID_FILE
-    Full path to the PID file. Default is /var/run/celeryd%n.pid
+    Full path to the PID file. Default is /var/run/celery/%N.pid
 
 * CELERYD_LOG_FILE
-    Full path to the celeryd log file. Default is /var/log/celeryd@%n.log
+    Full path to the worker log file. Default is /var/log/celery/%N.log
 
 * CELERYD_LOG_LEVEL
-    Log level to use for celeryd. Default is INFO.
+    Worker log level. Default is INFO.
 
 * CELERYD_USER
-    User to run celeryd as. Default is current user.
+    User to run the worker as. Default is current user.
 
 * CELERYD_GROUP
-    Group to run celeryd as. Default is current user.
+    Group to run worker as. Default is current user.
 
 * CELERY_CREATE_DIRS
     Always create directories (log directory and pid file directory).
@@ -294,15 +297,11 @@ Available options
 * CELERYBEAT_LOG_LEVEL
     Log level to use for celeryd. Default is INFO.
 
-* CELERYBEAT
-    Path to the celeryd program. Default is `celeryd`.
-    You can point this to an virtualenv, or even use manage.py for django.
-
 * CELERYBEAT_USER
-    User to run celeryd as. Default is current user.
+    User to run beat as. Default is current user.
 
 * CELERYBEAT_GROUP
-    Group to run celeryd as. Default is current user.
+    Group to run beat as. Default is current user.
 
 * CELERY_CREATE_DIRS
     Always create directories (log directory and pid file directory).
@@ -337,15 +336,15 @@ For example my `sh -x` output does this:
 
     ++ start-stop-daemon --start --chdir /opt/App/release/app --quiet \
         --oknodo --background --make-pidfile --pidfile /var/run/celeryd.pid \
-        --exec /opt/App/release/app/manage.py celeryd -- --time-limit=300 \
+        --exec /opt/App/release/app/manage.py celery worker -- --time-limit=300 \
         -f /var/log/celeryd.log -l INFO
 
-Run the celeryd command after `--exec` (without the `--`) to show the
+Run the worker command after `--exec` (without the `--`) to show the
 actual resulting output:
 
 .. code-block:: bash
 
-    $ /opt/App/release/app/manage.py celeryd --time-limit=300 \
+    $ /opt/App/release/app/manage.py celery worker --time-limit=300 \
         -f /var/log/celeryd.log -l INFO
 
 .. _daemon-supervisord:

+ 6 - 4
docs/userguide/monitoring.rst

@@ -20,7 +20,7 @@ features related to monitoring, like events and broadcast commands.
 Workers
 =======
 
-.. _monitoring-celeryctl:
+.. _monitoring-control:
 
 
 ``celery``: Management Command-line Utilities
@@ -150,17 +150,17 @@ Commands
 
 .. note::
 
-    All ``inspect`` commands supports a ``--timeout`` argument,
+    All ``inspect`` and ``control`` commands supports a ``--timeout`` argument,
     This is the number of seconds to wait for responses.
     You may have to increase this timeout if you're not getting a response
     due to latency.
 
-.. _celeryctl-inspect-destination:
+.. _inspect-destination:
 
 Specifying destination nodes
 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 
-By default the inspect commands operates on all workers.
+By default the inspect and control commands operates on all workers.
 You can specify a single, or a list of workers by using the
 `--destination` argument:
 
@@ -168,6 +168,8 @@ You can specify a single, or a list of workers by using the
 
     $ celery inspect -d w1,w2 reserved
 
+    $ celery control -d w1,w2 enable_events
+
 
 .. _monitoring-flower:
 

+ 1 - 1
docs/userguide/optimizing.rst

@@ -124,7 +124,7 @@ the tasks according to the run-time. (see :ref:`guide-routing`).
        all messages will be delivered to the active node.
 
 .. [*] This is the concurrency setting; :setting:`CELERYD_CONCURRENCY` or the
-       :option:`-c` option to :program:`celeryd`.
+       :option:`-c` option to the :program:`celery worker` program.
 
 
 Reserve one task at a time

+ 1 - 1
docs/userguide/signals.rst

@@ -247,7 +247,7 @@ Provides arguments:
 celeryd_init
 ~~~~~~~~~~~~
 
-This is the first signal sent when :program:`celeryd` starts up.
+This is the first signal sent when :program:`celery worker` starts up.
 The ``sender`` is the host name of the worker, so this signal can be used
 to setup worker specific configuration:
 

+ 2 - 2
docs/userguide/workers.rst

@@ -26,7 +26,7 @@ You can start the worker in the foreground by executing the command:
     $ celery worker --app=app -l info
 
 For a full list of available command-line options see
-:mod:`~celery.bin.celeryd`, or simply do:
+:mod:`~celery.bin.worker`, or simply do:
 
 .. code-block:: bash
 
@@ -154,7 +154,7 @@ Remote control
 
     The :program:`celery` program is used to execute remote control
     commands from the command-line.  It supports all of the commands
-    listed below.  See :ref:`monitoring-celeryctl` for more information.
+    listed below.  See :ref:`monitoring-control` for more information.
 
 pool support: *processes, eventlet, gevent*, blocking:*threads/solo* (see note)
 broker support: *amqp, redis, mongodb*

+ 1 - 1
examples/app/myapp.py

@@ -10,7 +10,7 @@ Usage:
    32
 
 
-You can also specify the app to use with celeryd::
+You can also specify the app to use with the celery command:
 
     $ celery worker -l info --app=myapp
 

+ 3 - 2
examples/eventlet/README.rst

@@ -13,10 +13,11 @@ module (when this is installed all name lookups will be asynchronous)::
     $ pip install eventlet
     $ pip install dnspython
 
-Before you run any of the example tasks you need to start celeryd::
+Before you run any of the example tasks you need to start
+the worker::
 
     $ cd examples/eventlet
-    $ celeryd -l info --concurrency=500 --pool=eventlet
+    $ celery worker -l info --concurrency=500 --pool=eventlet
 
 As usual you need to have RabbitMQ running, see the Celery getting started
 guide if you haven't installed it yet.

+ 2 - 2
extra/centos/celeryd.init

@@ -37,8 +37,8 @@ prog="$(basename $0)"
 
 CELERYD=${CELERYD:-"-m celery worker --detach"}
 CELERY=${CELERY:-"/usr/bin/celery"}
-CELERYD_PID_FILE=${CELERYD_PID_FILE:-"/var/run/celeryd/$prog.pid"}
-CELERYD_LOG_FILE=${CELERYD_LOG_FILE:-"/var/log/celeryd/$prog.log"}
+CELERYD_PID_FILE=${CELERYD_PID_FILE:-"/var/run/celery/$prog.pid"}
+CELERYD_LOG_FILE=${CELERYD_LOG_FILE:-"/var/log/celery/$prog.log"}
 CELERYD_LOG_LEVEL=${CELERYD_LOG_LEVEL:-"INFO"}
 
 # This is used to change how Celery loads in the configs.  It does not need to

+ 2 - 2
extra/centos/celeryd.sysconfig

@@ -1,4 +1,4 @@
-# Path to the virtualenv. This could be /usr if python 
+# Path to the virtualenv. This could be /usr if python
 #   is in /usr/bin
 #BIN_PATH="/path/to/.virtualenvs/myvenv"
 
@@ -11,7 +11,7 @@
 # Path to celery - which might be in a virtualenv
 #CELERY="$BIN_PATH/bin/celery"
 
-# Passed into celeryd multi
+# Passed into celery multi
 #CELERYD="-m celery worker --detach"
 
 # Log and PID files

+ 4 - 3
extra/mac/org.celeryq.celerybeat.plist → extra/mac/org.celeryq.beat.plist

@@ -5,15 +5,16 @@
     <key>Disabled</key>
     <false/>
     <key>GroupName</key>
-    <string>celerybeat</string>
+    <string>celery-beat</string>
     <key>KeepAlive</key>
     <true/>
     <key>Label</key>
-    <string>org.celeryq.celerybeat</string>
+    <string>org.celeryq.beat</string>
     <key>Program</key>
-    <string>celery beat</string>
+    <string>celery</string>
     <key>ProgramArguments</key>
     <array>
+        <string>beat</string>
         <string>--loglevel=WARNING</string>
     </array>
     <key>RunAtLoad</key>

+ 0 - 28
extra/mac/org.celeryq.celerymon.plist

@@ -1,28 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
-<plist version="1.0">
-<dict>
-    <key>Disabled</key>
-    <false/>
-    <key>GroupName</key>
-    <string>celerymon</string>
-    <key>KeepAlive</key>
-    <true/>
-    <key>Label</key>
-    <string>org.celeryq.celerymon</string>
-    <key>Program</key>
-    <string>celerymon</string>
-    <key>ProgramArguments</key>
-    <array>
-        <string>--loglevel=WARNING</string>
-    </array>
-    <key>RunAtLoad</key>
-    <true/>
-    <key>Umask</key>
-    <integer>7</integer>
-    <key>UserName</key>
-    <string>nobody</string>
-    <key>WorkingDirectory</key>
-    <string>/</string>
-</dict>
-</plist>

+ 4 - 3
extra/mac/org.celeryq.celeryd.plist → extra/mac/org.celeryq.worker.plist

@@ -5,15 +5,16 @@
     <key>Disabled</key>
     <false/>
     <key>GroupName</key>
-    <string>celeryd</string>
+    <string>celery-worker</string>
     <key>KeepAlive</key>
     <true/>
     <key>Label</key>
-    <string>org.celeryq.celeryd</string>
+    <string>org.celeryq.worker</string>
     <key>Program</key>
-    <string>celeryd</string>
+    <string>celery</string>
     <key>ProgramArguments</key>
     <array>
+        <string>worker</string>
         <string>--loglevel=WARNING</string>
     </array>
     <key>RunAtLoad</key>

+ 1 - 1
extra/release/core-modules.txt

@@ -18,7 +18,7 @@ celery//backends/__init__.py
 celery//backends/base.py
 celery//bin/__init__.py
 celery//bin/base.py
-celery//bin/celeryd.py
+celery//bin/worker.py
 celery//canvas.py
 celery//concurrency/__init__.py
 celery//concurrency/base.py

+ 0 - 1
extra/release/doc4allmods

@@ -7,7 +7,6 @@ SKIP_FILES="celery.five.rst
             celery.task.sets.rst
             celery.bin.rst
             celery.bin.celeryd_detach.rst
-            celery.bin.celeryctl.rst
             celery.contrib.rst
             celery.contrib.bundles.rst
             celery.local.rst

+ 7 - 5
extra/supervisord/celeryd.conf

@@ -1,6 +1,6 @@
-; ============================
-;  celeryd supervisor example
-; ============================
+; ==================================
+;  celery worker supervisor example
+; ==================================
 
 ; NOTE: If you're using Django, you shouldn't use this file.
 ; Use
@@ -8,9 +8,11 @@
 ; instead!
 
 [program:celery]
-command=celeryd --loglevel=INFO
+command=celery worker -A proj.tasks:app --loglevel=INFO
 
-; Set PYTHONPATH to the directory containing celeryconfig.py
+; remove -A argument if you don't use a celery app instance.
+
+; Set PYTHONPATH to the directory containing app/celeryconfig.py
 environment=PYTHONPATH=/path/to/project
 
 directory=/path/to/project