control.py 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375
  1. # -*- coding: utf-8 -*-
  2. """
  3. celery.worker.control
  4. ~~~~~~~~~~~~~~~~~~~~~
  5. Remote control commands.
  6. """
  7. from __future__ import absolute_import
  8. import tempfile
  9. from kombu.utils.encoding import safe_repr
  10. from celery.five import UserDict, items, StringIO
  11. from celery.platforms import signals as _signals
  12. from celery.utils import timeutils
  13. from celery.utils.functional import maybe_list
  14. from celery.utils.log import get_logger
  15. from celery.utils import jsonify
  16. from . import state as worker_state
  17. from .state import revoked
  18. from .job import Request
  19. __all__ = ['Panel']
  20. DEFAULT_TASK_INFO_ITEMS = ('exchange', 'routing_key', 'rate_limit')
  21. logger = get_logger(__name__)
  22. class Panel(UserDict):
  23. data = dict() # Global registry.
  24. @classmethod
  25. def register(cls, method, name=None):
  26. cls.data[name or method.__name__] = method
  27. return method
  28. def _find_requests_by_id(ids, requests):
  29. found, total = 0, len(ids)
  30. for request in requests:
  31. if request.id in ids:
  32. yield request
  33. found += 1
  34. if found >= total:
  35. break
  36. @Panel.register
  37. def query_task(state, ids, **kwargs):
  38. ids = maybe_list(ids)
  39. def reqinfo(state, req):
  40. return state, req.info()
  41. reqs = dict((req.id, ('reserved', req.info()))
  42. for req in _find_requests_by_id(
  43. ids, worker_state.reserved_requests))
  44. reqs.update(dict(
  45. (req.id, ('active', req.info()))
  46. for req in _find_requests_by_id(
  47. ids, worker_state.active_requests,
  48. )
  49. ))
  50. return req
  51. @Panel.register
  52. def revoke(state, task_id, terminate=False, signal=None, **kwargs):
  53. """Revoke task by task id."""
  54. # supports list argument since 3.1
  55. task_ids, task_id = set(maybe_list(task_id) or []), None
  56. size = len(task_ids)
  57. terminated = set()
  58. revoked.update(task_ids)
  59. if terminate:
  60. signum = _signals.signum(signal or 'TERM')
  61. # reserved_requests changes size during iteration
  62. # so need to consume the items first, then terminate after.
  63. requests = set(_find_requests_by_id(
  64. task_ids,
  65. worker_state.reserved_requests,
  66. ))
  67. for request in requests:
  68. if request.id not in terminated:
  69. terminated.add(request.id)
  70. logger.info('Terminating %s (%s)', request.id, signum)
  71. request.terminate(state.consumer.pool, signal=signum)
  72. if len(terminated) >= size:
  73. break
  74. if not terminated:
  75. return {'ok': 'terminate: tasks unknown'}
  76. return {'ok': 'terminate: {0}'.format(', '.join(terminated))}
  77. idstr = ', '.join(task_ids)
  78. logger.info('Tasks flagged as revoked: %s', idstr)
  79. return {'ok': 'tasks {0} flagged as revoked'.format(idstr)}
  80. @Panel.register
  81. def report(state):
  82. return {'ok': state.app.bugreport()}
  83. @Panel.register
  84. def enable_events(state):
  85. dispatcher = state.consumer.event_dispatcher
  86. if 'task' not in dispatcher.groups:
  87. dispatcher.groups.add('task')
  88. logger.info('Events of group {task} enabled by remote.')
  89. return {'ok': 'task events enabled'}
  90. return {'ok': 'task events already enabled'}
  91. @Panel.register
  92. def disable_events(state):
  93. dispatcher = state.consumer.event_dispatcher
  94. if 'task' in dispatcher.groups:
  95. dispatcher.groups.discard('task')
  96. logger.info('Events of group {task} disabled by remote.')
  97. return {'ok': 'task events disabled'}
  98. return {'ok': 'task events already disabled'}
  99. @Panel.register
  100. def heartbeat(state):
  101. logger.debug('Heartbeat requested by remote.')
  102. dispatcher = state.consumer.event_dispatcher
  103. dispatcher.send('worker-heartbeat', freq=5, **worker_state.SOFTWARE_INFO)
  104. @Panel.register
  105. def rate_limit(state, task_name, rate_limit, **kwargs):
  106. """Set new rate limit for a task type.
  107. See :attr:`celery.task.base.Task.rate_limit`.
  108. :param task_name: Type of task.
  109. :param rate_limit: New rate limit.
  110. """
  111. try:
  112. timeutils.rate(rate_limit)
  113. except ValueError as exc:
  114. return {'error': 'Invalid rate limit string: {0!r}'.format(exc)}
  115. try:
  116. state.app.tasks[task_name].rate_limit = rate_limit
  117. except KeyError:
  118. logger.error('Rate limit attempt for unknown task %s',
  119. task_name, exc_info=True)
  120. return {'error': 'unknown task'}
  121. state.consumer.reset_rate_limits()
  122. if not rate_limit:
  123. logger.info('Rate limits disabled for tasks of type %s', task_name)
  124. return {'ok': 'rate limit disabled successfully'}
  125. logger.info('New rate limit for tasks of type %s: %s.',
  126. task_name, rate_limit)
  127. return {'ok': 'new rate limit set successfully'}
  128. @Panel.register
  129. def time_limit(state, task_name=None, hard=None, soft=None, **kwargs):
  130. try:
  131. task = state.app.tasks[task_name]
  132. except KeyError:
  133. logger.error('Change time limit attempt for unknown task %s',
  134. task_name, exc_info=True)
  135. return {'error': 'unknown task'}
  136. task.soft_time_limit = soft
  137. task.time_limit = hard
  138. logger.info('New time limits for tasks of type %s: soft=%s hard=%s',
  139. task_name, soft, hard)
  140. return {'ok': 'time limits set successfully'}
  141. @Panel.register
  142. def dump_schedule(state, safe=False, **kwargs):
  143. def prepare_entries():
  144. for waiting in state.consumer.timer.schedule.queue:
  145. try:
  146. arg0 = waiting.entry.args[0]
  147. except (IndexError, TypeError):
  148. continue
  149. else:
  150. if isinstance(arg0, Request):
  151. yield {'eta': arg0.eta.isoformat() if arg0.eta else None,
  152. 'priority': waiting.priority,
  153. 'request': arg0.info(safe=safe)}
  154. return list(prepare_entries())
  155. @Panel.register
  156. def dump_reserved(state, safe=False, **kwargs):
  157. reserved = worker_state.reserved_requests - worker_state.active_requests
  158. if not reserved:
  159. return []
  160. return [request.info(safe=safe) for request in reserved]
  161. @Panel.register
  162. def dump_active(state, safe=False, **kwargs):
  163. return [request.info(safe=safe)
  164. for request in worker_state.active_requests]
  165. @Panel.register
  166. def stats(state, **kwargs):
  167. return state.consumer.controller.stats()
  168. @Panel.register
  169. def objgraph(state, num=200, max_depth=10, type='Request'): # pragma: no cover
  170. try:
  171. import objgraph
  172. except ImportError:
  173. raise ImportError('Requires the objgraph library')
  174. print('Dumping graph for type %r' % (type, ))
  175. with tempfile.NamedTemporaryFile(prefix='cobjg',
  176. suffix='.png', delete=False) as fh:
  177. objects = objgraph.by_type(type)[:num]
  178. objgraph.show_backrefs(
  179. objects,
  180. max_depth=max_depth, highlight=lambda v: v in objects,
  181. filename=fh.name,
  182. )
  183. return {'filename': fh.name}
  184. @Panel.register
  185. def memsample(state, **kwargs): # pragma: no cover
  186. from celery.utils.debug import sample_mem
  187. return sample_mem()
  188. @Panel.register
  189. def memdump(state, samples=10, **kwargs): # pragma: no cover
  190. from celery.utils.debug import memdump
  191. out = StringIO()
  192. memdump(file=out)
  193. return out.getvalue()
  194. @Panel.register
  195. def clock(state, **kwargs):
  196. return {'clock': state.app.clock.value}
  197. @Panel.register
  198. def dump_revoked(state, **kwargs):
  199. return list(worker_state.revoked)
  200. @Panel.register
  201. def hello(state, from_node, revoked=None, **kwargs):
  202. if from_node != state.hostname:
  203. logger.info('sync with %s', from_node)
  204. if revoked:
  205. worker_state.revoked.update(revoked)
  206. return {'revoked': worker_state.revoked._data,
  207. 'clock': state.app.clock.forward()}
  208. @Panel.register
  209. def dump_tasks(state, taskinfoitems=None, **kwargs):
  210. tasks = state.app.tasks
  211. taskinfoitems = taskinfoitems or DEFAULT_TASK_INFO_ITEMS
  212. def _extract_info(task):
  213. fields = dict((field, str(getattr(task, field, None)))
  214. for field in taskinfoitems
  215. if getattr(task, field, None) is not None)
  216. if fields:
  217. info = ['='.join(f) for f in items(fields)]
  218. return '{0} [{1}]'.format(task.name, ' '.join(info))
  219. return task.name
  220. return [_extract_info(tasks[task]) for task in sorted(tasks)]
  221. @Panel.register
  222. def ping(state, **kwargs):
  223. return {'ok': 'pong'}
  224. @Panel.register
  225. def pool_grow(state, n=1, **kwargs):
  226. if state.consumer.controller.autoscaler:
  227. state.consumer.controller.autoscaler.force_scale_up(n)
  228. else:
  229. state.consumer.pool.grow(n)
  230. state.consumer._update_prefetch_count(n)
  231. return {'ok': 'pool will grow'}
  232. @Panel.register
  233. def pool_shrink(state, n=1, **kwargs):
  234. if state.consumer.controller.autoscaler:
  235. state.consumer.controller.autoscaler.force_scale_down(n)
  236. else:
  237. state.consumer.pool.shrink(n)
  238. state.consumer._update_prefetch_count(-n)
  239. return {'ok': 'pool will shrink'}
  240. @Panel.register
  241. def pool_restart(state, modules=None, reload=False, reloader=None, **kwargs):
  242. if state.app.conf.CELERYD_POOL_RESTARTS:
  243. state.consumer.controller.reload(modules, reload, reloader=reloader)
  244. return {'ok': 'reload started'}
  245. else:
  246. raise ValueError('Pool restarts not enabled')
  247. @Panel.register
  248. def autoscale(state, max=None, min=None):
  249. autoscaler = state.consumer.controller.autoscaler
  250. if autoscaler:
  251. max_, min_ = autoscaler.update(max, min)
  252. return {'ok': 'autoscale now min={0} max={1}'.format(max_, min_)}
  253. raise ValueError('Autoscale not enabled')
  254. @Panel.register
  255. def shutdown(state, msg='Got shutdown from remote', **kwargs):
  256. logger.warning(msg)
  257. raise SystemExit(msg)
  258. @Panel.register
  259. def add_consumer(state, queue, exchange=None, exchange_type=None,
  260. routing_key=None, **options):
  261. state.consumer.add_task_queue(queue, exchange, exchange_type,
  262. routing_key, **options)
  263. return {'ok': 'add consumer {0}'.format(queue)}
  264. @Panel.register
  265. def cancel_consumer(state, queue=None, **_):
  266. state.consumer.cancel_task_queue(queue)
  267. return {'ok': 'no longer consuming from {0}'.format(queue)}
  268. @Panel.register
  269. def active_queues(state):
  270. """Return information about the queues a worker consumes from."""
  271. return [dict(queue.as_dict(recurse=True))
  272. for queue in state.consumer.task_consumer.queues]
  273. def _wanted_config_key(key):
  274. return key.isupper() and not key.startswith('__')
  275. @Panel.register
  276. def dump_conf(state, with_defaults=False, **kwargs):
  277. return jsonify(state.app.conf.table(with_defaults=with_defaults),
  278. keyfilter=_wanted_config_key,
  279. unknown_type_filter=safe_repr)
  280. @Panel.register
  281. def election(state, id, topic, action=None, **kwargs):
  282. state.consumer.gossip.election(id, topic, action)