Celery Tasks在Django应用程序中多次执行

时间:2019-09-20 13:38:09

标签: python django redis celery django-celery

我有一个Django应用程序,其中我在@task下定义了一些task.py函数以在给定的周期性任务上执行。我100%确定问题不是由task.py或任何相关代码引起的,而是由于settings.py或我的芹菜工人中的某些配置所致。

任务确实会在定期任务上执行,但会多次执行。

这是芹菜工作者的日志:

celery -A cimexmonitor worker --loglevel=info -B -c 4

[2019-09-19 21:22:16,360: INFO/ForkPoolWorker-5] Project Monitor Started : APPProject1
[2019-09-19 21:22:16,361: INFO/ForkPoolWorker-4] Project Monitor Started : APPProject1
[2019-09-19 21:25:22,108: INFO/ForkPoolWorker-4] Project Monitor DONE : APPProject1
[2019-09-19 21:25:45,255: INFO/ForkPoolWorker-5] Project Monitor DONE : APPProject1
[2019-09-20 00:22:16,395: INFO/ForkPoolWorker-4] Project Monitor Started : APPProject2
[2019-09-20 00:22:16,398: INFO/ForkPoolWorker-5] Project Monitor Started : APPProject2
[2019-09-20 01:22:11,554: INFO/ForkPoolWorker-5] Project Monitor DONE : APPProject2
[2019-09-20 01:22:12,047: INFO/ForkPoolWorker-4] Project Monitor DONE : APPProject2
  • 如果您在上述时间间隔内检查,tasks.py将执行一项任务,但是2名芹菜工人将执行该任务并以相同的间隔执行同一任务。我不确定为什么2个工人要完成一项任务?

  • settings.py

..
..
# Internationalization
# https://docs.djangoproject.com/en/2.1/topics/i18n/

LANGUAGE_CODE = 'en-us'

TIME_ZONE = 'Asia/Kolkata'

USE_I18N = True

USE_L10N = True

USE_TZ = True
..
..
..
######## CELERY : CONFIG
CELERY_BROKER_URL = 'redis://localhost:6379'
CELERY_RESULT_BACKEND = 'redis://localhost:6379'
CELERY_ACCEPT_CONTENT = ['application/json']
CELERY_TASK_SERIALIZER = 'json'
CELERY_RESULT_SERIALIZER = 'json'
CELERY_ENABLE_UTC = True
CELERYBEAT_SCHEDULER = 'django_celery_beat.schedulers:DatabaseScheduler'
  • celery.py
from __future__ import absolute_import, unicode_literals
from celery import Celery 
import os
from django.conf import settings

os.environ.setdefault('DJANGO_SETTINGS_MODULE','cimexmonitor.settings')
## set the default Django settings module for the 'celery' program.

# Using a string here means the worker don't have to serialize
# the configuration object to child processes.
# - namespace='CELERY' means all celery-related configuration keys
#   should have a `CELERY_` prefix.

app = Celery('cimexmonitor')
#app.config_from_object('django.conf:settings', namespace='CELERY') 
app.config_from_object('django.conf:settings')
# Load task modules from all registered Django app configs.
app.autodiscover_tasks(settings.INSTALLED_APPS)

@app.task(bind=True)
def debug_task(self):
    print('Request: {0!r}'.format(self.request))
  • 其他信息:
→ celery --version
4.3.0 (rhubarb)

→ redis-server --version
Redis server v=3.0.6 sha=00000000:0 malloc=jemalloc-3.6.0 bits=64 build=7785291a3d2152db

django-admin-interface==0.9.2
django-celery-beat==1.5.0
  • 请帮助我调试问题的方法:

谢谢

2 个答案:

答案 0 :(得分:1)

worker和beat服务都需要同时运行才能按照https://github.com/celery/django-celery-beat定期执行任务

  • 工作者:
 $ celery -A [project-name] worker --loglevel=info -B -c 5
  • Django调度程序:
celery -A [project-name] beat -l info --scheduler django_celery_beat.schedulers:DatabaseScheduler
  • 我同时在运行两个worker和数据库调度程序,根据文档所述,这导致问题在同一时间执行,我真的不确定celery worker如何开始工作同时作为数据库调度程序。
  • 仅运行celery worker就解决了我的问题。

答案 1 :(得分:0)

摘自官方文档:Ensuring a task is only executed one at a time

此外,我希望您不要以相同的方式(celery -A cimexmonitor worker --loglevel=info -B -c 4来运行多个工作人员,因为这将意味着您要安排多个芹菜节拍来安排任务的运行时间...总之,请确保您只有一个芹菜节拍。运行!