我正在尝试在Celery框架之上构建一个应用程序。
我有一个模块settings/celery_settings.py
,其中包含初始化Celery应用程序的代码(我扩展了一些变量):
from __future__ import absolute_import
from celery import Celery
pfiles = ['other_tasks.test123', 'balance_log.balance_log']
app = Celery('myapp')
# here I just have some parameters defined like broker, result backend, etc
# app.config_from_object(settings)
# TRYING to discover tasks
app.autodiscover_tasks(pfiles)
文件other_tasks/test123.py
和balance_log/balance_log.py
包含以下任务定义:
# file other_tasks/test123.py
from celery import shared_task, Task
@shared_task()
def mytask():
print("Test 1234!")
class TestTask01(Task):
def run(self, client_id=None):
logger.debug("TestTask01: run")
return client_id
我经营芹菜工人:
python3 /usr/local/bin/celery -A settings.celery_settings worker
这样可以发现任务。我可以称之为这些任务。
但后来我尝试使用IPython:
In [1]: from settings.celery_settings import app
In [2]: app.tasks
Out[2]:
{'celery.backend_cleanup': <@task: celery.backend_cleanup of XExchange:0x7f9f50267ac8>,
'celery.chain': <@task: celery.chain of XExchange:0x7f9f50267ac8>,
'celery.chord': <@task: celery.chord of XExchange:0x7f9f50267ac8>,
'celery.chord_unlock': <@task: celery.chord_unlock of XExchange:0x7f9f50267ac8>,
'celery.chunks': <@task: celery.chunks of XExchange:0x7f9f50267ac8>,
'celery.group': <@task: celery.group of XExchange:0x7f9f50267ac8>,
'celery.map': <@task: celery.map of XExchange:0x7f9f50267ac8>,
'celery.starmap': <@task: celery.starmap of XExchange:0x7f9f50267ac8>}
显然没有发现任务。
似乎当我明确地调用任务时,我首先导入它们并在调用时指定芹菜的确切路径。这就是它起作用的原因。
问题:如何让他们发现有已知任务列表?
答案 0 :(得分:4)
最后我发现autodiscover_tasks
函数有一个额外的参数:
def autodiscover_tasks(self, packages, related_name='tasks', force=False):
...
因此,在设置force=True
后,它开始运作了!
app.autodiscover_tasks(pfiles, force=True)
答案 1 :(得分:1)
这是我的示例配置:
CONF / celeryconfig
from conf import settings
CELERYD_CHDIR='/usr/local/src/imbue/application/imbue'
CELERY_IGNORE_RESULT = False
CELERY_RESULT_BACKEND = "amqp"
CELERY_TASK_RESULT_EXPIRES = 360000
CELERY_RESULT_PERSISTENT = True
BROKER_URL='amqp://<USERNAME>:<PASSWORD>@rabbitmq:5672'
CELERY_ENABLE_UTC=True
CELERY_TIMEZONE= "US/Eastern"
CELERY_IMPORTS=("hypervisor.esxi.vm_operations",
"tools.deploy_tools",)
管理程序/ ESXi的/ vm_operations.py
@task(bind=True, default_retry_delay=300, max_retries=5)
def cancel_job(self, host_id=None, vm_id=None, job=None, get_job=False, **kwargs):
pass
call_task.py
def call_task():
log.info('api() | Sending task: ' + job_instance.reference)
celery = Celery()
celery.config_from_object('conf.celeryconfig')
celery.send_task("hypervisor.esxi.vm_operations.cancel_job",
kwargs={'job': job_instance,
'get_job': True},
task_id=job_instance.reference)
我将Celery与Supervisor一起使用,然后从conf目录启动它:
source ~/.profile
CELERY_LOGFILE=/usr/local/src/imbue/application/imbue/log/celeryd.log
CELERYD_OPTS=" --loglevel=INFO --autoscale=10,5"
cd /usr/local/src/imbue/application/imbue/conf
exec celery worker -n celeryd@%h -f $CELERY_LOGFILE $CELERYD_OPTS