使用Celery

时间:2016-12-14 14:04:45

标签: python python-3.x asynchronous celery celery-task

我正在编写一个应用程序,它将异步执行一组多个同步任务链。

换句话说,对于某些foo(a,b,c) -> boo(a,b,c)列表,我可能会有管道bs

我的理解是为此列表中的每个b创建一个foo(a,b,c) | boo(a,b,c)链。然后这些链形成一个芹菜组,可以异步应用。

我的代码如下:

  

my_app.py

#!/usr/bin/env python3

import functools
import time

from celery import chain, group, Celery
from celery.utils.log import get_task_logger

logger = get_task_logger(__name__)

app = Celery("my_app", broker='redis://localhost:6379/0', backend='redis://localhost:6379/0')

@app.task
def foo(a, b, c):
    logger.info("foo from {0}!".format(b))
    return b

@app.task
def boo(a, b, c):
    logger.info("boo from {0}!".format(b))
    return b

def break_up_tasks(tasks):
    try:
        first_task, *remaining_tasks = tasks
    except ValueError as e:
        first_task, remaining_tasks = [], []
    return first_task, remaining_tasks

def do_tasks(a, bs, c, opts):
    tasks = [foo, boo]

    # There should be an option for each task
    if len(opts) != len(tasks):
        raise ValueError("There should be {0} provided options".format(len(tasks)))

    # Create a list of tasks that should be included per the list of options' boolean values
    tasks = [task for opt, task in zip(opts, tasks) if opt]

    first_task, remaining_tasks = break_up_tasks(tasks)

    # If there are no tasks, we're done.
    if not first_task: return

    chains = (
        functools.reduce(
            # `a` should be provided by `apply_async`'s `args` kwarg
            # `b` should be provided by previous partials in chain
            lambda x, y: x | y.s(c),
            remaining_tasks, first_task.s(a, b, c)
        ) for b in bs
    )

    g = group(*chains)
    res = g.apply_async(args=(a,), queue="default")
    print("Applied async... waiting for termination.")

    total_tasks = len(tasks)

    while not res.ready():
        print("Waiting... {0}/{1} tasks complete".format(res.completed_count(), total_tasks))
        time.sleep(1)

if __name__ == "__main__":
    a = "whatever"
    bs = ["hello", "world"]
    c = "baz"

    opts = [
        # do "foo"
        True,
        # do "boo"
        True
    ]

    do_tasks(a, bs, c, opts)
  

跑芹菜

celery worker -A my_app -l info -c 5 -Q default

我发现,当我运行上述操作时,我的服务器客户端运行无限循环,因为boo缺少一个参数:

  

TypeError: boo() missing 1 required positional argument: 'c'

我的理解是apply_async将为每个链提供args kwarg,并且链中的先前链接将为后续链接提供返回值。

为什么boo没有正确接收参数?我确信这些任务写得不好,因为这是我第一次涉足Celery。如果你有其他建议,我很乐意招待他们。

1 个答案:

答案 0 :(得分:3)

在调试代码之后(我也是Celery的新手!:))我已经知道每个链接的函数都会将第一个参数替换为前一个链接函数调用的结果 - 所以用它说我相信你的问题的解决方案是在reduce中的ys中添加一个缺少的参数(第二个):

chains = (
    functools.reduce(
        # `a` should be provided by `apply_async`'s `args` kwarg
        # `b` should be provided by previous partials in chain
        lambda x, y: x | y.s(b,c), # <- here is the 'new guy'
        remaining_tasks, first_task.s(a, b, c)
    ) for b in bs
)

希望它有所帮助。