Answer The offending line has been removed from Celery nearly 6 years ago. You should update the celery package to a recent version.
Tag: celery-task
Use existing celery workers for Airflow’s Celeryexecutor workers
I am trying to introduce dynamic workflows into my landscape that involves multiple steps of different model inference where the output from one model gets fed into another model.Currently we have few Celery workers spread across hosts to manage the inference chain. As the complexity increase, we are attempting to build workflows on the fly. For that purpose, I got
python celery monitoring events not being triggered
I have a following project directory: tasks.py main.py monitor.py Started celery worker with And started monitor.py with But Only worker-online event is being triggered, while other events like task-succeeded is not triggered or handled. What am I missing here? Answer Enable worker task- group events with cli option -E or –task-events and try to capture all events:
python celery invalid value for -A unable to load application
I have a following project directory: task.py main.py When I run a command to initialize celery workers, I get this error: But when I rename main.py to celery.py as it was earlier there’s no issue. What am I missing here? Answer There are two approaches import your app to azima/__init__.py You can omit the last line, celery will recognize the
Django Celery delay() always pushing to default ‘celery’ queue
I’m ripping my hair out with this one. The crux of my issue is that, using the Django CELERY_DEFAULT_QUEUE setting in my settings.py is not forcing my tasks to go to that particular queue that I’ve set up. It always goes to the default celery queue in my broker. However, if I specify queue=proj:dev in the shared_task decorator, it goes