To 'adequately' debug Celery under Windows, there are several ways such as:
> celery worker --app=demo_app.core --pool=solo --loglevel=INFO
# core/settings.py
CELERY_BROKER_URL = 'redis://demo_app_redis:6379'
CELERY_ACCEPT_CONTENT = ['json']
CELERY_TASK_SERIALIZER = 'json'
# core/celery.py
from __future__ import absolute_import, unicode_literals
import os
from celery import Celery
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'core.settings')
app = Celery('core')
app.config_from_object('django.conf:settings', namespace='CELERY')
app.autodiscover_tasks()
# core/__init__.py
from __future__ import absolute_import, unicode_literals
from .celery import app as celery_app
__all__ = ('celery_app',)
# app/tasks.py
from celery.task import periodic_task
from datetime import timedelta
@periodic_task(run_every=(timedelta(seconds=5)), name='hello')
def hello():
print("Hello there")
# Dockerfile
FROM python:3.7
RUN pip install pipenv
RUN mkdir /code
WORKDIR /code
COPY Pipfile* /code/
RUN pipenv install --system --deploy --ignore-pipfile
ADD core /code/
# docker-compose.yml
version: '3'
services:
redis:
image: redis
restart: always
container_name: 'demo_app_redis'
command: redis-server
ports:
- '6379:6379'
web:
build: .
restart: always
container_name: 'demo_app_django'
command: python manage.py runserver 0.0.0.0:8000
volumes:
- .:/code
ports:
- '8000:8000'
celery:
build: .
container_name: 'demo_app_celery'
command: celery -A core worker -B
volumes:
- .:/code
links:
- redis
depends_on:
- web
- redis
К сожалению, не доступен сервер mySQL