![]() Raise six.reraise(type(error), error, _stacktrace) File "/home/airflow/.local/lib/python3.7/site-packages/urllib3/packages/six.py", "/home/airflow/.local/lib/python3.7/site-packages/urllib3/util/retry.py", Method, url, error=e, _pool=self, _stacktrace=sys.exc_info() 2 File Timeout=timeout File "/home/airflow/.local/lib/python3.7/site-packages/urllib3/connectionpool.py", "/home/airflow/.local/lib/python3.7/site-packages/requests/adapters.py", No the DAG still fails with the following log error: I added - /var/run/docker.sock:/var/run/docker.sock to the volumes section in my airflow docker-compose file. My DAG script looks like this: from airflow import DAGįrom .docker import DockerOperatorĭescription='Testing the docker operator',Ĭommand='echo "this is a test message shown from within the container',Īfter digging around a bit, I assumed that this is a Docker-in-Docker issue, the most likely solution was found in this tutorial. So far so good, however if I trigger the DAG with my docker-operator it fails and I get a severalįileNotFoundError: No such file or directory Running this from the CLI with docker run -name docker-test-container docker-test-image will give me the expected output: Hello World I can create my image with docker build -t docker-test-image. Which together with the following simple docker-file FROM python:3.9 To test this I have created a simple "Hello World!" example-script: import numpy as np This works fine, however I would like to move the python scripts, which are my tasks out of the mounted plugins folder and into their own docker containers. I have a docker container running on my windows machine, which was build with an adapted version of the docker-compose file provided in the official docs.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |