-
Notifications
You must be signed in to change notification settings - Fork 14.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Task fails and cannot read logs. Invalid URL 'http://:8793/log/...': No host supplied #42136
Comments
having the same issue with 2.10.1 in k8s, using the CeleryKubernetesExecutor. Could this be related to the inheritance issue that was discussed in #41891? |
Additionally I checked the logs directory for the dag_id/run_id and it's missing the respective task_id folder. |
Having the same issue on |
We have upgraded on |
got the same behavior since upgrading from version 2.9.3 to 2.10.1. |
I have the same issue with 2.10.0, using the CeleryExecutor.
eventlog
Scheduler has a error log at the same hour as eventlog.
|
same issue for us when upgrading to 2.10.2 |
We’re encountering the same issue as well. |
We have switched from Bitnami docker-compose to official Apache docker-compose and we could make it run successfuly 🤩 |
try to check that dags exist in worker, schedule and webserver. I deploy Airflow in K8S and get this error when putting my dags into scheduler(expecting that in will replicate into another pods), but when I check dags folder in worker it was empty |
At this time (#42136 (comment)), I used the |
We encountered the same problem in Airflow 2.9.3
I can also note that increasing the sql_alchemy_pool_size parameter to 50 reduced the number of such errors, but did not eliminate them completely. |
The same issue in Airflow 2.10.2 |
TL;DR look for invalid python scripts on the malfunctioning worker. Try creating a DagBag on the worker and see what happens. # Ensure the AIRFLOW_HOME points to the right location, then run on the worker
>>> from airflow.models import DagBag
>>> DagBag(include_examples=False) I had this issue too, turns out I edited one of the files through vim, pasted some code, and it pasted tabs instead of spaces, so the file became an invalid python script due to Note that the problematic file doesn't have to be imported by the failing DAG/task. If I understand the issue correctly, a DagBag cannot be created if one of the DAG definition files or their imports isn't a valid python file. Then the issue manifests as DAGs supposedly not being found. In my case, the filesystem isn't shared between the scheduler and the malfunctioning celery worker, and the affected file was unmodified on the scheduler (or modified in a correct way) - so no "big red import error" was displayed in the webserver UI. |
Hi @quack39 and all, Error Could not read served logs: HTTPConnectionPool(host='test-dag-config-nlp8suol', port=8793): Max retries exceeded with url: /log/dag_id=test_dag/run_id=manual__2024-11-06T06:43:18.256272+00:00/task_id=config/attempt=1.log (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f8ef9eaecd0>: Failed to establish a new connection: [Errno -2] Name or service not known')) |
I had the same problem when changing sequentialexecutor to localexecutor. with t3.large(CPU core :2) But is this the expected action ? I'm not sure . |
same issue for us when upgrading to 2.10.3,using k8s. |
The same issue in Airflow 2.10.2 with KubernetesExecutor. |
I was just able to get the following to work:
initially it failed with the same error, then succeeded on retry, I believe because the log stream was still in 'create' and not available for the first attempt. |
Same issue here, even if the task is successful |
Apache Airflow version
Other Airflow 2 version (please specify below)
If "Other Airflow 2 version" selected, which one?
2.10.1
What happened?
I'm having an issue with an airflow instance where a task fails and I cannot read the logs.
Logs:
Event logs:
Additionally I checked the logs directory for the dag_id/run_id and it's missing the respective task_id folder.
What you think should happen instead?
I should be able to access the logs.
How to reproduce
Not sure how to.
Operating System
Ubuntu 24.04 LTS
Versions of Apache Airflow Providers
No response
Deployment
Other Docker-based deployment
Deployment details
Deployed with docker-compose on Docker Swarm setup on 2 VMs.
Anything else?
Additionally I checked the logs directory for the
dag_id/run_id
and it's missing the respectivetask_id
folder.Are you willing to submit PR?
Code of Conduct
The text was updated successfully, but these errors were encountered: