Airflow "This DAG isnt available in the webserver DagBag object "

PythonAirflowWorkflow

Python Problem Overview


when I put a new DAG python script in the dags folder, I can view a new entry of DAG in the DAG UI but it was not enabled automatically. On top of that, it seems does not loaded properly as well. I can only click on the Refresh button few times on the right side of the list and toggle the on/off button on the left side of the list to be able to schedule the DAG. These are manual process as I need to trigger something even though the DAG Script was put inside the dag folder.

Anyone can help me on this ? Did I missed something ? Or this is a correct behavior in airflow ?

By the way, as mentioned in the post title, there is an indicator with this message "This DAG isn't available in the webserver DagBag object. It shows up in this list because the scheduler marked it as active in the metdata database" tagged with the DAG title before i trigger all this manual process.

Python Solutions


Solution 1 - Python

It is not you nor it is correct or expected behavior. It is a current 'bug' with Airflow. The web server is caching the DagBag in a way that you cannot really use it as expected.

"Attempt removing DagBag caching for the web server" remains on the official TODO as part of the roadmap, indicating that this bug may not yet be fully resolved, but here are some suggestions on how to proceed:

only use builders in airflow v1.9+

Prior to airflow v1.9 this occurs when a dag is instantiated by a function which is imported into the file where instantiation happens. That is: when a builder or factory pattern is used. Some reports of this issue on github 2 and JIRA 3 led to a fix released with in airflow v1.9.

If you are using an older version of airflow, don't use builder functions.

airflow backfill to reload the cache

As Dmitri suggests, running airflow backfill '<dag_id>' -s '<date>' -e '<date>' for the same start and end date can sometimes help. Thereafter you may end up with the (non)-issue that Priyank points, but that is expected behavior (state: paused or not) depending on the configuration you have in your installation.

Solution 2 - Python

Restart the airflow webserver solves my issue.

Solution 3 - Python

This error can be misleading. If hitting refresh button or restarting airflow webserver doesn't fix this issue, check the DAG (python script) for errors.

Running airflow list_dags can display the DAG errors (in addition to listing out the dags) or even try running/testing your dag as a normal python script.

After fixing the error, this indicator should go away.

Solution 4 - Python

The issue is because the DAG by default is put in the DagBag in paused state so that the scheduler is not overwhelmed with lots of backfill activity on start/restart.

To work around this change the below setting in your airflow.cfg file:

# Are DAGs paused by default at creation 
dags_are_paused_at_creation = False

Hope this helps. Cheers!

Solution 5 - Python

I have a theory about possible cause of this issue in Google Composer. There is section about dag failures on webserver in troubleshooting documentation for Composer, which says:

> Avoid running heavyweight computation at DAG parse time. Unlike the > worker and scheduler nodes, whose machine types can be customized to > have greater CPU and memory capacity, the webserver uses a fixed > machine type, which can lead to DAG parsing failures if the parse-time > computation is too heavyweight.

And I was trying to load configuration from external source (which actually took negligible amount of time comparing to other operations to create DAG, but still broke something, because webserver of Airflow in composer runs on App Engine, which has strange behaviours).

I found the workaround in discussion of this Google issue, and it is to create separate DAG with task which loads all the data needed and stores that data in airflow variable:

Variable.set("pipeline_config", config, serialize_json=True)

Then I could do

Variable.get("pipeline_config", deserialize_json=True)

And successfully generate pipeline from that. Additional benefit is that I get logs from that task, which I get from web server, because of this issue.

Attributions

All content for this solution is sourced from the original question on Stackoverflow.

The content on this page is licensed under the Attribution-ShareAlike 4.0 International (CC BY-SA 4.0) license.

Content TypeOriginal AuthorOriginal Content on Stackoverflow
QuestionsantiView Question on Stackoverflow
Solution 1 - PythonGuilleView Answer on Stackoverflow
Solution 2 - PythoncozyssView Answer on Stackoverflow
Solution 3 - PythonrahulView Answer on Stackoverflow
Solution 4 - PythonPriyank MehtaView Answer on Stackoverflow
Solution 5 - PythonBunykView Answer on Stackoverflow