I feel the right higher-level solution to this problem (which is "Adding
Consistency to Airflow") is DAG serialization, that is all DAGs should be
represented as e.g. JSON (similar to the current SimpleDAGBag object used
by the Scheduler). This solves the webserver issue, and also adds
consistency between Scheduler/Workers (all DAGruns can be ensured to run at
the same version of a DAG instead of whatever happens to live on the worker
at the time).

On Thu, Jan 31, 2019 at 9:44 AM Peter van ‘t Hof <
petervant...@godatadriven.com> wrote:

> Hi All,
>
> As most of you guys know, airflow got an issue when loading new dags where
> the webserver sometimes sees it and sometimes not.
> Because of this we did wrote this AIP to solve this issue:
>
> https://cwiki.apache.org/confluence/display/AIRFLOW/AIP-12+Persist+DAG+into+DB
>
> Any feedback is welcome.
>
> Gr,
> Peter van 't Hof
> Big Data Engineer
>
> GoDataDriven
> Wibautstraat 202
> 1091 GS Amsterdam
> https://godatadriven.com
>
>

Reply via email to