That sounds like a good idea at first, but falls down with possible code 
changes in operators between one task and the next.

(I would like this, but there are definite complexities)

-ash


On 31 January 2019 16:56:54 GMT, Dan Davydov <ddavy...@twitter.com.INVALID> 
wrote:
>I feel the right higher-level solution to this problem (which is
>"Adding
>Consistency to Airflow") is DAG serialization, that is all DAGs should
>be
>represented as e.g. JSON (similar to the current SimpleDAGBag object
>used
>by the Scheduler). This solves the webserver issue, and also adds
>consistency between Scheduler/Workers (all DAGruns can be ensured to
>run at
>the same version of a DAG instead of whatever happens to live on the
>worker
>at the time).
>
>On Thu, Jan 31, 2019 at 9:44 AM Peter van ‘t Hof <
>petervant...@godatadriven.com> wrote:
>
>> Hi All,
>>
>> As most of you guys know, airflow got an issue when loading new dags
>where
>> the webserver sometimes sees it and sometimes not.
>> Because of this we did wrote this AIP to solve this issue:
>>
>>
>https://cwiki.apache.org/confluence/display/AIRFLOW/AIP-12+Persist+DAG+into+DB
>>
>> Any feedback is welcome.
>>
>> Gr,
>> Peter van 't Hof
>> Big Data Engineer
>>
>> GoDataDriven
>> Wibautstraat 202
>> 1091 GS Amsterdam
>> https://godatadriven.com
>>
>>

Reply via email to