t4n1o edited a comment on issue #19192:
URL: https://github.com/apache/airflow/issues/19192#issuecomment-1003779286


   >     1. Do you actually have scheduler running at all?  Does it have 
coninuous access to the DB?
   Yes, the scheduler is  visible in the running processes, and also systemd. 
   The scheduler is the only program accessing the db. The db is on the same 
RAID array/filesystem as the data. 
   >     2. Are you absolutely sure you are not using SequentialExecutior ? 
What does your `airflow info` say  - can you paste-bin output of it ? (airlfow 
has built-in flag to send to pastebin). Please make sure also that you do it in 
exactly the way your scheduler works. Miost likely you run your scheduler with 
a different configuration than your webserver and that causes the problem.
    I can't find the flag for pastebin but I pasted above.
   >     3. Are you sure you are using Postgres and not Sqlite?
    Yes
   >     4. Where is your Python code (non-DAG)?  Did you .airflowignore 
non-DAG files from airflow's DAG folder?
          /opt/data-workflows  <= code
          /opt/data-workflows/rust <= code
          /opt/data-workflows/dags
          There are only dags in the dag folder, I kept everything modular and 
separated.
   >     5. can you upgrade to Airlfow 2.2.3 (latest released) - it has 
built-in warnings in case you use Sequential Executor/SQLite in the  UI.
    Yes, still the same problem on 2.2.3. No warnings about sqllite or 
sequential executor
   >     6. Can you change your DAGs to:
    When I run your dag with the `sleep 1000`, everything works properly, can 
launch many dags in parallel. But launching my standalone scripts (any of them 
- python or rust ones) with BashOperator in a new process creates this problem 
;/ I considered whether there could be contention between reading/writing from 
the disks, because the postgres db is on the same disks, but this weird state 
can persist indefinitely, and I believe linux would allow postgres to at least 
take turns reading/writing from the disk.
         
   Perhaps there is something weird in the way BashOperator spawns subprocesses 
that can block the entire scheduler when used this way. Or a bug with 
subprocess. Or maybe it doesn't like the CWD to be changed.
   ```
     bash_command="umask 002 && cd /opt/data_workflows/rust/ && "
           "./parser",
   ```
   
   By the way, I'm sorry this turned into an unpaid debugging session, it 
wasn't my intention.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to