Replies: 1 comment 5 replies
-
A lot depends on resources you have and your deployment. You have not explained which executor you have, how many machines you have, what is the parallelism for those machines, what kind of tasks you have them. Airflow as you see - is correclty scheduling the tasks, now it's your job as deployment manager, to make sure that you have enough workers (for celery) with enough resources (CPU/machines, memory, sometimes GPUS etc) to be able to run as many parallel processes you migh want. In Kubernetes - you need to have enough resources to be able to run as many PODs you have. Depending on the capacity of your system, the number of parallel things that might be executed. It's also your job to make sure that you allocate appropriately the resources (for example via K8S so I suggest you look at your deployment and make sure those resources are not blocking you. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
We have Airflow version 2.5.0 and we're trying to execute a job that contains 60 tasks which can be executed in parallel. Despite trying several concurrency variable values in Airlfow service and environment settings, the airflow is still unable to execute 60 tasks in parallel.
While executing, it starts out with 32 tasks(screenshot 1,2,3).
And after a few moments 16 out of the 32 change status to 'up_for_try' state(screenshot 4,5) and eventually get queued again.
On the other hand these 32 tasks that are initiated at once are still being executed in ECS as shown in screenshot 6. So it seems like those tasks are still executing while Aiflow is unable to keep track of them and sets them to be rerun.
Any help would be appreciated. Thank you
Beta Was this translation helpful? Give feedback.
All reactions