Configuring Amazon MWAA worker automatic scaling
The auto scaling mechanism automatically increases the number of Apache Airflow workers in response to running and queued tasks on your Amazon Managed Workflows for Apache Airflow environment and disposes of extra workers when there are no more tasks queued or executing. This topic describes how you can configure auto scaling by specifying the maximum number of Apache Airflow workers that run on your environment using the Amazon MWAA console.
Note
Amazon MWAA uses Apache Airflow metrics to determine when additional Celery Executormax-workers
. As the additional workers complete work and work load
decreases, Amazon MWAA removes them, thus downscaling back to the value set by min-workers
.
If workers pick up new tasks while downscaling, Amazon MWAA keeps the Fargate resource and does not remove the worker. For more information, see How Amazon MWAA auto scaling works.
Sections
How worker scaling works
Amazon MWAA uses RunningTasks
and QueuedTasks
metrics, where
(tasks running + tasks queued) / (tasks per worker) = (required workers). If the required number of workers is greater than the current number of workers,
Amazon MWAA will add Fargate worker containers to that value, up to the maximum value specified by max-workers
.
As the workload decreases and the RunningTasks
and QueuedTasks
metric sum reduces, Amazon MWAA requests Fargate to scale down the workers for the environment.
Any workers which still completing work remain protected during downscaling until they complete their work. Depending on the workload, tasks may be queued while
workers downscale.
Using the Amazon MWAA console
You can choose the maximum number of workers that can run on your environment concurrently on the Amazon MWAA console. By default, you can specify a maximum value up to 25.
To configure the number of workers
-
Open the Environments page
on the Amazon MWAA console. -
Choose an environment.
-
Choose Edit.
-
Choose Next.
-
On the Environment class pane, enter a value in Maximum worker count.
-
Choose Save.
Note
It can take a few minutes before changes take effect on your environment.
Example high performance use case
The following section describes the type of configurations you can use to enable high performance and parallelism on an environment.
On-premise Apache Airflow
Typically, in an on-premise Apache Airflow platform, you would configure task parallelism, auto scaling, and concurrency settings in your airflow.cfg
file:
-
core.parallelism
– The maximum number of task instances that can run simultaneously per scheduler. -
core.dag_concurrency
– The maximum concurrency for DAGs (not workers). -
celery.worker_autoscale
– The maximum and minimum number of tasks that can run concurrently on any worker.
For example, if core.parallelism
was set to 100
and core.dag_concurrency
was set to 7
, you would still only be able to run a total of 14
tasks concurrently if you had 2 DAGs. Given, each DAG is set to run only seven tasks concurrently (in core.dag_concurrency
), even though overall parallelism is set to 100
(in core.parallelism
).
On an Amazon MWAA environment
On an Amazon MWAA environment, you can configure these settings directly on the Amazon MWAA console using Using Apache Airflow configuration options on Amazon MWAA, Configuring the Amazon MWAA environment class, and the Maximum worker count auto scaling mechanism. While core.dag_concurrency
is not available in the drop down list as an Apache Airflow configuration option on the Amazon MWAA console, you can add it as a custom Apache Airflow configuration option.
Let's say, when you created your environment, you chose the following settings:
-
The mw1.small environment class which controls the maximum number of concurrent tasks each worker can run by default and the vCPU of containers.
-
The default setting of
10
Workers in Maximum worker count. -
An Apache Airflow configuration option for
celery.worker_autoscale
of5,5
tasks per worker.
This means you can run 50 concurrent tasks in your environment. Any tasks beyond 50 will be queued, and wait for the running tasks to complete.
Run more concurrent tasks. You can modify your environment to run more tasks concurrently using the following configurations:
-
Increase the maximum number of concurrent tasks each worker can run by default and the vCPU of containers by choosing the
mw1.medium
(10 concurrent tasks by default) environment class. -
Add
celery.worker_autoscale
as an Apache Airflow configuration option. -
Increase the Maximum worker count. In this example, increasing maximum workers from
10
to20
would double the number of concurrent tasks the environment can run.
Specify Minimum workers. You can also specify the minimum and maximum number of Apache Airflow Workers that run in your environment using the Amazon Command Line Interface (Amazon CLI). For example:
aws mwaa update-environment --max-workers 10 --min-workers 10 --name
YOUR_ENVIRONMENT_NAME
To learn more, see the update-environment command in the Amazon CLI.
Troubleshooting tasks stuck in the running state
In rare cases, Apache Airflow may think there are tasks still running. To resolve this issue, you need to clear the stranded task in your Apache Airflow UI. For more information, see the I see my tasks stuck or not completing troubleshooting topic.
What's next?
-
Learn more about the best practices we recommend to tune the performance of your environment in Performance tuning for Apache Airflow on Amazon MWAA.