Airflow — The Easy Way

“Running Airflow on AWS EC2 & RDS using docker-compose”
5 min readJan 6, 2022

Hello Folks,

Image Source — Google

Lets start the year on the roll. Wishing you all a successful learning year.


For quick set up of Apache Airflow, we will deploy airflow using docker-compose and run it on AWS EC2 & RDS Instance.

Some of the readers reached out to me for more easy & development friendly playground for Airflow Setup on AWS.

Here I am with Airflow — The Easy Way

Table Of Contents

  • Introduction
  • Prerequisites
  • Architecture
  • AWS Infrastructure Provisioning
  • Airflow Provisioning
  • Environment Validation
  • Cleanup

Introduction -

Airflow — Please check my first blog

docker-compose — It is used to run multiple containers as a single service. For example, suppose you had an application which required NGNIX and MySQL, you could create one file which would start both the containers as a service without the need to start each one separately.

The docker-compose.yaml contains several service definitions:
airflow-scheduler — The scheduler monitors all tasks and DAGs, then triggers the task instances once their dependencies are complete.
airflow-webserver — The webserver available at http://localhost:8080.
airflow-worker — The worker that executes the tasks given by the scheduler.
airflow-init — The initialization service.
flower — The flower app for monitoring the environment & available at http://localhost:5555.
redis — The redis — broker that forwards messages from scheduler to worker.

Some directories in the container are mounted, which means that their contents are synchronized between the services.

  • ./dags — you can put your DAG files here.
  • ./logs — contains logs from task execution and scheduler.
  • ./plugins — you can put your custom plugins here.

Prerequisites -

  • Must have access to an AWS account with the required roles or permissions. The below steps can be run from AWS EC2 Instance(Ubuntu) in the given AWS account with necessary access permissions.
  • AWS Services — Full Access to RDS, EC2, IAM, S3, VPC
  • Tools Dependencies — AWS CLI (V2), Cron, docker-compose

Architecture -

High Level — Airflow on EC2 & RDS Architecture

AWS Infrastructure Provisioning -

$ aws configure

AWS Access Key ID [None]: (Your Access Key)

AWS Secret Access Key [None]: (Your Secret Key)

Default region name [None]: (Your Region)

Default output format [None]: json

  • Install Ubuntu Desktop & XRDP for remote RDP.

# sudo apt-get update && sudo apt-get upgrade

# sudo apt install tasksel

# sudo tasksel install ubuntu-desktop

# reboot (You have to Log In Again to EC2 Instance & run the below command)

# sudo apt-get install xrdp

  • Now you can either change the user ubuntu password or create a new user.
  • This will be used for RDP authentication.
  • Install vim editor -> apt install vim
  • Install Cron -> apt install cron
  • (Optional) Install Google Chrome browser. Run below mentioned commands in the given order.

# wget

# sudo apt install ./google-chrome-stable_current_amd64.deb

Airflow Provisioning -

  • Copy the docker-compose.yaml file on AWS EC2 Instance & update below parameters.



  • set the env variable -> echo -e “AIRFLOW_UID=50000\nAIRFLOW_GID=0” > .env
  • Create local folders on EC2 instance -> mkdir ./dags ./logs ./plugins
  • Install docker-compose ->

sudo curl -L “$(uname -s)-$(uname -m)” -o /usr/bin/docker-compose

  • Set Crontab to Sync s3 folder to EC2 local folder.

# crontab -e

# add below commands inside the editor.

# * * * * * /usr/local/bin/aws s3 sync s3://<S3 Folder Path> /root/dags/

# * * * * * /usr/local/bin/aws s3 sync s3://<S3 Folder Path> /root/plugins/

# Change s3 folder as per your environment bucket folder.

  • Start the Cron service -> service cron start
  • Deploy Airflow through docker-compose -> docker-compose up -d
  • Please verify container status using below commands from EC2 bash terminal

# docker ps

# docker-compose run airflow-worker airflow info

docker ps — output
  • To upload custom DAGs on Airflow Web UI -
  • We need to upload DAGs & plugins file in the respective created s3 bucket.

Environment Validation -

Airflow Web UI
  • Enter Credentials

username — airflow

password — airflow

  • After login Check the DAGs & start running it.
Example DAGs
  • As you trigger the DAG, Airflow will create pods to execute the code included in the DAG.
DAGs Running Status
  • Check RDS connections on AWS Console it will show current connections from Airflow docker.
  • Voilaaaa..!! Airflow is ready on AWS EC2 & RDS.
  • Pros- Easy, Fast, developer friendly setup
  • Cons- Not production ready, Performance issues, Slowness

Cleanup -

  • docker compose stop.
  • Delete the CloudFormation template of AWS EC2 & RDS.
  • Delete the S3 buckets created from console.


I had fun deploying this setup & playing around AWS EC2, RDS & AIRFLOW.

Hope you guys like it & start playing around.

More things lined up around AWS Stay Tuned..

“Nothing is particularly hard if you break it down into small bits”

Image Source — Google

I am Kunal Shah, AWS Certified Solutions Architect, helping clients to achieve optimal solutions on the Cloud. Cloud Enabler by choice, DevOps Practitioner having 7+ Years of overall experience in the IT industry.

I love to talk about Cloud Technology, DevOps, Digital Transformation, Analytics, Infrastructure, Dev Tools, Operational efficiency, Serverless, Cost Optimization, Cloud Networking & Security.

#aws #community #builders #devops #airflow #docker #ec2 #infrastructure

You can reach out to me @



AWS Community Builder | AWS Solutions Architect | Infrastructure | DevOps | Cloud Network & Security | CFT | Terraform | K8s | SysOps | FinOps | Blog | Share