Why Am I conducting this course ?
When I started configuring airflow in my organisation, I spent many weeks on writing Docker Compose files for each airflow component.
Airflow community provides a single docker compose file which installs all the components in a single machine. It works fine for practising purposes. But in production, we setup each component on different machines. Also, there is no docker image available on Docker registry to start airflow through Docker Swarm.
Overall, I contributed many sleepless nights to achieve fault tolerant, resilient, distributed, Highly Available airflow using Docker Swarm. I consolidated all my learnings, knowledge into this course. So that others don't need to struggle like mine.
What do you Learn From this Course ?
The primary objective of this course is to achieve resilient airflow using the Docker and Docker Swarm. I am using the latest stable airflow (2.0.1) throughout this course.
At first, we study all the required docker concepts. Don't worry if you have no prior experience on docker. I cover all docker necessary concepts which are used in this course. In the second module, we investigate Airflow 2.0 and understand the additional advantage over airflow 1.x. We discover the airflow HA architecture and discuss each system requirement.
After this, we acquire these machines from AWS and start containerising one by one applications using docker compose.
At the end, we run multiple airflow schedulers and benchmark it.
Last module is on docker swarm and we witness how easy it is to setup the entire airflow just by running a few swarm commands.
In the beginning, we learn all the swarm concepts, architecture, commands and networking. We translate the docker compose file into the docker service command. Subsequently, We create a Swarm Cluster and trigger one by one airflow services on it.