The outline for this document in GitHub is available at top-right corner button (with 3-dots and 3 lines).
Airflow is a complex project, but setting up a working environment is quite simple if you follow the guide.
There are three ways you can run the Airflow dev env:
- With a local virtual environment (on your local machine)
- With Docker Containers and Docker Compose (on your local machine). This environment is managed with the Breeze tool written in Python that makes environment management, yeah you guessed it - a breeze.
- With a remote, managed environment (via remote development environment)
Before deciding which method to choose, there are a couple of factors to consider:
- In most cases, installing Airflow in a local environment might be sufficient. For a comprehensive local virtualenv tutorial, visit Local virtualenv
- Running Airflow in a container is the most reliable and repeatable way: it provides a more consistent environment - with almost no dependencies (except docker) on your Host OS / machine and allows integration tests with a number of integrations (Cassandra, MongoDB, MySQL, etc.). However, it also requires 4GB RAM, 40GB disk space and at least 2 cores.
- You need to have a (usually paid) account to access managed, remote virtual environments.
If you do not work in a remote development environment, you will need these prerequisites:
- UV is recommended for managing Python versions and virtual environments
- Docker Community Edition (you can also use Colima or others, see instructions below)
- Docker buildx
- Docker Compose
The below setup describes Ubuntu installation. It might be slightly different on different machines. For Windows, click on Start menu and type WSL to enter Linux console. Please note, not doing this will result in errors in next steps.
- Installing required packages for Docker and setting up docker repo
sudo apt-get update
sudo apt-get install \
ca-certificates \
curl \
gnupg \
lsb-release
sudo mkdir -p /etc/apt/keyrings
curl -fsSL https://download.docker.com/linux/ubuntu/gpg | sudo gpg --dearmor -o /etc/apt/keyrings/docker.gpg
echo \
"deb [arch=$(dpkg --print-architecture) signed-by=/etc/apt/keyrings/docker.gpg] https://download.docker.com/linux/ubuntu \
$(lsb_release -cs) stable" | sudo tee /etc/apt/sources.list.d/docker.list > /dev/null
- Install Docker Engine, containerd
sudo apt-get update
sudo apt-get install docker-ce docker-ce-cli containerd.io
- Manage docker as non-root user
sudo groupadd docker
sudo usermod -aG docker $USER
Note
This is done so a non-root user can access the docker command.
After adding user to docker group Logout and Login again for group membership re-evaluation.
On some Linux distributions, the system automatically creates this group.
- Test Docker installation
docker run hello-world
Note
Read more about Linux post-installation steps for Docker Engine.
If you use Colima as your container runtimes engine, please follow the next steps:
- Install buildx manually and follow its instructions
- Link the Colima socket to the default socket path. Note that this may break other Docker servers
sudo ln -sf $HOME/.colima/default/docker.sock /var/run/docker.sock
- Change docker context to use default
docker context use default
- Installing latest version of the Docker Compose plugin
on Debian / Ubuntu, Install using the repository:
sudo apt-get update
sudo apt-get install docker-compose-plugin
Install manually:
COMPOSE_VERSION="$(curl -s https://api.github.com/repos/docker/compose/releases/latest | grep '"tag_name":'\
| cut -d '"' -f 4)"
COMPOSE_URL="https://github.com/docker/compose/releases/download/${COMPOSE_VERSION}/\
docker-compose-$(uname -s)-$(uname -m)"
sudo curl -L "${COMPOSE_URL}" -o /usr/local/bin/docker-compose
sudo chmod +x /usr/local/bin/docker-compose
Note
This option requires you to manage updates manually. It is recommended that you set up Docker's repository for easier maintenance.
on macOS, you can also install docker-compose via
brew install docker-compose
- Verifying installation
docker-compose --version
- While you can use any virtualenv manager, we recommend using UV as your build and integration frontend. You can read more about UV and its use in Airflow in Local virtualenv.
- After creating the environment, you need to install a few more required packages for Airflow. The below command adds basic system-level dependencies on Debian/Ubuntu-like system. You will have to adapt it to install similar packages if your operating system is MacOS or another flavour of Linux
sudo apt install openssl sqlite3 default-libmysqlclient-dev libmysqlclient-dev postgresql
If you want to install all Airflow providers, more system dependencies might be needed. For example on Debian/Ubuntu
like system, this command will install all necessary dependencies that should be installed when you use
all extras while installing airflow.
sudo apt install apt-transport-https apt-utils build-essential ca-certificates dirmngr \
freetds-bin freetds-dev git graphviz graphviz-dev krb5-user ldap-utils libffi-dev \
libkrb5-dev libldap2-dev libpq-dev libsasl2-2 libsasl2-dev libsasl2-modules \
libssl-dev locales lsb-release openssh-client sasl2-bin \
software-properties-common sqlite3 sudo unixodbc unixodbc-dev
Go to https://github.com/apache/airflow/ and fork the project
Go to your github account's fork of Airflow click on
Codeyou will find the link to your repoFollow Cloning a repository to clone the repo locally (you can also do it in your IDE - see the Using your IDE chapter below)
Note
For windows based machines, on cloning, the Git line endings may be different from unix based systems and might lead to unexpected behaviour on running breeze tooling. Manually setting a property will mitigate this issue. Set it to true for windows.
git config core.autocrlf true
Before committing changes to github or raising a pull request, the code needs to be checked for certain quality standards such as spell check, code syntax, code formatting, compatibility with Apache License requirements etc. This set of tests are applied when you commit your code.
To avoid burden on our CI infrastructure and to save time, prek hooks can be run locally before committing changes.
Note
We have recently started to recommend uv for our local development.
Note
Remember to have global python set to Python >= 3.10 - Python 3.10 is end-of-life already and we've started to use Python 3.10+ features in Airflow and accompanying scripts.
Installing prek is best done with uv (recommended) or pipx.
- Installing required packages
on Debian / Ubuntu, install via
sudo apt install libxml2-utils
on macOS, install via
brew install libxml2
- Installing prek:
Note
You might need to pass --python <python> to force the python version if not it uses the latest system python version.
python value can be fetched from uv python list
uv tool install prek
or with pipx:
pipx install prek
- Go to your project directory
cd ~/Projects/airflow
- Running prek hooks
prek --all-files
No-tabs checker......................................................Passed
Add license for all SQL files........................................Passed
Add license for all other files......................................Passed
Add license for all rst files........................................Passed
Add license for all JS/CSS/PUML files................................Passed
Add license for all JINJA template files.............................Passed
Add license for all shell files......................................Passed
Add license for all python files.....................................Passed
Add license for all XML files........................................Passed
Add license for all yaml files.......................................Passed
Add license for all md files.........................................Passed
Add license for all mermaid files....................................Passed
Add TOC for md files.................................................Passed
Add TOC for upgrade documentation....................................Passed
Check hooks apply to the repository..................................Passed
black................................................................Passed
Check for merge conflicts............................................Passed
Debug Statements (Python)............................................Passed
Check builtin type constructor use...................................Passed
Detect Private Key...................................................Passed
Fix End of Files.....................................................Passed
...........................................................................
- Running prek for selected files
prek --files airflow-core/src/airflow/utils/decorators.py airflow-core/tests/unit/utils/test_task_group.py
- Running specific hook for selected files
prek black --files airflow-core/src/airflow/utils/decorators.py airflow-core/tests/unit/utils/test_task_group.py
black...............................................................Passed
prek ruff --files airflow-core/src/airflow/utils/decorators.py airflow-core/tests/unit/utils/test_task_group.py
Run ruff............................................................Passed
- Enabling prek hook check before push
It will run prek hooks automatically before committing and stops the commit on failure
cd ~/Projects/airflow
prek install
git commit -m "Added xyz"
- If you want to disable prek hooks
cd ~/Projects/airflow
prek uninstall
- For more information on this visit 08_static_code_checks.rst
Following are some of the important links of 08_static_code_checks.rst
For many of the development tasks you will need Breeze to be configured. Breeze is a development
environment which uses docker and docker-compose and its main purpose is to provide a consistent
and repeatable environment for all the contributors and CI. When using Breeze you avoid the "works for me"
syndrome - because not only others can reproduce easily what you do, but also the CI of Airflow uses
the same environment to run all tests - so you should be able to easily reproduce the same failures you
see in CI in your local environment.
- Install
uvorpipx. We recommend to installuvas the general purpose python development environment - you can install it via https://docs.astral.sh/uv/getting-started/installation/ or you can installpipx(>=1.2.1) - follow the instructions in Install pipx It is important to install version of pipx >= 1.2.1 to workaroundpackagingbreaking change introduced in September 2023 - Run
uv tool install -e ./dev/breeze(orpipx install -e ./dev/breeze) in your checked-out repository. Make sure to follow any instructions printed during the installation - this is needed to make sure that thebreezecommand is available in your PATH
Warning
If you see below warning while running pipx - it means that you have hit the
known issue with packaging version 23.2:
⚠️ Ignoring --editable install option. pipx disallows it for anything but a local path,
to avoid having to create a new src/ directory.
The workaround is to downgrade packaging to 23.1 and re-running the pipx install command, for example
by running pip install "packaging<23.2".
pip install "packaging==23.1"
pipx install -e ./dev/breeze --force
- Initialize breeze autocomplete
breeze setup autocomplete
- Initialize breeze environment with required python version and backend. This may take a while for first time.
breeze --python 3.10 --backend postgres
Note
If you encounter an error like "docker.credentials.errors.InitializationError: docker-credential-secretservice not installed or not available in PATH", you may execute the following command to fix it:
sudo apt install golang-docker-credential-helpers
Once the package is installed, execute the breeze command again to resume image building.
If you encounter an error such as
jinja2.exceptions.TemplateNotFound: '/index.html' not found in search path: '/opt/airflow/airflow-core/src/airflow/ui/dist'
you may need to compile the UI assets before starting the Breeze environment. To do so, run the following command before executing step 4:
breeze ui compile-assets
After running this, verify that the compiled UI assets have been added to /airflow/.build/ui.
Then, proceed with:
breeze --python 3.10 --backend postgres
- When you enter the Breeze environment you should see a prompt similar to
[Breeze:3.10.19] root@e4756f6ac886:/opt/airflow#. This means that you are inside the Breeze container and ready to run most of the development tasks. You can leave the environment withexitand re-enter it with justbreezecommand - Once you enter the Breeze environment, create Airflow tables and users from the breeze CLI.
airflow db resetis required to execute at least once for Airflow Breeze to get the database/tables created. If you run tests, however - the test database will be initialized automatically for you
[Breeze:3.10.19] root@b76fcb399bb6:/opt/airflow# airflow db reset
[Breeze:3.10.19] root@b76fcb399bb6:/opt/airflow# airflow users create \
--username admin \
--firstname FIRST_NAME \
--lastname LAST_NAME \
--role Admin \
--email admin@example.org
Note
airflow users command is only available when FAB auth manager is enabled.
- Exiting the Breeze environment. After successfully finishing above command will leave you in container,
type
exitto exit the container. The database created before will remain and servers will be running though, until you stop the Breeze environment completely
[Breeze:3.10.19] root@b76fcb399bb6:/opt/airflow# exit
- You can stop the environment (which means deleting the databases and database servers running in the
background) via
breeze downcommand
breeze down
- Starting the Breeze environment using
breeze start-airflowstarts the Breeze environment with last configuration run( In this case Python version and backend are picked up from last executionbreeze --python 3.10 --backend postgres) It also automatically starts the API server (FastAPI api and UI), triggerer, dag processor and scheduler. It drops you in tmux with triggerer to the right, and Scheduler, API server (FastAPI api and UI), Dag processor from left to right at the bottom. Use[Ctrl + B] and Arrow keysto navigate.
breeze start-airflow
Use CI image.
Branch name: main
Docker image: ghcr.io/apache/airflow/main/ci/python3.10:latest
Airflow source version: 2.4.0.dev0
Python version: 3.10
Backend: mysql 5.7
* Port forwarding:
Ports are forwarded to the running docker containers for components and database
* 12322 -> forwarded to Airflow ssh server -> airflow:22
* 28080 -> forwarded to Airflow api server API -> airflow:8080
* 25555 -> forwarded to Flower dashboard -> airflow:5555
* 25433 -> forwarded to Postgres database -> postgres:5432
* 23306 -> forwarded to MySQL database -> mysql:3306
* 26379 -> forwarded to Redis broker -> redis:6379
Direct links to those services that you can use from the host:
* ssh connection for remote debugging: ssh -p 12322 airflow@localhost (password: airflow)
* API server: http://localhost:28080
* Flower: http://localhost:25555
* Postgres: jdbc:postgresql://localhost:25433/airflow?user=postgres&password=airflow
* Mysql: jdbc:mysql://localhost:23306/airflow?user=root
* Redis: redis://localhost:26379/0
Alternatively you can start the same using the following commands
- Start Breeze
breeze --python 3.10 --backend postgres
- Open tmux
tmux
- Press Ctrl + B and "
airflow scheduler
- Press Ctrl + B and %
airflow api-server
- Press Ctrl + B and %
airflow dag-processor
- Press Ctrl + B and up arrow followed by Ctrl + B and %
airflow triggerer
- Press Ctrl + B followed by (Optional step for better tile arrangement)
:select-layout tiled
Now you can access Airflow web interface on your local machine at http://localhost:28080 with user name
adminand passwordadminSetup a PostgreSQL database in your database management tool of choice (e.g. DBeaver, DataGrip) with host
localhost, port25433, userpostgres, passwordairflow, and default schemaairflowStopping breeze
If breeze was started with breeze start-airflow, this command will stop breeze and Airflow:
[Breeze:3.10.19] root@f3619b74c59a:/opt/airflow# stop_airflow
breeze down
If breeze was started with breeze --python 3.10 --backend postgres (or similar):
[Breeze:3.10.19] root@f3619b74c59a:/opt/airflow# exit
breeze down
Note
stop_airflow is available only when breeze is started with breeze start-airflow.
By default, breeze start-airflow uses mprocs to manage Airflow components. You can use tmux as an
alternative with the --terminal-multiplexer tmux argument. The last choice of yours will be remembered
for the next run, you can switch back to mprocs anytime by using --terminal-multiplexer mprocs.
breeze start-airflow --terminal-multiplexer tmux
and
breeze start-airflow --terminal-multiplexer mprocs
You can also switch terminal multiplexer via breeze config:
breeze setup config --terminal_multiplexer tmux
breeze setup config --terminal_multiplexer mprocs
** Benefits of using tmux:**
- Familiar terminal multiplexer for many developers
- More control over panes and windows
- Customizable key bindings and layouts
- Wide range of plugins and extensions
- ability to see more than one log at a time
Benefits of using mprocs:
- Modern terminal UI with better visual feedback
- Easier navigation with mouse and keyboard
- Individual process controls (start, stop, restart)
- Process status indicators
- Better cross-platform support
For more information on mprocs, look at mprocs documentation.
- Knowing more about Breeze
breeze --help
Following are some of important topics of Breeze documentation:
- Breeze Installation
- Installing Additional tools to the Docker Image
- Regular developer tasks
- Cleaning the environment
- Troubleshooting Breeze environment
- It may require some packages to be installed; watch the output of the command to see which ones are missing
sudo apt-get install sqlite3 libsqlite3-dev default-libmysqlclient-dev postgresql
./scripts/tools/initialize_virtualenv.py
- Add following line to ~/.bashrc in order to call breeze command from anywhere
export PATH=${PATH}:"/home/${USER}/Projects/airflow"
source ~/.bashrc
You can usually conveniently run tests in your IDE (see IDE below) using virtualenv but with Breeze you can be sure that all the tests are run in the same environment as tests in CI.
All Tests are inside ./tests directory.
Running Unit tests inside Breeze environment.
Just run
pytest filepath+filenameto run the tests.
[Breeze:3.10.19] root@63528318c8b1:/opt/airflow# pytest tests/utils/test_dates.py
============================================================= test session starts ==============================================================
platform linux -- Python 3.10.20, pytest-8.3.3, pluggy-1.5.0 -- /usr/python/bin/python
cachedir: .pytest_cache
rootdir: /opt/airflow
configfile: pyproject.toml
plugins: anyio-4.6.0, time-machine-2.15.0, icdiff-0.9, rerunfailures-14.0, instafail-0.5.0, custom-exit-code-0.3.0, xdist-3.6.1, mock-3.14.0, cov-5.0.0, asyncio-0.24.0, requests-mock-1.12.1, timeouts-1.2.1
asyncio: mode=strict, default_loop_scope=None
setup timeout: 0.0s, execution timeout: 0.0s, teardown timeout: 0.0s
collected 4 items
tests/utils/test_dates.py::TestDates::test_parse_execution_date PASSED [ 25%]
tests/utils/test_dates.py::TestDates::test_round_time PASSED [ 50%]
tests/utils/test_dates.py::TestDates::test_infer_time_unit PASSED [ 75%]
tests/utils/test_dates.py::TestDates::test_scale_time_units PASSED [100%]
================================================================== 4 passed in 3.30s ===================================================================
- Running All the tests with Breeze by specifying the required python version, backend, backend version
breeze --backend postgres --postgres-version 15 --python 3.10 --db-reset testing tests --test-type All
Running specific type of test
breeze --backend postgres --postgres-version 15 --python 3.10 --db-reset testing tests --test-type Core
Running Integration test for specific test type
breeze --backend postgres --postgres-version 15 --python 3.10 --db-reset testing tests --test-type All --integration mongo
For more information on Testing visit 09_testing.rst
Similarly to regular development, you can also debug while testing using your IDE, for more information, you may refer to
- To know how to contribute to the project visit README.rst
Following are some of the important links of Contribution documentation
Go to your GitHub account and open your fork project and click on Branches
Click on
New pull requestbutton on branch from which you want to raise a pull requestAdd title and description as per Contributing guidelines and click on
Create pull request
Often it takes several days or weeks to discuss and iterate with the PR until it is ready to merge.
In the meantime new commits are merged, and you might run into conflicts, therefore you should periodically
synchronize main in your fork with the apache/airflow main and rebase your PR on top of it. Following
describes how to do it.
If you are familiar with Python development and use your favourite editors, Airflow can be setup similarly to other projects of yours. However, if you need specific instructions for your IDE you will find more detailed instructions here:
In order to use remote development environment, you usually need a paid account, but you do not have to setup local machine for development.
Once you have your environment set up, you can start contributing to Airflow. You can find more about ways you can contribute in the How to contribute document.








