Rerun Airflow DAGs and tasks
You can set when to run Airflow DAGs using a wide variety of scheduling options. Some uses cases where you might want tasks or DAGs to run outside of their regular schedule include:
- You want one or more tasks to automatically run again if they fail.
- You need to manually rerun a failed task for one or multiple DAG runs.
- You want to deploy a DAG with a start date of one year ago and trigger all DAG runs that would have been scheduled in the past year.
- You have a running DAG and realize you need it to process data for two months prior to the DAG's start date.
In this guide, you'll learn how to configure automatic retries, rerun tasks or DAGs, trigger historical DAG runs, and review the Airflow concepts of catchup and backfill.
To get the most out of this guide, you should have an understanding of:
- DAG scheduling. See Scheduling and timetables in Airflow
Automatically retry tasks
In Airflow, you can configure individual tasks to retry automatically in case of a failure. The default number of times a task will retry before failing permanently can be defined at the Airflow configuration level using the core config
default_task_retries. You can set this configuration either in
airflow.cfg or with the environment variable
You can overwrite the
default_task_retries of an Airflow environment at the task level by using the
retry_delay parameter (default:
timedelta(seconds=300)) defines the time spent between retries. As of Airflow 2.6, you can set a maximum value for the retry delay in the core Airflow config
AIRFLOW__CORE__MAX_TASK_RETRY_DELAY), which, by default, is set at 24 hours. Or, for individual tasks, you can set the maximum retry delay with the parameter,
To progressively increase the wait time between retries until
max_retry_delay is reached, set
It is common practice to set the number of retries for all tasks in a DAG by using
default_args and override it for specific tasks as needed. To override specific tasks, provide a different value to the task level
The DAG below contains 4 tasks that will always fail. Each of the tasks uses a different retry parameter configuration.
from airflow.decorators import dag
from airflow.operators.bash import BashOperator
from pendulum import datetime, duration
start_date=datetime(2023, 4, 1),
t1 = BashOperator(task_id="t1", bash_command="echo I get 3 retries! && False")
t2 = BashOperator(
bash_command="echo I get 6 retries and never wait long! && False",
t3 = BashOperator(
bash_command="echo I wait exactly 20 seconds between each of my 4 retries! && False",
t4 = BashOperator(
bash_command="echo I have to get it right the first time! && False",
Manually rerun tasks or DAGs
Rerunning tasks or full DAGs in Airflow is a common workflow.
To rerun a task in Airflow you clear the task status to update the
max_tries and current task instance state values in the metastore. After the task reruns, the
max_tries value updates to
0, and the current task instance state updates to
To clear the task status, go to the Grid View in the Airflow UI, select the task instance you want to rerun and click the Clear task button.
A popup window appears, giving you the following options to clear and rerun additional task instances related to the selected task:
- Past: Clears any instances of the task in DAG runs with a data interval before the selected task instance.
- Future: Clears any instances of the task in DAG runs with a data interval after the selected task instance.
- Upstream: Clears any tasks in the current DAG run which are upstream from the selected task instance.
- Downstream: Clears any tasks in the current DAG run which are downstream from the selected task instance.
- Recursive: Clears any task instances of the task in the child DAG and any parent DAGs if you have cross-DAG dependencies.
- Only Failed: Clears only failed instances of any task instances selected based on the above options.
Check which task instances will be cleared with the current settings by expanding the dropdown menu Affected tasks: X. Click Clear to confirm and the task(s) will be cleared and rescheduled for another run.
You can also clear task statuses fromm the Graph view.
If you want to use the Airflow CLI to clear task statuses, you can use the following command:
airflow tasks clear [-h] [-R] [-d] [-e END_DATE] [-X] [-x] [-f] [-r]
[-s START_DATE] [-S SUBDIR] [-t TASK_REGEX] [-u] [-y]
For more info on the positional arguments for the
clear command, see the Airflow documentation.
To clear a full DAG run, go to the Grid View in the Airflow UI, click on the DAG run, and then click Clear as shown in the following image.
Don't clear or change task statuses directly in the Airflow metastore. This can cause unexpected behavior in Airflow.
Add notes to cleared tasks and DAGs
As of Airflow 2.5, you can add notes to task instances and DAG runs from the Grid View in the Airflow UI. This feature is useful for tracking manual changes to task instances, such as reruns or task status changes. Astronomer recommends leaving a note on a task or DAG whenever you manually update a task instance through the Airflow UI.
To add a note to a task instance or DAG run:
- Go to the Grid View of the Airflow UI.
- Select a task instance or DAG run.
- Click Details > Task Instance Notes or DAG Run notes > Add Note.
- Write a note and click Save Note.
If a task instance or DAG run has a note, its grid box is marked with a grey corner.
This feature is useful for tracking and maintaining visibility of manual changes made to task instances such as rerunning or changing the task status. We recommend utilizing this feature any time task or DAG statuses are altered.
Clear all tasks
- In the Airflow UI, go to Browse > Task Instances.
- Select the tasks to rerun.
- In the Actions list select Clear.
To rerun multiple DAGs, click Browse > DAG Runs, select the DAGs to rerun, and in the Actions list select Clear the state.
You can use the built-in catchup DAG argument to process data for logical dates between the set
start_date of a DAG and the current date.
When the catchup parameter for a DAG is set to
True, at the time the DAG is turned on in Airflow, the scheduler starts a DAG run for every data interval that has not been run between the DAG's
start_date and the current data interval. For example, if your DAG is scheduled to run daily and has a
start_date of 1/1/2023, and you deploy that DAG and turn it on 2/1/2023, Airflow will schedule and start all of the daily DAG runs for January. Catchup is also triggered when you turn a DAG off for a period and then turn it on again.
Catchup can be controlled by setting the parameter in your DAG's arguments. By default, catchup is set to
True. This example DAG doesn't use catchup:
start_date=datetime(2023, 4, 23),
Catchup is a powerful feature, but it should be used with caution. For example, if you deploy a DAG that runs every 5 minutes with a start date of 1 year ago and don't set catchup to
False, Airflow will schedule numerous DAG runs all at once. When using catchup, keep in mind what resources Airflow has available and how many DAG runs you can support at one time. To avoid overloading your scheduler or external systems, you can use the following parameters in conjunction with catchup:
max_active_runs: Set at the DAG level and limits the number of DAG runs that Airflow will execute for that particular DAG at any given time. For example, if you set this value to 3 and the DAG had 15 catchup runs to complete, they would be executed in 5 chunks of 3 runs.
depends_on_past: Set at the task level or as a
default_argfor all tasks at the DAG level. When set to
True, the task instance must wait for the same task in the most recent DAG run to be successful. This ensures sequential data loads and allows only one DAG run to be executed at a time in most cases.
wait_for_downstream: Set at the DAG level and similar to a DAG-level implementation of
depends_on_past. The entire DAG needs to run successfully for the next DAG run to start.
catchup_by_default: Set at the Airflow level in your
airflow.cfgor as an environment variable. If you set this parameter to
Falseall DAGs in your Airflow environment will not catchup unless you turn it on.
If you want to deploy your DAG with catchup enabled but there are some tasks you don't want to run during the catchup, you can use the
LatestOnlyOperator in your DAG. This operator only runs during the DAG's most recent scheduled interval. In every other DAG run it is ignored, along with any tasks downstream of it.
Backfilling lets you use a DAG to process data prior to the DAG's start date. Backfilling is the concept of running a DAG for a specified historical period. Unlike catchup, which triggers missed DAG runs from the DAG's
start_date through the current data interval, backfill periods can be specified explicitly and can include periods prior to the DAG's
Backfilling can be accomplished in Airflow using the CLI by specifying the DAG ID and the start and end dates for the backfill period. This command runs the DAG for all intervals between the start date and end date. DAGs in your backfill interval are still rerun even if they already have DAG runs. For example:
airflow dags backfill [-h] [-c CONF] [--delay-on-limit DELAY_ON_LIMIT] [-x]
[-n] [-e END_DATE] [-i] [-I] [-l] [-m] [--pool POOL]
[--rerun-failed-tasks] [--reset-dagruns] [-B]
[-s START_DATE] [-S SUBDIR] [-t TASK_REGEX] [-v] [-y]
airflow dags backfill -s 2023-04-01 -e 2021-04-02 example_dag backfills
example_dag from April 1st-2nd 2023. For more information about the backfill parameter, see the Airflow documentation.
When using backfill keep the following considerations in mind:
- Consider your available resources. If your backfill will trigger many DAG runs, you might want to add some of the catchup parameters to your DAG.
- Clearing the task or DAG status of a backfilled DAG run does not rerun the task or DAG.
Alternatively, you can deploy a copy of the DAG using a new name with the start date as the date you want to backfill to. Airflow considers this a separate DAG, so you won't see all the DAG runs and task instances in the same place, but it runs the DAG for data in the desired time period. If you have a small number of DAG runs to backfill, you can trigger them manually from the Airflow UI via
Trigger DAG w/ config and choose the desired logical date as shown in the following image: