Kedro-Airflow makes it easy to deploy Kedro projects to Airflow
Project description
Kedro-Airflow
develop |
master |
---|---|
Apache Airflow is a tool for orchestrating complex workflows and data processing pipelines. The Kedro-Airflow plugin can be used for:
- Rapid pipeline creation in the prototyping phase. You can write Python functions in Kedro without worrying about schedulers, daemons, services or having to recreate the Airflow DAG file.
- Automatic dependency resolution in Kedro. This allows you to bypass Airflow's need to specify the order of your tasks.
- Distributing Kedro tasks across many workers. You can also enable monitoring and scheduling of the tasks' runtimes.
How do I install Kedro-Airflow?
kedro-airflow
is a Python plugin. To install it:
pip install kedro-airflow
How do I use Kedro-Airflow?
The Kedro-Airflow plugin adds a kedro airflow create
CLI command that generates an Airflow DAG file in the airflow_dags
folder of your project. At runtime, this file translates your Kedro pipeline into Airflow Python operators. This DAG object can be modified according to your needs and you can then deploy your project to Airflow by running kedro airflow deploy
.
Prerequisites
The following conditions must be true for Airflow to run your pipeline:
- Your project directory must be available to the Airflow runners in the directory listed at the top of the DAG file.
- Your source code must be on the Python path (by default the DAG file takes care of this).
- All datasets must be explicitly listed in
catalog.yml
and reachable for the Airflow workers. Kedro-Airflow does not supportMemoryDataSet
or datasets that require Spark. - All local paths in configuration files (notably in
catalog.yml
andlogging.yml
) should be absolute paths and not relative paths.
Process
- Run
kedro airflow create
to generate a DAG file for your project. - If needed, customize the DAG file as described below.
- Run
kedro airflow deploy
which will copy the DAG file from theairflow_dags
folder in your Kedro project into thedags
folder in the Airflow home directory.
Note: The generated DAG file will be placed in
$AIRFLOW_HOME/dags/
whenkedro airflow deploy
is run, whereAIRFLOW_HOME
is an environment variable. If the environment variable is not defined, Kedro-Airflow will create~/airflow
and~/airflow/dags
(if required) and copy the DAG file into it.
Customization
There are a number of items in the DAG file that you may want to customize including:
- Source location,
- Project location,
- DAG construction,
- Default operator arguments,
- Operator-specific arguments,
- And / or Airflow context and execution date.
The following sections guide you to the appropriate location within the file.
Source location
The line sys.path.append("/Users/<user-name>/new-kedro-project/src")
enables Python and Airflow to find your project source.
Project location
The line project_path = "/Users/<user-name>/new-kedro-project"
sets the location for your project directory. This is passed to your get_config
method.
DAG construction
The construction of the actual DAG object can be altered as needed. You can learn more about how to do this by going through the Airflow tutorial.
Default operator arguments
The default arguments for the Airflow operators are contained in the default_args
dictionary.
Operator-specific arguments
The operator_specific_arguments
callback is called to retrieve any additional arguments specific to individual operators. It is passed the Airflow task_id
and should return a dictionary of additional arguments. For example, to change the number of retries on node named analysis
to 5 you may have:
def operator_specific_arguments(task_id):
if task_id == "analysis":
return {"retries": 5}
return {}
The easiest way to find the correct task_id
is to use Airflow's list_tasks
command.
Airflow context and execution date
The process_context
callback provides a hook for ingesting Airflow's Jinja context. It is called before every node, receives the context and catalog and must return a catalog. A common use of this is to pick up the execution date and either insert it into the catalog or modify the catalog based on it.
The list of default context variables is available in the Airflow documentation.
What licence do you use?
Kedro-Airflow is licensed under the Apache 2.0 License.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file kedro-airflow-0.3.0.tar.gz
.
File metadata
- Download URL: kedro-airflow-0.3.0.tar.gz
- Upload date:
- Size: 8.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.2.0 pkginfo/1.5.0.1 requests/2.24.0 setuptools/47.3.1.post20200622 requests-toolbelt/0.9.1 tqdm/4.47.0 CPython/3.6.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 64ccf98bd01042f6540e3144314640b4f2e94820ca510962edad217ebdebcae1 |
|
MD5 | d6c2e59acdc2fb71aac767f8f03fb676 |
|
BLAKE2b-256 | 2f6c7cfe36d7f150fa847bf463bd14c9cd8c003a68befd53983d0638af42a0ba |
File details
Details for the file kedro_airflow-0.3.0-py3-none-any.whl
.
File metadata
- Download URL: kedro_airflow-0.3.0-py3-none-any.whl
- Upload date:
- Size: 11.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.2.0 pkginfo/1.5.0.1 requests/2.24.0 setuptools/47.3.1.post20200622 requests-toolbelt/0.9.1 tqdm/4.47.0 CPython/3.6.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | edad02794164c030eccd0b118383465a16898b6b99cc501780489cc1d591f8f7 |
|
MD5 | 379246e4f06c0825b4d4d69bf9a5be7a |
|
BLAKE2b-256 | 12196e2076008d03d67149f6ce6fa7c267c11e5e8e13c2bc0485b5c452224260 |