- Notifications
You must be signed in to change notification settings - Fork0
Automate your data pipelines using Apache Airflow with this ready-to-use DAG for data integration, ETL and workflow automation.
License
AnthonyByansi/Airflow-Data-Pipeline-Automation
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
This project provides a workflow using Apache Airflow to automate the data pipeline for a specific use case. The purpose of this project is to demonstrate how Airflow can be used to automate the process of extracting, transforming, and loading data from various sources into a data warehouse or data lake.
To get started with this project, you will need to have Apache Airflow installed on your system. You can follow the installation instructions provided by the official Airflow documentationhere.
Once you have Airflow installed, you can clone this repository to your local machine and create a new virtual environment using the requirements.txt file provided:
$ git clone https://github.com/AnthonyByansi/airflow-data-pipeline-automation.git$cd airflow-data-pipeline-automation$ python -m venv venv$source venv/bin/activate$ pip install -r requirements.txt
Next, you can initialize the Airflow database and start the Airflow web server:
$ airflow db init$ airflow webserver --port 8080
Finally, you can trigger the DAG by navigating to the Airflow web interface athttp://localhost:8080
and clicking on the "Trigger DAG" button. This will start the process of extracting, transforming, and loading data according to the specific use case defined in the DAG.
If you are interested in contributing to this project, please feel free to submit a pull request. We welcome contributions of all kinds, including bug fixes, new features, and documentation improvements.
Before submitting a pull request, please make sure to run the tests and format your code usingflake8
andblack
:
$ pytest$ flake8$ black.
This project is licensed under the MIT License - see theLICENSE file for details.
About
Automate your data pipelines using Apache Airflow with this ready-to-use DAG for data integration, ETL and workflow automation.
Topics
Resources
License
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Releases
Packages0
Uh oh!
There was an error while loading.Please reload this page.
Contributors2
Uh oh!
There was an error while loading.Please reload this page.