- Notifications
You must be signed in to change notification settings - Fork23
A Python framework for data processing on GCP.
License
allegro/bigflow
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
- What is BigFlow?
- Getting started
- Installing Bigflow
- Help me
- BigFlow tutorial
- CLI
- Configuration
- Project structure and build
- Deployment
- Workflow & Job
- Starter
- Technologies
- Development
BigFlow is a Python framework for data processing pipelines onGCP.
The main features are:
- Dockerized deployment environment
- Powerful CLI
- Automated build,deployment,versioning andconfiguration
- Unified project structure
- Support for GCP data processing technologies —Dataflow (Apache Beam) andBigQuery
- Project starter
Start from installing BigFlow on your local machine.Next, go through the BigFlowtutorial.
Prerequisites. Before you start, make sure you have the following software installed:
You can install thebigflow
package globally, but we recommendinstalling it locally withvenv
, in your project's folder:
python -m venv .bigflow_envsource .bigflow_env/bin/activate
Install thebigflow
PIP package:
pip install bigflow[bigquery,dataflow]
Test it:
bigflow -h
Read more aboutBigFlow CLI.
To interact with GCP you need to set a default project and log in:
gcloud configset project<your-gcp-project-id>gcloud auth application-default login
Finally, check if your Docker is running:
docker info
You can ask questions on ourgitter channel orstackoverflow.
About
A Python framework for data processing on GCP.
Topics
Resources
License
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Uh oh!
There was an error while loading.Please reload this page.