Skip to content

Skeleton project for Apache Airflow training participants to work on.

License

Notifications You must be signed in to change notification settings

godatadriven/airflow-training-skeleton

Repository files navigation

Apache Airflow training

This project serves as a skeleton project on which Airflow training participants build exercises. There are two versions, one cloud based training and one on-premise based training.

Cloud Composer based training

At the start of the training, you will receive a Google Cloud Platform project, with a pre-configured Cloud Composer environment (Google's managed Airflow).

Training participants generally like taking their code home after the training, so we provide them with a skeleton Airflow project on GitHub. In order to push and deploy, we kindly ask the participants to setup their own CI because we don't know the GitHub handles of participants beforehand. However the setup is just a few clicks, described below.

A CI pipeline is included with the project (cloudbuild.yml), which defines steps to execute after each push. However, there are variables which must be entered before applying the CI pipeline. The CI pipeline tests and deploys your code to Google Cloud Composer.

  1. Fork the repository.
  2. In the GCP console, go to your GCP project, and browse to Cloud Build.
  3. Go to Build triggers.
  4. Click "Create trigger".
  5. Select source GitHub and click Continue.
  6. You have to allow Google Cloud Platform to access your repositories.
  7. In Cloud Build, you can now select your repositories. Select this forked repository and check the consent box.
  8. Under Build configuration, select cloudbuild.yaml and click Create trigger at the bottom.

Done. The just configured build trigger now watches changes on your repository.

On-Premise training

The on premise based training will run Apache Airflow based on a Docker stack. This includes docker-compose to hook up multiple containers to simulate a multi container setup. A guide on how to setup k8s locally: https://xebia.com/blog/running-kubernetes-locally-docker-mac-os-x/

git clone [email protected]:godatadriven/airflow-training-skeleton.git
cd airflow-training-skeleton/
kubectl apply -f k8s/

Hit up a browser and browse to http://localhost:30808/admin/ to see the admin screen.

About

Skeleton project for Apache Airflow training participants to work on.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages