This repository contains a number of sample projects for Conveyor
- Clone this repository
- Verify you have the Conveyor CLI installed by executing
conveyor doctor
- Authenticate the Conveyor CLI
conveyor auth login
- Create a new conveyor environment
conveyor environment create --name samples
- Execute the additional steps listed in the prerequisites folder
- pi_spark: use Apache Spark to calculate pi.
- first_project_dbt: use dbt and DuckDB for the first time. Using this project is described in the Conveyor getting started guide.
- coffee_shop_dbt: use dbt and DuckDB for cleaning and transforming the coffee shop input data and writing the results to S3.
- titanic: use the dataset from Kaggle to build a ML pipeline that will produce a model to predict survival.
- housing: use the dataset from Kaggle to build a ML pipeline that will produce a model to predict the housing prices.
- alerting: use native Apache Airflow features to send notifications to Slack.