Visit the full docs here to see additional examples and the API reference.
With prefect-dbt, you can easily trigger and monitor dbt Cloud jobs, execute dbt Core CLI commands, and incorporate other services, like Snowflake, into your dbt runs!
Check out the examples below to get started!
Be sure to install prefect-dbt and save a block to run the examples below!
If you have an existing dbt Cloud job, take advantage of the flow, run_dbt_cloud_job
.
This flow triggers the job and waits until the job run is finished.
If certain nodes fail, run_dbt_cloud_job
efficiently retries the specific, unsuccessful nodes.
from prefect import flow
from prefect_dbt.cloud import DbtCloudJob
from prefect_dbt.cloud.jobs import run_dbt_cloud_job
@flow
def run_dbt_job_flow():
result = run_dbt_cloud_job(
dbt_cloud_job=DbtCloudJob.load("my-block-name"),
targeted_retries=5,
)
return result
run_dbt_job_flow()
prefect-dbt
also supports execution of dbt Core CLI commands.
To get started, if you don't have a DbtCoreOperation
block already saved,
set the commands that you want to run; it can include a mix of dbt and non-dbt commands.
Then, optionally specify the project_dir
.
If profiles_dir
is unset, it will try to use the DBT_PROFILES_DIR
environment variable.
If that's also not set, it will use the default directory $HOME/.dbt/
.
If you already have an existing dbt profile, specify the profiles_dir
where profiles.yml
is located.
from prefect import flow
from prefect_dbt.cli.commands import DbtCoreOperation
@flow
def trigger_dbt_flow() -> str:
result = DbtCoreOperation(
commands=["pwd", "dbt debug", "dbt run"],
project_dir="PROJECT-DIRECTORY-PLACEHOLDER",
profiles_dir="PROFILES-DIRECTORY-PLACEHOLDER"
).run()
return result
trigger_dbt_flow()
To setup a new profile, first save and load a DbtCliProfile block and use it in DbtCoreOperation
.
Then, specify profiles_dir
where profiles.yml
will be written.
from prefect import flow
from prefect_dbt.cli import DbtCliProfile, DbtCoreOperation
@flow
def trigger_dbt_flow():
dbt_cli_profile = DbtCliProfile.load("DBT-CORE-OPERATION-BLOCK-NAME-PLACEHOLDER")
with DbtCoreOperation(
commands=["dbt debug", "dbt run"],
project_dir="PROJECT-DIRECTORY-PLACEHOLDER",
profiles_dir="PROFILES-DIRECTORY-PLACEHOLDER",
dbt_cli_profile=dbt_cli_profile,
) as dbt_operation:
dbt_process = dbt_op.trigger()
# do other things before waiting for completion
dbt_process.wait_for_completion()
result = dbt_process.fetch_result()
return result
trigger_dbt_flow()
If you need help getting started with or using dbt, please consult the dbt documentation.
To use prefect-dbt
with dbt Cloud:
pip install prefect-dbt
To use dbt Core (CLI):
pip install "prefect-dbt[cli]"
To use dbt Core with Snowflake profiles:
pip install "prefect-dbt[snowflake]"
To use dbt Core with BigQuery profiles:
pip install "prefect-dbt[bigquery]"
To use dbt Core with Postgres profiles:
pip install "prefect-dbt[postgres]"
!!! warning "Some dbt Core profiles require additional installation"
According to dbt's [Databricks setup page](https://docs.getdbt.com/reference/warehouse-setups/databricks-setup), users must first install the adapter:
```bash
pip install dbt-databricks
```
Check out the [desired profile setup page](https://docs.getdbt.com/reference/profiles.yml) on the sidebar for others.
Requires an installation of Python 3.7+.
We recommend using a Python virtual environment manager such as pipenv, conda or virtualenv.
These tasks are designed to work with Prefect 2. For more information about how to use Prefect, please refer to the Prefect documentation.
Note, to use the load
method on Blocks, you must already have a block document saved through code or saved through the UI.
!!! info "Registering blocks"
Register blocks in this module to
[view and edit them](https://orion-docs.prefect.io/ui/blocks/)
on Prefect Cloud:
```bash
prefect block register -m prefect_dbt
```
A list of available blocks in prefect-dbt
and their setup instructions can be found here.
To create a dbt Cloud credentials block:
- Head over to your dbt Cloud profile.
- Login to your dbt Cloud account.
- Scroll down to "API" or click "API Access" on the sidebar.
- Copy the API Key.
- Click Projects on the sidebar.
- Copy the account ID from the URL:
https://cloud.getdbt.com/settings/accounts/<ACCOUNT_ID>
. - Create a short script, replacing the placeholders.
from prefect_dbt.cloud import DbtCloudCredentials
DbtCloudCredentials(
api_key="API-KEY-PLACEHOLDER",
account_id="ACCOUNT-ID-PLACEHOLDER"
).save("BLOCK-NAME-PLACEHOLDER")
Then, to create a dbt Cloud job block:
- Head over to your dbt home page.
- On the top nav bar, click on Deploy -> Jobs.
- Select a job.
- Copy the job ID from the URL:
https://cloud.getdbt.com/deploy/<ACCOUNT_ID>/projects/<PROJECT_ID>/jobs/<JOB_ID>
- Create a short script, replacing the placeholders.
from prefect_dbt.cloud import DbtCloudCredentials, DbtCloudJob
dbt_cloud_credentials = DbtCloudCredentials.load("BLOCK-NAME-PLACEHOLDER")
dbt_cloud_job = DbtCloudJob.load(
dbt_cloud_credentials=dbt_cloud_credentials,
job_id="JOB-ID-PLACEHOLDER"
)
Congrats! You can now easily load the saved block, which holds your credentials:
from prefect_dbt.cloud import DbtCloudJob
DbtCloudJob.load("BLOCK-NAME-PLACEHOLDER")
!!! info "Available TargetConfigs
blocks"
The following may vary slightly depending on the service you want to incorporate.
Visit the [API Reference](cli/configs/base) to see other built-in `TargetConfigs` blocks.
If the desired service profile is not available, check out the
[Examples Catalog](examples_catalog/#clicredentials-module) to see how you can
build one from the generic `TargetConfigs` class.
To create dbt Core target config and profile blocks for BigQuery:
- Save and load a
GcpCredentials
block. - Determine the schema / dataset you want to use in BigQuery.
- Create a short script, replacing the placeholders.
from prefect_gcp.credentials import GcpCredentials
from prefect_dbt.cli import BigQueryTargetConfigs, DbtCliProfile
credentials = GcpCredentials.load("CREDENTIALS-BLOCK-NAME-PLACEHOLDER")
target_configs = BigQueryTargetConfigs(
schema="SCHEMA-NAME-PLACEHOLDER", # also known as dataset
credentials=credentials,
)
target_configs.save("TARGET-CONFIGS-BLOCK-NAME-PLACEHOLDER")
dbt_cli_profile = DbtCliProfile(
name="PROFILE-NAME-PLACEHOLDER",
target="TARGET-NAME-placeholder",
target_configs=target_configs,
)
dbt_cli_profile.save("DBT-CLI-PROFILE-BLOCK-NAME-PLACEHOLDER")
Then, to create a dbt Core operation block:
- Determine the dbt commands you want to run.
- Create a short script, replacing the placeholders.
from prefect_dbt.cli import DbtCliProfile, DbtCoreOperation
dbt_cli_profile = DbtCliProfile.load("DBT-CLI-PROFILE-BLOCK-NAME-PLACEHOLDER")
dbt_core_operation = DbtCoreOperation(
commands=["DBT-CLI-COMMANDS-PLACEHOLDER"],
dbt_cli_profile=dbt_cli_profile,
overwrite_profiles=True,
)
dbt_core_operation.save("DBT-CORE-OPERATION-BLOCK-NAME-PLACEHOLDER")
Congrats! You can now easily load the saved block, which holds your credentials:
from prefect_dbt.cloud import DbtCoreOperation
DbtCoreOperation.load("DBT-CORE-OPERATION-BLOCK-NAME-PLACEHOLDER")
If you encounter any bugs while using prefect-dbt
, feel free to open an issue in the prefect-dbt repository.
If you have any questions or issues while using prefect-dbt
, you can find help in either the Prefect Discourse forum or the Prefect Slack community.
Feel free to star or watch prefect-dbt
for updates too!
If you'd like to help contribute to fix an issue or add a feature to prefect-dbt
, please propose changes through a pull request from a fork of the repository.
Here are the steps:
- Fork the repository
- Clone the forked repository
- Install the repository and its dependencies:
pip install -e ".[dev]"
- Make desired changes
- Add tests
- Insert an entry to CHANGELOG.md
- Install
pre-commit
to perform quality checks prior to commit:
pre-commit install
git commit
,git push
, and create a pull request