Set default file format to parquet #422
Closed
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
resolves #363
Description
We already set this default here, but it's clearly not flowing through properly, because
AdapterConfig
doesn't really seem to be working indbt-core
(dbt-labs/dbt-core#5236):dbt-spark/dbt/adapters/spark/impl.py
Line 38 in 8744cf1
Let's set the default explicitly in the macro, so that if the user hasn't provided a
file_format
,dbt-spark
uses Parquet. This prevents weird issues with query rewrite in Databricks SQL warehouses, too (#236).Update: If the user is connecting via
ODBC
(= Databricks), let's usedelta
as the default file format instead. This helps out Databricks SQL warehouse connections, which can only write Delta-formatted tables.Checklist
CHANGELOG.md
and added information about my change to the "dbt-spark next" section.