-
Notifications
You must be signed in to change notification settings - Fork 161
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ADAP-433] [Feature] Granular execution project #647
Comments
Thanks for making this feature suggestion @VasiliiSurov ! Are you thinking this would be similar to the I read briefly about BigQuery editions, but I couldn't determine how to specify the "execution project" differently from the "dataset project" in BigQuery. Do you know how? e.g., can you explain what the API call, Python connection parameters, or SQL session parameters would look like? |
@dbeatty10
The way how I saw it we will need more than a single connection, but a connection per |
Thanks for connecting the dots for me @VasiliiSurov ! In Snowflake, we can do something like this: use warehouse my_purple_tshirt Do you know if we can do something similar in BigQuery? SET @@dataset_project_id = 'MyProject'; If not, then the comments here and here will be especially relevant. Either way, it sounds like you are asking the similar/same thing as #343 (which itself is similar to databricks/dbt-databricks#59). |
Looks like this one is very much a like what I'm asking. As far as I know you can define "compute" project only on a connection level and once defined it can't be overwritten on a job level. That's why I was proposing to have a set of connections instead of a single one in current implementation. |
I'd like this, too :) It's going to be a much bigger lift, but something that I'd like to see us building over the medium term. |
If BQ's doing it, it feels more like a trend than just Snowflake? So should we consider moving the choice of "compute environment" at the model level across the board, rather than the connection level? Intuitively, if most platforms would allow the switch to be made in session, like Snowflake with 'USE WAREHOUSE', that would be an easy pull. But for something like BQ, if that means spinning up and down heterogeneous connections as we move through the DAG (or keeping them alive and using them as we go I guess), that sounds much more interesting. @jtcohen6 do you think that's a core thing, or a BQ thing? Personally I would start here first, and generalize if that makes sense later. We have enough top down initiatives in flight anyway. @VasiliiSurov I'm thinking that the next step is to spike it on our side to map the territory. From there, we may use your help, but that could be a complicated one. |
This issue has been marked as Stale because it has been open for 180 days with no activity. If you would like the issue to remain open, please comment on the issue or else it will be closed in 7 days. |
Although we are closing this issue as stale, it's not gone forever. Issues can be reopened if there is renewed community interest. Just add a comment to notify the maintainers. |
Although we are closing this issue as stale, it's not gone forever. Issues can be reopened if there is renewed community interest. Just add a comment to notify the maintainers. |
Is this your first time submitting a feature request?
Describe the feature
With Google introducing new BigQuery editions we can make dbt run more cost efficient if we could manage execution project on a model like, somewhat how it's implemented in Snowflake adapter.
Some of the models are CPU-bound and will cost less if they are executed using on-demand price model, other could by IO-bound and will cost less if executed on Standard/Enterprise Edition.
This feature should work as a new model configuration that can be set on a project / folder / model level.
Describe alternatives you've considered
No dbt-native support, only artificially split dbt DAG into slices and use external orchestration with multiple targets
Who will this benefit?
All users of dbt-bigquery adapter who can invest time into advanced cost management
Are you interested in contributing this feature?
Yes, if provided some guidence.
Anything else?
No response
The text was updated successfully, but these errors were encountered: