diff --git a/docs/content/concepts/assets/asset-materializations.mdx b/docs/content/concepts/assets/asset-materializations.mdx index 1b6b72f3a3dc4..4aec0bee57425 100644 --- a/docs/content/concepts/assets/asset-materializations.mdx +++ b/docs/content/concepts/assets/asset-materializations.mdx @@ -114,7 +114,7 @@ There are a variety of types of metadata that can be associated with a materiali #### Example: Op body ```python file=concepts/assets/materialization_ops.py startafter=start_materialization_ops_marker_2 endbefore=end_materialization_ops_marker_2 -from dagster import op, AssetMaterialization, MetadataValue +from dagster import AssetMaterialization, MetadataValue, op @op @@ -173,7 +173,7 @@ Check our API docs for for If you are materializing a single slice of an asset (e.g. a single day's worth of data on a larger table), rather than mutating or creating it entirely, you can indicate this to Dagster by including the `partition` argument on the object. ```python file=/concepts/assets/materialization_ops.py startafter=start_partitioned_asset_materialization endbefore=end_partitioned_asset_materialization -from dagster import op, AssetMaterialization +from dagster import AssetMaterialization, op @op(config_schema={"date": str}) diff --git a/docs/content/concepts/assets/asset-observations.mdx b/docs/content/concepts/assets/asset-observations.mdx index 9fef494fb51da..e9ed5cd4ccb44 100644 --- a/docs/content/concepts/assets/asset-observations.mdx +++ b/docs/content/concepts/assets/asset-observations.mdx @@ -55,7 +55,7 @@ height={917} There are a variety of types of metadata that can be associated with an observation event, all through the class. Each observation event optionally takes a dictionary of metadata entries that are then displayed in the event log and the [Asset Details](/concepts/dagit/dagit#asset-details) page. Check our API docs for for more details on the types of event metadata available. ```python file=concepts/assets/observations.py startafter=start_observation_asset_marker_2 endbefore=end_observation_asset_marker_2 -from dagster import op, AssetObservation, MetadataValue +from dagster import AssetObservation, MetadataValue, op @op @@ -93,7 +93,7 @@ height={1146} If you are observing a single slice of an asset (e.g. a single day's worth of data on a larger table), rather than mutating or creating it entirely, you can indicate this to Dagster by including the `partition` argument on the object. ```python file=/concepts/assets/observations.py startafter=start_partitioned_asset_observation endbefore=end_partitioned_asset_observation -from dagster import op, AssetMaterialization +from dagster import AssetMaterialization, op @op(config_schema={"date": str}) diff --git a/docs/content/concepts/assets/software-defined-assets.mdx b/docs/content/concepts/assets/software-defined-assets.mdx index ae78ba2bab8d1..6050d828e22cd 100644 --- a/docs/content/concepts/assets/software-defined-assets.mdx +++ b/docs/content/concepts/assets/software-defined-assets.mdx @@ -427,7 +427,7 @@ def nabisco_cereals(): **This recommended approach** constructs a group of assets from a specified module in your project. Using the `load_assets_from_package_module` function, you can import all assets in a module and apply a grouping: ```python file=/concepts/assets/asset_group_module.py startafter=start_example endbefore=end_example -import my_package.cereal as cereal +from my_package import cereal cereal_assets = load_assets_from_package_module( cereal, diff --git a/docs/content/concepts/dagit/graphql-client.mdx b/docs/content/concepts/dagit/graphql-client.mdx index a450c3a201ba3..bcceb1d0dfb56 100644 --- a/docs/content/concepts/dagit/graphql-client.mdx +++ b/docs/content/concepts/dagit/graphql-client.mdx @@ -63,6 +63,7 @@ You can use the client to get the status of a job run as follows: ```python file=/concepts/dagit/graphql/client_example.py startafter=start_run_status_marker endbefore=end_run_status_marker from dagster_graphql import DagsterGraphQLClientError + from dagster import DagsterRunStatus try: @@ -83,10 +84,7 @@ You can also reload a repository location in a Dagster deployment. This reloads all repositories in that repository location. This is useful in a variety of contexts, including refreshing Dagit without restarting the server. Example usage is as follows: ```python file=/concepts/dagit/graphql/client_example.py startafter=start_reload_repo_location_marker endbefore=end_reload_repo_location_marker -from dagster_graphql import ( - ReloadRepositoryLocationInfo, - ReloadRepositoryLocationStatus, -) +from dagster_graphql import ReloadRepositoryLocationInfo, ReloadRepositoryLocationStatus reload_info: ReloadRepositoryLocationInfo = client.reload_repository_location(REPO_NAME) if reload_info.status == ReloadRepositoryLocationStatus.SUCCESS: diff --git a/docs/content/concepts/logging/loggers.mdx b/docs/content/concepts/logging/loggers.mdx index 653a66994d9b8..6100bd9ac0b57 100644 --- a/docs/content/concepts/logging/loggers.mdx +++ b/docs/content/concepts/logging/loggers.mdx @@ -200,7 +200,7 @@ Default loggers can be specified on a Note that if you explicitly specify loggers on a job, they will override those provided to `default_logger_defs`. ```python file=/concepts/logging/custom_logger.py startafter=start_default_logger_repo endbefore=end_default_logger_repo -from dagster import repository, define_asset_job, asset +from dagster import asset, define_asset_job, repository @asset diff --git a/docs/content/concepts/ops-jobs-graphs/op-events.mdx b/docs/content/concepts/ops-jobs-graphs/op-events.mdx index bfd5e798616e3..b611694b2e5c3 100644 --- a/docs/content/concepts/ops-jobs-graphs/op-events.mdx +++ b/docs/content/concepts/ops-jobs-graphs/op-events.mdx @@ -159,7 +159,7 @@ To learn more about assets and how they are surfaced once you send this event, c Attaching metadata to Asset Materializations is an important way of tracking aspects of a given asset over time. This functions essentially identically to other events which accept a `metadata` parameter, allowing you to attach a set of structured labels and values to display. ```python file=concepts/assets/materialization_ops.py startafter=start_materialization_ops_marker_2 endbefore=end_materialization_ops_marker_2 -from dagster import op, AssetMaterialization, MetadataValue +from dagster import AssetMaterialization, MetadataValue, op @op diff --git a/docs/content/concepts/ops-jobs-graphs/op-hooks.mdx b/docs/content/concepts/ops-jobs-graphs/op-hooks.mdx index 4074ae09caa33..4e624c6956734 100644 --- a/docs/content/concepts/ops-jobs-graphs/op-hooks.mdx +++ b/docs/content/concepts/ops-jobs-graphs/op-hooks.mdx @@ -25,7 +25,17 @@ A , which is use You can use to test the type check function of a custom Dagster Type: ```python file=/concepts/types/types.py startafter=start_test_dagster_type endbefore=end_test_dagster_type -from dagster import check_dagster_type, Dict, Any +from dagster import Any, Dict, check_dagster_type def test_dagster_type(): diff --git a/docs/content/deployment/executors.mdx b/docs/content/deployment/executors.mdx index e56986124d730..d9663d660d98b 100644 --- a/docs/content/deployment/executors.mdx +++ b/docs/content/deployment/executors.mdx @@ -23,7 +23,7 @@ Every job has an executor. The default executor is the to the `executor_def` parameter of or . ```python file=/deploying/executors/executors.py startafter=start_executor_on_job endbefore=end_executor_on_job -from dagster import multiprocess_executor, job, graph +from dagster import graph, job, multiprocess_executor # Providing an executor using the job decorator @job(executor_def=multiprocess_executor) @@ -43,7 +43,7 @@ other_job = the_graph.to_job(executor_def=multiprocess_executor) A default executor can be specified for all jobs and assets provided to a repository using the `default_executor_def` argument of . All jobs that don't specify an executor will use this default executor, but if a job explicitly specifies an executor, then the default provided to the repository will not be used. ```python file=/deploying/executors/executors.py startafter=start_executor_on_repo endbefore=end_executor_on_repo -from dagster import multiprocess_executor, define_asset_job, asset, repository +from dagster import asset, define_asset_job, multiprocess_executor, repository @asset diff --git a/docs/content/guides/dagster/branch_deployments.mdx b/docs/content/guides/dagster/branch_deployments.mdx index 78711da258d98..770ce5d7e7487 100644 --- a/docs/content/guides/dagster/branch_deployments.mdx +++ b/docs/content/guides/dagster/branch_deployments.mdx @@ -167,7 +167,7 @@ def repo(): def get_current_env(): is_branch_depl = os.getenv("DAGSTER_CLOUD_IS_BRANCH_DEPLOYMENT") == "1" - assert is_branch_depl != None # env var must be set + assert is_branch_depl is not None # env var must be set return "branch" if is_branch_depl else "prod" return [ diff --git a/docs/content/guides/dagster/enriching-with-software-defined-assets.mdx b/docs/content/guides/dagster/enriching-with-software-defined-assets.mdx index 831f91de224fa..128c3ecbcbfa2 100644 --- a/docs/content/guides/dagster/enriching-with-software-defined-assets.mdx +++ b/docs/content/guides/dagster/enriching-with-software-defined-assets.mdx @@ -205,14 +205,14 @@ from .mylib import create_db_connection, pickle_to_s3, train_recommender_model @op def build_users(): - raw_users_df = read_sql(f"select * from raw_users", con=create_db_connection()) + raw_users_df = read_sql("select * from raw_users", con=create_db_connection()) users_df = raw_users_df.dropna() users_df.to_sql(name="users", con=create_db_connection()) @op(ins={"users": In(Nothing)}) def build_user_recommender_model(): - users_df = read_sql(f"select * from users", con=create_db_connection()) + users_df = read_sql("select * from users", con=create_db_connection()) users_recommender_model = train_recommender_model(users_df) pickle_to_s3(users_recommender_model, key="users_recommender_model") @@ -241,14 +241,14 @@ from .mylib import create_db_connection, pickle_to_s3, train_recommender_model @asset(non_argument_deps={"raw_users"}) def users(): - raw_users_df = read_sql(f"select * from raw_users", con=create_db_connection()) + raw_users_df = read_sql("select * from raw_users", con=create_db_connection()) users_df = raw_users_df.dropna() users_df.to_sql(name="users", con=create_db_connection()) @asset(non_argument_deps={"users"}) def user_recommender_model(): - users_df = read_sql(f"select * from users", con=create_db_connection()) + users_df = read_sql("select * from users", con=create_db_connection()) users_recommender_model = train_recommender_model(users_df) pickle_to_s3(users_recommender_model, key="users_recommender_model") diff --git a/docs/content/guides/dagster/re-execution.mdx b/docs/content/guides/dagster/re-execution.mdx index 48e6685f52587..bafe2a22d7363 100644 --- a/docs/content/guides/dagster/re-execution.mdx +++ b/docs/content/guides/dagster/re-execution.mdx @@ -80,9 +80,10 @@ Re-execution can be triggered via the API as well. Again, let's revist the job `unreliable_job`, which has a op named `unreliable`. ```python file=/guides/dagster/reexecution/reexecution_api.py startafter=start_initial_execution_marker endbefore=end_initial_execution_marker -from dagster import DagsterInstance, ReexecutionOptions, execute_job, reconstructable from docs_snippets.guides.dagster.reexecution.unreliable_job import unreliable_job +from dagster import DagsterInstance, ReexecutionOptions, execute_job, reconstructable + instance = DagsterInstance.ephemeral() # Initial execution diff --git a/docs/content/guides/dagster/software-defined-assets.mdx b/docs/content/guides/dagster/software-defined-assets.mdx index 64dc6c99a47d8..db337091a54eb 100644 --- a/docs/content/guides/dagster/software-defined-assets.mdx +++ b/docs/content/guides/dagster/software-defined-assets.mdx @@ -70,9 +70,10 @@ It's common to use a utility like ```python file=../../assets_pandas_pyspark/assets_pandas_pyspark/assets/weather_assets.py startafter=gather_assets_start endbefore=gather_assets_end # imports the module called "assets" from the package containing the current module # the "assets" module contains the asset definitions -from . import table_assets from dagster import load_assets_from_modules, with_resources +from . import table_assets + weather_assets = with_resources( load_assets_from_modules(modules=[table_assets]), resource_defs={ @@ -109,9 +110,11 @@ class LocalFileSystemIOManager(IOManager): Not all the assets in the same dependency graph need to have the same Python type. Here's an asset whose computation is defined using Spark DataFrames, that depends on the `daily_temperature_highs` asset we defined above using Pandas. ```python file=../../assets_pandas_pyspark/assets_pandas_pyspark/assets/spark_asset.py -from pyspark.sql import DataFrame as SparkDF -from pyspark.sql import Window -from pyspark.sql import functions as f +from pyspark.sql import ( + DataFrame as SparkDF, + Window, + functions as f, +) from dagster import asset @@ -132,9 +135,10 @@ def daily_temperature_high_diffs(daily_temperature_highs: SparkDF) -> SparkDF: Here's an extended version of `weather_assets` that contains the new asset: ```python file=../../assets_pandas_pyspark/assets_pandas_pyspark/assets/spark_weather_assets.py startafter=gather_assets_start endbefore=gather_assets_end -from . import table_assets, spark_asset from dagster import load_assets_from_modules, with_resources +from . import spark_asset, table_assets + spark_weather_assets = with_resources( load_assets_from_modules(modules=[table_assets, spark_asset]), resource_defs={ diff --git a/docs/content/integrations/dbt-cloud.mdx b/docs/content/integrations/dbt-cloud.mdx index cb4c321cf43a0..a0a3128206705 100644 --- a/docs/content/integrations/dbt-cloud.mdx +++ b/docs/content/integrations/dbt-cloud.mdx @@ -99,7 +99,7 @@ Now that your dbt Cloud assets are loaded, you can define a Dagster job that mat You can explicitly define when your assets should be materialized. For example, you can schedule assets based on their upstream or downstream dependencies, external events using a sensor, or a cron schedule. ```python startafter=start_schedule_dbt_cloud_assets endbefore=end_schedule_dbt_cloud_assets file=/integrations/dbt/dbt_cloud.py dedent=4 -from dagster import ScheduleDefinition, define_asset_job, repository, AssetSelection +from dagster import AssetSelection, ScheduleDefinition, define_asset_job, repository # Materialize all assets in the repository run_everything_job = define_asset_job("run_everything_job", AssetSelection.all()) diff --git a/examples/docs_snippets/docs_snippets/concepts/assets/asset_input_managers_numpy.py b/examples/docs_snippets/docs_snippets/concepts/assets/asset_input_managers_numpy.py index 04ea170352bd9..de368058b0c31 100644 --- a/examples/docs_snippets/docs_snippets/concepts/assets/asset_input_managers_numpy.py +++ b/examples/docs_snippets/docs_snippets/concepts/assets/asset_input_managers_numpy.py @@ -4,7 +4,11 @@ from dagster import AssetIn, IOManager, asset, io_manager, repository, with_resources -from .asset_input_managers import load_numpy_array, load_pandas_dataframe, store_pandas_dataframe +from .asset_input_managers import ( + load_numpy_array, + load_pandas_dataframe, + store_pandas_dataframe, +) # start_numpy_example diff --git a/examples/docs_snippets/docs_snippets/concepts/assets/subset_graph_backed_asset.py b/examples/docs_snippets/docs_snippets/concepts/assets/subset_graph_backed_asset.py index 87d2bb85b39d2..6845606b97408 100644 --- a/examples/docs_snippets/docs_snippets/concepts/assets/subset_graph_backed_asset.py +++ b/examples/docs_snippets/docs_snippets/concepts/assets/subset_graph_backed_asset.py @@ -1,4 +1,13 @@ -from dagster import AssetsDefinition, GraphOut, Out, Output, define_asset_job, graph, op, repository +from dagster import ( + AssetsDefinition, + GraphOut, + Out, + Output, + define_asset_job, + graph, + op, + repository, +) # start_graph_backed_asset_foo diff --git a/examples/docs_snippets/docs_snippets/concepts/dagit/graphql/client_example.py b/examples/docs_snippets/docs_snippets/concepts/dagit/graphql/client_example.py index 66a0ecc8e64fd..3e1b7ea8d3d19 100644 --- a/examples/docs_snippets/docs_snippets/concepts/dagit/graphql/client_example.py +++ b/examples/docs_snippets/docs_snippets/concepts/dagit/graphql/client_example.py @@ -89,7 +89,10 @@ def do_something_with_exc(some_exception): # pylint: disable=W0613 # end_reload_repo_location_marker # start_shutdown_repo_location_marker -from dagster_graphql import ShutdownRepositoryLocationInfo, ShutdownRepositoryLocationStatus +from dagster_graphql import ( + ShutdownRepositoryLocationInfo, + ShutdownRepositoryLocationStatus, +) shutdown_info: ShutdownRepositoryLocationInfo = client.shutdown_repository_location( REPO_NAME diff --git a/examples/docs_snippets/docs_snippets/concepts/ops_jobs_graphs/nested_graphs.py b/examples/docs_snippets/docs_snippets/concepts/ops_jobs_graphs/nested_graphs.py index db1bd1f1a2d18..12d95e1da4e20 100644 --- a/examples/docs_snippets/docs_snippets/concepts/ops_jobs_graphs/nested_graphs.py +++ b/examples/docs_snippets/docs_snippets/concepts/ops_jobs_graphs/nested_graphs.py @@ -4,7 +4,12 @@ from dagster import graph, job, op -from .unnested_ops import add_thirty_two, log_number, multiply_by_one_point_eight, return_fifty +from .unnested_ops import ( + add_thirty_two, + log_number, + multiply_by_one_point_eight, + return_fifty, +) # start_composite_solid_example_marker diff --git a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_asset_job.py b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_asset_job.py index 5996d11fe7d37..8d88e0778e381 100644 --- a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_asset_job.py +++ b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_asset_job.py @@ -1,4 +1,10 @@ -from dagster import AssetSelection, HourlyPartitionsDefinition, asset, define_asset_job, repository +from dagster import ( + AssetSelection, + HourlyPartitionsDefinition, + asset, + define_asset_job, + repository, +) hourly_partitions_def = HourlyPartitionsDefinition(start_date="2022-05-31-00:00") diff --git a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_job_test.py b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_job_test.py index 7ab6908632d9f..062e6ec90e798 100644 --- a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_job_test.py +++ b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_job_test.py @@ -1,6 +1,8 @@ # isort: skip_file -from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import do_stuff_partitioned +from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import ( + do_stuff_partitioned, +) # start diff --git a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/run_status_run_requests.py b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/run_status_run_requests.py index a5c128782c5c9..08c0de8713789 100644 --- a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/run_status_run_requests.py +++ b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/run_status_run_requests.py @@ -1,4 +1,10 @@ -from dagster import DagsterRunStatus, RunRequest, SkipReason, run_failure_sensor, run_status_sensor +from dagster import ( + DagsterRunStatus, + RunRequest, + SkipReason, + run_failure_sensor, + run_status_sensor, +) status_reporting_job = None diff --git a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/sensor_alert.py b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/sensor_alert.py index 223453f2e4fba..df8aace89b719 100644 --- a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/sensor_alert.py +++ b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/sensor_alert.py @@ -13,7 +13,10 @@ def my_slack_on_run_failure(context: RunFailureSensorContext): slack_client.chat_postMessage( channel="#alert-channel", - message=f'Job "{context.dagster_run.job_name}" failed. Error: {context.failure_event.message}', + message=( + f'Job "{context.dagster_run.job_name}" failed. Error:' + f" {context.failure_event.message}" + ), ) @@ -29,7 +32,10 @@ def email_alert(_): @run_failure_sensor def my_email_failure_sensor(context: RunFailureSensorContext): - message = f'Job "{context.dagster_run.job_name}" failed. Error: {context.failure_event.message}' + message = ( + f'Job "{context.dagster_run.job_name}" failed. Error:' + f" {context.failure_event.message}" + ) email_alert(message) diff --git a/examples/docs_snippets/docs_snippets/guides/dagster/development_to_production/branch_deployments/repository_v3.py b/examples/docs_snippets/docs_snippets/guides/dagster/development_to_production/branch_deployments/repository_v3.py index 5d321e78a5ef7..a968390eecf29 100644 --- a/examples/docs_snippets/docs_snippets/guides/dagster/development_to_production/branch_deployments/repository_v3.py +++ b/examples/docs_snippets/docs_snippets/guides/dagster/development_to_production/branch_deployments/repository_v3.py @@ -3,7 +3,14 @@ from dagster import graph, repository, with_resources from .clone_and_drop_db import drop_database_clone -from .repository_v2 import clone_prod, comments, get_current_env, items, resource_defs, stories +from .repository_v2 import ( + clone_prod, + comments, + get_current_env, + items, + resource_defs, + stories, +) # start_drop_db diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_asset_dependency.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_asset_dependency.py index 29c47e96dc952..3ea1c41e35593 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_asset_dependency.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_asset_dependency.py @@ -1,4 +1,7 @@ -from docs_snippets.concepts.assets.asset_dependency import downstream_asset, upstream_asset +from docs_snippets.concepts.assets.asset_dependency import ( + downstream_asset, + upstream_asset, +) def test_asset_dependency(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_cross_repository_asset.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_cross_repository_asset.py index 537b61e3dc68d..01976a1395e47 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_cross_repository_asset.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_cross_repository_asset.py @@ -1,4 +1,7 @@ -from docs_snippets.concepts.assets.cross_repository_asset import repository_a, repository_b +from docs_snippets.concepts.assets.cross_repository_asset import ( + repository_a, + repository_b, +) def test_repository_asset_groups(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/configuration_tests/test_run_config.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/configuration_tests/test_run_config.py index 4ba636766b588..f36941566d701 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/configuration_tests/test_run_config.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/configuration_tests/test_run_config.py @@ -1,7 +1,9 @@ import tempfile from docs_snippets.concepts.configuration.make_values_resource_any import file_dir_job -from docs_snippets.concepts.configuration.make_values_resource_config_schema import file_dirs_job +from docs_snippets.concepts.configuration.make_values_resource_config_schema import ( + file_dirs_job, +) def test_make_values_resource_any(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_config_input_manager.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_config_input_manager.py index 0d29bee5a9013..c5afac9878acb 100755 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_config_input_manager.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_config_input_manager.py @@ -1,4 +1,6 @@ -from docs_snippets.concepts.io_management.config_input_manager import execute_with_config +from docs_snippets.concepts.io_management.config_input_manager import ( + execute_with_config, +) def test_execute_job(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_load_custom_type_from_config.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_load_custom_type_from_config.py index 103f63148d21d..3e9c2cb8b724d 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_load_custom_type_from_config.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_load_custom_type_from_config.py @@ -1,4 +1,6 @@ -from docs_snippets.concepts.io_management.load_custom_type_from_config import execute_with_config +from docs_snippets.concepts.io_management.load_custom_type_from_config import ( + execute_with_config, +) def test_execute_job(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_output_config.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_output_config.py index 7b28824cf8f83..68d4bd4b61505 100755 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_output_config.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_output_config.py @@ -1,4 +1,6 @@ -from docs_snippets.concepts.io_management.output_config import execute_my_job_with_config +from docs_snippets.concepts.io_management.output_config import ( + execute_my_job_with_config, +) def test_execute_job(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_subselection.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_subselection.py index 59e86603d2bca..dd74be1482e92 100755 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_subselection.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_subselection.py @@ -1,4 +1,7 @@ -from docs_snippets.concepts.io_management.subselection import execute_full, execute_subselection +from docs_snippets.concepts.io_management.subselection import ( + execute_full, + execute_subselection, +) def test_execute_job(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/graph_tests/test_graphs.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/graph_tests/test_graphs.py index 380165877db66..e5a4c88580c23 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/graph_tests/test_graphs.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/graph_tests/test_graphs.py @@ -2,8 +2,12 @@ from docs_snippets.concepts.ops_jobs_graphs.graphs.fan_in_graph import fan_in from docs_snippets.concepts.ops_jobs_graphs.graphs.graphs import alias, one_plus_one from docs_snippets.concepts.ops_jobs_graphs.graphs.linear_graph import linear -from docs_snippets.concepts.ops_jobs_graphs.graphs.multiple_io_graph import inputs_and_outputs -from docs_snippets.concepts.ops_jobs_graphs.graphs.order_based_dependency import nothing_dependency +from docs_snippets.concepts.ops_jobs_graphs.graphs.multiple_io_graph import ( + inputs_and_outputs, +) +from docs_snippets.concepts.ops_jobs_graphs.graphs.order_based_dependency import ( + nothing_dependency, +) def test_one_plus_one(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_jobs.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_jobs.py index 5b8775293cc57..25f31491c16d2 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_jobs.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_jobs.py @@ -21,8 +21,13 @@ from docs_snippets.concepts.ops_jobs_graphs.jobs_from_graphs import local_job, prod_job from docs_snippets.concepts.ops_jobs_graphs.linear_job import linear from docs_snippets.concepts.ops_jobs_graphs.multiple_io_job import inputs_and_outputs -from docs_snippets.concepts.ops_jobs_graphs.order_based_dependency_job import nothing_dependency -from docs_snippets.concepts.ops_jobs_graphs.retries import default_and_override_job, retry_job +from docs_snippets.concepts.ops_jobs_graphs.order_based_dependency_job import ( + nothing_dependency, +) +from docs_snippets.concepts.ops_jobs_graphs.retries import ( + default_and_override_job, + retry_job, +) def test_one_plus_one(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_nested_graphs.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_nested_graphs.py index cd77d534330a7..a421a0741bc3a 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_nested_graphs.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_nested_graphs.py @@ -10,7 +10,10 @@ subgraph_config_job, subgraph_multiple_outputs_job, ) -from docs_snippets.concepts.ops_jobs_graphs.unnested_ops import all_together_unnested, return_fifty +from docs_snippets.concepts.ops_jobs_graphs.unnested_ops import ( + all_together_unnested, + return_fifty, +) from dagster import job from dagster._utils import file_relative_path diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_asset_job.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_asset_job.py index 38eae7c444cca..aea28f654adbb 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_asset_job.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_asset_job.py @@ -1,4 +1,6 @@ -from docs_snippets.concepts.partitions_schedules_sensors.partitioned_asset_job import repo +from docs_snippets.concepts.partitions_schedules_sensors.partitioned_asset_job import ( + repo, +) def test(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_job.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_job.py index 1f2ea2d404e54..64a154a7c99c9 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_job.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_job.py @@ -1,4 +1,6 @@ -from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import do_stuff_partitioned +from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import ( + do_stuff_partitioned, +) def test_do_stuff(): diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_schedule_from_partitions.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_schedule_from_partitions.py index a125dc5f35b96..d5c94ddc13b12 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_schedule_from_partitions.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_schedule_from_partitions.py @@ -1,4 +1,6 @@ -from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import do_stuff_partitioned +from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import ( + do_stuff_partitioned, +) from docs_snippets.concepts.partitions_schedules_sensors.schedule_from_partitions import ( do_stuff_partitioned_schedule, ) diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_static_partitioned_job.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_static_partitioned_job.py index 1f8ecb1484a22..f4aa60693cace 100644 --- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_static_partitioned_job.py +++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_static_partitioned_job.py @@ -1,4 +1,6 @@ -from docs_snippets.concepts.partitions_schedules_sensors.static_partitioned_job import continent_job +from docs_snippets.concepts.partitions_schedules_sensors.static_partitioned_job import ( + continent_job, +) def test_continent_job(): diff --git a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_graph.py b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_graph.py index 9930fc53c6da0..ca7b409f1cb63 100644 --- a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_graph.py +++ b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_graph.py @@ -1,4 +1,6 @@ -from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_graph import repo +from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_graph import ( + repo, +) def test_sda_graph(): diff --git a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_io_manager.py b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_io_manager.py index 4b42bb8e79dcd..3260032e1f030 100644 --- a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_io_manager.py +++ b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_io_manager.py @@ -1,4 +1,6 @@ -from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_io_manager import repo +from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_io_manager import ( + repo, +) def test_sda_nothing(): diff --git a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_nothing.py b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_nothing.py index 5e64a92a1cd5b..12956d8fc44d0 100644 --- a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_nothing.py +++ b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_nothing.py @@ -1,4 +1,6 @@ -from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_nothing import repo +from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_nothing import ( + repo, +) def test_sda_nothing(): diff --git a/examples/docs_snippets/docs_snippets_tests/legacy_tests/dagster_pandas_guide_tests/test_pipelines.py b/examples/docs_snippets/docs_snippets_tests/legacy_tests/dagster_pandas_guide_tests/test_pipelines.py index d0090ea4499b8..bcc7a30c2d9bc 100644 --- a/examples/docs_snippets/docs_snippets_tests/legacy_tests/dagster_pandas_guide_tests/test_pipelines.py +++ b/examples/docs_snippets/docs_snippets_tests/legacy_tests/dagster_pandas_guide_tests/test_pipelines.py @@ -3,7 +3,9 @@ from docs_snippets.legacy.dagster_pandas_guide.custom_column_constraint import ( custom_column_constraint_trip, ) -from docs_snippets.legacy.dagster_pandas_guide.shape_constrained_trip import shape_constrained_trip +from docs_snippets.legacy.dagster_pandas_guide.shape_constrained_trip import ( + shape_constrained_trip, +) from docs_snippets.legacy.dagster_pandas_guide.summary_stats import summary_stats_trip diff --git a/python_modules/dagster-graphql/dagster_graphql/schema/asset_graph.py b/python_modules/dagster-graphql/dagster_graphql/schema/asset_graph.py index 8c9b783fa7508..0c982921ad504 100644 --- a/python_modules/dagster-graphql/dagster_graphql/schema/asset_graph.py +++ b/python_modules/dagster-graphql/dagster_graphql/schema/asset_graph.py @@ -281,8 +281,13 @@ def get_partition_keys(self) -> Sequence[str]: # TODO: Add functionality for dynamic partitions definition partitions_def_data = self._external_asset_node.partitions_def_data if partitions_def_data: - if ( - isinstance(partitions_def_data, (ExternalStaticPartitionsDefinitionData, ExternalTimeWindowPartitionsDefinitionData, ExternalMultiPartitionsDefinitionData)) + if isinstance( + partitions_def_data, + ( + ExternalStaticPartitionsDefinitionData, + ExternalTimeWindowPartitionsDefinitionData, + ExternalMultiPartitionsDefinitionData, + ), ): return [ partition.name diff --git a/python_modules/dagster/dagster/__init__.py b/python_modules/dagster/dagster/__init__.py index 07232fff0d918..afa8b1abb6624 100644 --- a/python_modules/dagster/dagster/__init__.py +++ b/python_modules/dagster/dagster/__init__.py @@ -528,7 +528,6 @@ ) - _DEPRECATED: Final[Mapping[str, TypingTuple[str, str, str]]] = { "dagster_type_materializer": ( "dagster._core.types.config_schema", diff --git a/python_modules/dagster/dagster/_core/definitions/decorators/repository_decorator.py b/python_modules/dagster/dagster/_core/definitions/decorators/repository_decorator.py index a8e0b3f4f9e44..e09e5aabd6483 100644 --- a/python_modules/dagster/dagster/_core/definitions/decorators/repository_decorator.py +++ b/python_modules/dagster/dagster/_core/definitions/decorators/repository_decorator.py @@ -69,7 +69,20 @@ def __call__( for i, definition in enumerate(_flatten(repository_definitions)): if isinstance(definition, CacheableAssetsDefinition): defer_repository_data = True - elif not isinstance(definition, (PipelineDefinition, PartitionSetDefinition, ScheduleDefinition, SensorDefinition, GraphDefinition, AssetGroup, AssetsDefinition, SourceAsset, UnresolvedAssetJobDefinition)): + elif not isinstance( + definition, + ( + PipelineDefinition, + PartitionSetDefinition, + ScheduleDefinition, + SensorDefinition, + GraphDefinition, + AssetGroup, + AssetsDefinition, + SourceAsset, + UnresolvedAssetJobDefinition, + ), + ): bad_defns.append((i, type(definition))) else: repository_defns.append(definition) diff --git a/python_modules/dagster/dagster/_core/host_representation/external_data.py b/python_modules/dagster/dagster/_core/host_representation/external_data.py index 0886f9a9e0d9b..d1c0f2da5bad4 100644 --- a/python_modules/dagster/dagster/_core/host_representation/external_data.py +++ b/python_modules/dagster/dagster/_core/host_representation/external_data.py @@ -1235,7 +1235,10 @@ def external_multi_partitions_definition_from_def( if any( [ - not isinstance(dimension.partitions_def, (TimeWindowPartitionsDefinition, StaticPartitionsDefinition)) + not isinstance( + dimension.partitions_def, + (TimeWindowPartitionsDefinition, StaticPartitionsDefinition), + ) for dimension in partitions_def.partitions_defs ] ): diff --git a/python_modules/dagster/dagster/_core/storage/event_log/sql_event_log.py b/python_modules/dagster/dagster/_core/storage/event_log/sql_event_log.py index 05ebbf5a992a5..10a56ff40a1dd 100644 --- a/python_modules/dagster/dagster/_core/storage/event_log/sql_event_log.py +++ b/python_modules/dagster/dagster/_core/storage/event_log/sql_event_log.py @@ -370,10 +370,7 @@ def get_records_for_run( check.str_param(run_id, "run_id") check.opt_str_param(cursor, "cursor") - check.invariant( - not of_type - or isinstance(of_type, (DagsterEventType, frozenset, set)) - ) + check.invariant(not of_type or isinstance(of_type, (DagsterEventType, frozenset, set))) dagster_event_types = ( {of_type} diff --git a/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_schedule_commands.py b/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_schedule_commands.py index 9971b48ca8e7f..67a39fb857ce1 100644 --- a/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_schedule_commands.py +++ b/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_schedule_commands.py @@ -61,7 +61,7 @@ def test_schedules_start_and_stop(gen_schedule_args): ) assert result.exit_code == 0 - assert result.output == 'Started schedule foo_schedule\n' + assert result.output == "Started schedule foo_schedule\n" result = runner.invoke( schedule_stop_command, @@ -69,7 +69,7 @@ def test_schedules_start_and_stop(gen_schedule_args): ) assert result.exit_code == 0 - assert result.output == 'Stopped schedule foo_schedule\n' + assert result.output == "Stopped schedule foo_schedule\n" @pytest.mark.parametrize("gen_schedule_args", schedule_command_contexts()) diff --git a/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_sensor_commands.py b/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_sensor_commands.py index eaac60dd4a327..3b6f414156700 100644 --- a/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_sensor_commands.py +++ b/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_sensor_commands.py @@ -46,7 +46,7 @@ def test_sensors_start_and_stop(gen_sensor_args): ) assert result.exit_code == 0 - assert result.output == 'Started sensor foo_sensor\n' + assert result.output == "Started sensor foo_sensor\n" result = runner.invoke( sensor_stop_command, @@ -54,7 +54,7 @@ def test_sensors_start_and_stop(gen_sensor_args): ) assert result.exit_code == 0 - assert result.output == 'Stopped sensor foo_sensor\n' + assert result.output == "Stopped sensor foo_sensor\n" @pytest.mark.parametrize("gen_sensor_args", sensor_command_contexts()) diff --git a/python_modules/dagster/dagster_tests/core_tests/config_types_tests/evaluator_tests/test_config_mappings.py b/python_modules/dagster/dagster_tests/core_tests/config_types_tests/evaluator_tests/test_config_mappings.py index ec1bf4c6e0dc6..f200e70e0a2bf 100644 --- a/python_modules/dagster/dagster_tests/core_tests/config_types_tests/evaluator_tests/test_config_mappings.py +++ b/python_modules/dagster/dagster_tests/core_tests/config_types_tests/evaluator_tests/test_config_mappings.py @@ -1,4 +1,3 @@ - import pytest from dagster import ( diff --git a/python_modules/dagster/dagster_tests/core_tests/definitions_tests/decorators_tests/test_solid.py b/python_modules/dagster/dagster_tests/core_tests/definitions_tests/decorators_tests/test_solid.py index 165447d27b489..6fc01a9600401 100644 --- a/python_modules/dagster/dagster_tests/core_tests/definitions_tests/decorators_tests/test_solid.py +++ b/python_modules/dagster/dagster_tests/core_tests/definitions_tests/decorators_tests/test_solid.py @@ -241,9 +241,7 @@ def hello_world(context): assert context.op_config == conf_value called["yup"] = True - execute_in_graph( - hello_world, run_config={"solids": {"hello_world": {"config": conf_value}}} - ) + execute_in_graph(hello_world, run_config={"solids": {"hello_world": {"config": conf_value}}}) assert called["yup"] diff --git a/python_modules/dagster/dagster_tests/core_tests/launcher_tests/test_default_run_launcher.py b/python_modules/dagster/dagster_tests/core_tests/launcher_tests/test_default_run_launcher.py index e5a755445e6eb..29a6c224ca1f4 100644 --- a/python_modules/dagster/dagster_tests/core_tests/launcher_tests/test_default_run_launcher.py +++ b/python_modules/dagster/dagster_tests/core_tests/launcher_tests/test_default_run_launcher.py @@ -33,6 +33,7 @@ default_mode_def = ModeDefinition(resource_defs={"io_manager": fs_io_manager}) + @solid def noop_solid(_): pass diff --git a/python_modules/dagster/dagster_tests/core_tests/storage_tests/test_fs_io_manager.py b/python_modules/dagster/dagster_tests/core_tests/storage_tests/test_fs_io_manager.py index 9d9ecc01ba3f3..c542210e8d74f 100644 --- a/python_modules/dagster/dagster_tests/core_tests/storage_tests/test_fs_io_manager.py +++ b/python_modules/dagster/dagster_tests/core_tests/storage_tests/test_fs_io_manager.py @@ -70,7 +70,7 @@ def test_fs_io_manager(): assert input_metadata_entry_a.label == "path" assert input_metadata_entry_a.value == MetadataValue.path(filepath_a) assert len(loaded_input_events) == 1 - assert loaded_input_events[0].event_specific_data.upstream_step_key == 'op_a' + assert loaded_input_events[0].event_specific_data.upstream_step_key == "op_a" filepath_b = os.path.join(tmpdir_path, result.run_id, "op_b", "result") result_metadata_entry_b = handled_output_events[1].event_specific_data.metadata_entries[0] diff --git a/python_modules/dagster/dagster_tests/core_tests/test_output_materialization.py b/python_modules/dagster/dagster_tests/core_tests/test_output_materialization.py index f81346cf83b26..fe4dd4e62dee1 100644 --- a/python_modules/dagster/dagster_tests/core_tests/test_output_materialization.py +++ b/python_modules/dagster/dagster_tests/core_tests/test_output_materialization.py @@ -361,7 +361,16 @@ def return_one(): assert result.success event_types = [event.event_type_value for event in result.event_list] - assert sum([True for event_type in event_types if event_type == DagsterEventType.ASSET_MATERIALIZATION.value]) == 2 + assert ( + sum( + [ + True + for event_type in event_types + if event_type == DagsterEventType.ASSET_MATERIALIZATION.value + ] + ) + == 2 + ) @dagster_type_materializer(Int) diff --git a/python_modules/libraries/dagstermill/dagstermill/asset_factory.py b/python_modules/libraries/dagstermill/dagstermill/asset_factory.py index b76a967797e7e..a59e749f71f80 100644 --- a/python_modules/libraries/dagstermill/dagstermill/asset_factory.py +++ b/python_modules/libraries/dagstermill/dagstermill/asset_factory.py @@ -91,7 +91,8 @@ def _t_fn(context, **inputs): ) except Exception as ex: step_execution_context.log.warn( - f"Error when attempting to materialize executed notebook: {serializable_error_info_from_exc_info(sys.exc_info())}" + "Error when attempting to materialize executed notebook:" + f" {serializable_error_info_from_exc_info(sys.exc_info())}" ) # pylint: disable=no-member # compat: