diff --git a/docs/content/concepts/assets/asset-materializations.mdx b/docs/content/concepts/assets/asset-materializations.mdx
index 1b6b72f3a3dc4..4aec0bee57425 100644
--- a/docs/content/concepts/assets/asset-materializations.mdx
+++ b/docs/content/concepts/assets/asset-materializations.mdx
@@ -114,7 +114,7 @@ There are a variety of types of metadata that can be associated with a materiali
#### Example: Op body
```python file=concepts/assets/materialization_ops.py startafter=start_materialization_ops_marker_2 endbefore=end_materialization_ops_marker_2
-from dagster import op, AssetMaterialization, MetadataValue
+from dagster import AssetMaterialization, MetadataValue, op
@op
@@ -173,7 +173,7 @@ Check our API docs for for
If you are materializing a single slice of an asset (e.g. a single day's worth of data on a larger table), rather than mutating or creating it entirely, you can indicate this to Dagster by including the `partition` argument on the object.
```python file=/concepts/assets/materialization_ops.py startafter=start_partitioned_asset_materialization endbefore=end_partitioned_asset_materialization
-from dagster import op, AssetMaterialization
+from dagster import AssetMaterialization, op
@op(config_schema={"date": str})
diff --git a/docs/content/concepts/assets/asset-observations.mdx b/docs/content/concepts/assets/asset-observations.mdx
index 9fef494fb51da..e9ed5cd4ccb44 100644
--- a/docs/content/concepts/assets/asset-observations.mdx
+++ b/docs/content/concepts/assets/asset-observations.mdx
@@ -55,7 +55,7 @@ height={917}
There are a variety of types of metadata that can be associated with an observation event, all through the class. Each observation event optionally takes a dictionary of metadata entries that are then displayed in the event log and the [Asset Details](/concepts/dagit/dagit#asset-details) page. Check our API docs for for more details on the types of event metadata available.
```python file=concepts/assets/observations.py startafter=start_observation_asset_marker_2 endbefore=end_observation_asset_marker_2
-from dagster import op, AssetObservation, MetadataValue
+from dagster import AssetObservation, MetadataValue, op
@op
@@ -93,7 +93,7 @@ height={1146}
If you are observing a single slice of an asset (e.g. a single day's worth of data on a larger table), rather than mutating or creating it entirely, you can indicate this to Dagster by including the `partition` argument on the object.
```python file=/concepts/assets/observations.py startafter=start_partitioned_asset_observation endbefore=end_partitioned_asset_observation
-from dagster import op, AssetMaterialization
+from dagster import AssetMaterialization, op
@op(config_schema={"date": str})
diff --git a/docs/content/concepts/assets/software-defined-assets.mdx b/docs/content/concepts/assets/software-defined-assets.mdx
index ae78ba2bab8d1..6050d828e22cd 100644
--- a/docs/content/concepts/assets/software-defined-assets.mdx
+++ b/docs/content/concepts/assets/software-defined-assets.mdx
@@ -427,7 +427,7 @@ def nabisco_cereals():
**This recommended approach** constructs a group of assets from a specified module in your project. Using the `load_assets_from_package_module` function, you can import all assets in a module and apply a grouping:
```python file=/concepts/assets/asset_group_module.py startafter=start_example endbefore=end_example
-import my_package.cereal as cereal
+from my_package import cereal
cereal_assets = load_assets_from_package_module(
cereal,
diff --git a/docs/content/concepts/dagit/graphql-client.mdx b/docs/content/concepts/dagit/graphql-client.mdx
index a450c3a201ba3..bcceb1d0dfb56 100644
--- a/docs/content/concepts/dagit/graphql-client.mdx
+++ b/docs/content/concepts/dagit/graphql-client.mdx
@@ -63,6 +63,7 @@ You can use the client to get the status of a job run as follows:
```python file=/concepts/dagit/graphql/client_example.py startafter=start_run_status_marker endbefore=end_run_status_marker
from dagster_graphql import DagsterGraphQLClientError
+
from dagster import DagsterRunStatus
try:
@@ -83,10 +84,7 @@ You can also reload a repository location in a Dagster deployment.
This reloads all repositories in that repository location. This is useful in a variety of contexts, including refreshing Dagit without restarting the server. Example usage is as follows:
```python file=/concepts/dagit/graphql/client_example.py startafter=start_reload_repo_location_marker endbefore=end_reload_repo_location_marker
-from dagster_graphql import (
- ReloadRepositoryLocationInfo,
- ReloadRepositoryLocationStatus,
-)
+from dagster_graphql import ReloadRepositoryLocationInfo, ReloadRepositoryLocationStatus
reload_info: ReloadRepositoryLocationInfo = client.reload_repository_location(REPO_NAME)
if reload_info.status == ReloadRepositoryLocationStatus.SUCCESS:
diff --git a/docs/content/concepts/logging/loggers.mdx b/docs/content/concepts/logging/loggers.mdx
index 653a66994d9b8..6100bd9ac0b57 100644
--- a/docs/content/concepts/logging/loggers.mdx
+++ b/docs/content/concepts/logging/loggers.mdx
@@ -200,7 +200,7 @@ Default loggers can be specified on a
Note that if you explicitly specify loggers on a job, they will override those provided to `default_logger_defs`.
```python file=/concepts/logging/custom_logger.py startafter=start_default_logger_repo endbefore=end_default_logger_repo
-from dagster import repository, define_asset_job, asset
+from dagster import asset, define_asset_job, repository
@asset
diff --git a/docs/content/concepts/ops-jobs-graphs/op-events.mdx b/docs/content/concepts/ops-jobs-graphs/op-events.mdx
index bfd5e798616e3..b611694b2e5c3 100644
--- a/docs/content/concepts/ops-jobs-graphs/op-events.mdx
+++ b/docs/content/concepts/ops-jobs-graphs/op-events.mdx
@@ -159,7 +159,7 @@ To learn more about assets and how they are surfaced once you send this event, c
Attaching metadata to Asset Materializations is an important way of tracking aspects of a given asset over time. This functions essentially identically to other events which accept a `metadata` parameter, allowing you to attach a set of structured labels and values to display.
```python file=concepts/assets/materialization_ops.py startafter=start_materialization_ops_marker_2 endbefore=end_materialization_ops_marker_2
-from dagster import op, AssetMaterialization, MetadataValue
+from dagster import AssetMaterialization, MetadataValue, op
@op
diff --git a/docs/content/concepts/ops-jobs-graphs/op-hooks.mdx b/docs/content/concepts/ops-jobs-graphs/op-hooks.mdx
index 4074ae09caa33..4e624c6956734 100644
--- a/docs/content/concepts/ops-jobs-graphs/op-hooks.mdx
+++ b/docs/content/concepts/ops-jobs-graphs/op-hooks.mdx
@@ -25,7 +25,17 @@ A , which is use
You can use to test the type check function of a custom Dagster Type:
```python file=/concepts/types/types.py startafter=start_test_dagster_type endbefore=end_test_dagster_type
-from dagster import check_dagster_type, Dict, Any
+from dagster import Any, Dict, check_dagster_type
def test_dagster_type():
diff --git a/docs/content/deployment/executors.mdx b/docs/content/deployment/executors.mdx
index e56986124d730..d9663d660d98b 100644
--- a/docs/content/deployment/executors.mdx
+++ b/docs/content/deployment/executors.mdx
@@ -23,7 +23,7 @@ Every job has an executor. The default executor is the to the `executor_def` parameter of or .
```python file=/deploying/executors/executors.py startafter=start_executor_on_job endbefore=end_executor_on_job
-from dagster import multiprocess_executor, job, graph
+from dagster import graph, job, multiprocess_executor
# Providing an executor using the job decorator
@job(executor_def=multiprocess_executor)
@@ -43,7 +43,7 @@ other_job = the_graph.to_job(executor_def=multiprocess_executor)
A default executor can be specified for all jobs and assets provided to a repository using the `default_executor_def` argument of . All jobs that don't specify an executor will use this default executor, but if a job explicitly specifies an executor, then the default provided to the repository will not be used.
```python file=/deploying/executors/executors.py startafter=start_executor_on_repo endbefore=end_executor_on_repo
-from dagster import multiprocess_executor, define_asset_job, asset, repository
+from dagster import asset, define_asset_job, multiprocess_executor, repository
@asset
diff --git a/docs/content/guides/dagster/branch_deployments.mdx b/docs/content/guides/dagster/branch_deployments.mdx
index 78711da258d98..770ce5d7e7487 100644
--- a/docs/content/guides/dagster/branch_deployments.mdx
+++ b/docs/content/guides/dagster/branch_deployments.mdx
@@ -167,7 +167,7 @@ def repo():
def get_current_env():
is_branch_depl = os.getenv("DAGSTER_CLOUD_IS_BRANCH_DEPLOYMENT") == "1"
- assert is_branch_depl != None # env var must be set
+ assert is_branch_depl is not None # env var must be set
return "branch" if is_branch_depl else "prod"
return [
diff --git a/docs/content/guides/dagster/enriching-with-software-defined-assets.mdx b/docs/content/guides/dagster/enriching-with-software-defined-assets.mdx
index 831f91de224fa..128c3ecbcbfa2 100644
--- a/docs/content/guides/dagster/enriching-with-software-defined-assets.mdx
+++ b/docs/content/guides/dagster/enriching-with-software-defined-assets.mdx
@@ -205,14 +205,14 @@ from .mylib import create_db_connection, pickle_to_s3, train_recommender_model
@op
def build_users():
- raw_users_df = read_sql(f"select * from raw_users", con=create_db_connection())
+ raw_users_df = read_sql("select * from raw_users", con=create_db_connection())
users_df = raw_users_df.dropna()
users_df.to_sql(name="users", con=create_db_connection())
@op(ins={"users": In(Nothing)})
def build_user_recommender_model():
- users_df = read_sql(f"select * from users", con=create_db_connection())
+ users_df = read_sql("select * from users", con=create_db_connection())
users_recommender_model = train_recommender_model(users_df)
pickle_to_s3(users_recommender_model, key="users_recommender_model")
@@ -241,14 +241,14 @@ from .mylib import create_db_connection, pickle_to_s3, train_recommender_model
@asset(non_argument_deps={"raw_users"})
def users():
- raw_users_df = read_sql(f"select * from raw_users", con=create_db_connection())
+ raw_users_df = read_sql("select * from raw_users", con=create_db_connection())
users_df = raw_users_df.dropna()
users_df.to_sql(name="users", con=create_db_connection())
@asset(non_argument_deps={"users"})
def user_recommender_model():
- users_df = read_sql(f"select * from users", con=create_db_connection())
+ users_df = read_sql("select * from users", con=create_db_connection())
users_recommender_model = train_recommender_model(users_df)
pickle_to_s3(users_recommender_model, key="users_recommender_model")
diff --git a/docs/content/guides/dagster/re-execution.mdx b/docs/content/guides/dagster/re-execution.mdx
index 48e6685f52587..bafe2a22d7363 100644
--- a/docs/content/guides/dagster/re-execution.mdx
+++ b/docs/content/guides/dagster/re-execution.mdx
@@ -80,9 +80,10 @@ Re-execution can be triggered via the API as well.
Again, let's revist the job `unreliable_job`, which has a op named `unreliable`.
```python file=/guides/dagster/reexecution/reexecution_api.py startafter=start_initial_execution_marker endbefore=end_initial_execution_marker
-from dagster import DagsterInstance, ReexecutionOptions, execute_job, reconstructable
from docs_snippets.guides.dagster.reexecution.unreliable_job import unreliable_job
+from dagster import DagsterInstance, ReexecutionOptions, execute_job, reconstructable
+
instance = DagsterInstance.ephemeral()
# Initial execution
diff --git a/docs/content/guides/dagster/software-defined-assets.mdx b/docs/content/guides/dagster/software-defined-assets.mdx
index 64dc6c99a47d8..db337091a54eb 100644
--- a/docs/content/guides/dagster/software-defined-assets.mdx
+++ b/docs/content/guides/dagster/software-defined-assets.mdx
@@ -70,9 +70,10 @@ It's common to use a utility like
```python file=../../assets_pandas_pyspark/assets_pandas_pyspark/assets/weather_assets.py startafter=gather_assets_start endbefore=gather_assets_end
# imports the module called "assets" from the package containing the current module
# the "assets" module contains the asset definitions
-from . import table_assets
from dagster import load_assets_from_modules, with_resources
+from . import table_assets
+
weather_assets = with_resources(
load_assets_from_modules(modules=[table_assets]),
resource_defs={
@@ -109,9 +110,11 @@ class LocalFileSystemIOManager(IOManager):
Not all the assets in the same dependency graph need to have the same Python type. Here's an asset whose computation is defined using Spark DataFrames, that depends on the `daily_temperature_highs` asset we defined above using Pandas.
```python file=../../assets_pandas_pyspark/assets_pandas_pyspark/assets/spark_asset.py
-from pyspark.sql import DataFrame as SparkDF
-from pyspark.sql import Window
-from pyspark.sql import functions as f
+from pyspark.sql import (
+ DataFrame as SparkDF,
+ Window,
+ functions as f,
+)
from dagster import asset
@@ -132,9 +135,10 @@ def daily_temperature_high_diffs(daily_temperature_highs: SparkDF) -> SparkDF:
Here's an extended version of `weather_assets` that contains the new asset:
```python file=../../assets_pandas_pyspark/assets_pandas_pyspark/assets/spark_weather_assets.py startafter=gather_assets_start endbefore=gather_assets_end
-from . import table_assets, spark_asset
from dagster import load_assets_from_modules, with_resources
+from . import spark_asset, table_assets
+
spark_weather_assets = with_resources(
load_assets_from_modules(modules=[table_assets, spark_asset]),
resource_defs={
diff --git a/docs/content/integrations/dbt-cloud.mdx b/docs/content/integrations/dbt-cloud.mdx
index cb4c321cf43a0..a0a3128206705 100644
--- a/docs/content/integrations/dbt-cloud.mdx
+++ b/docs/content/integrations/dbt-cloud.mdx
@@ -99,7 +99,7 @@ Now that your dbt Cloud assets are loaded, you can define a Dagster job that mat
You can explicitly define when your assets should be materialized. For example, you can schedule assets based on their upstream or downstream dependencies, external events using a sensor, or a cron schedule.
```python startafter=start_schedule_dbt_cloud_assets endbefore=end_schedule_dbt_cloud_assets file=/integrations/dbt/dbt_cloud.py dedent=4
-from dagster import ScheduleDefinition, define_asset_job, repository, AssetSelection
+from dagster import AssetSelection, ScheduleDefinition, define_asset_job, repository
# Materialize all assets in the repository
run_everything_job = define_asset_job("run_everything_job", AssetSelection.all())
diff --git a/examples/docs_snippets/docs_snippets/concepts/assets/asset_input_managers_numpy.py b/examples/docs_snippets/docs_snippets/concepts/assets/asset_input_managers_numpy.py
index 04ea170352bd9..de368058b0c31 100644
--- a/examples/docs_snippets/docs_snippets/concepts/assets/asset_input_managers_numpy.py
+++ b/examples/docs_snippets/docs_snippets/concepts/assets/asset_input_managers_numpy.py
@@ -4,7 +4,11 @@
from dagster import AssetIn, IOManager, asset, io_manager, repository, with_resources
-from .asset_input_managers import load_numpy_array, load_pandas_dataframe, store_pandas_dataframe
+from .asset_input_managers import (
+ load_numpy_array,
+ load_pandas_dataframe,
+ store_pandas_dataframe,
+)
# start_numpy_example
diff --git a/examples/docs_snippets/docs_snippets/concepts/assets/subset_graph_backed_asset.py b/examples/docs_snippets/docs_snippets/concepts/assets/subset_graph_backed_asset.py
index 87d2bb85b39d2..6845606b97408 100644
--- a/examples/docs_snippets/docs_snippets/concepts/assets/subset_graph_backed_asset.py
+++ b/examples/docs_snippets/docs_snippets/concepts/assets/subset_graph_backed_asset.py
@@ -1,4 +1,13 @@
-from dagster import AssetsDefinition, GraphOut, Out, Output, define_asset_job, graph, op, repository
+from dagster import (
+ AssetsDefinition,
+ GraphOut,
+ Out,
+ Output,
+ define_asset_job,
+ graph,
+ op,
+ repository,
+)
# start_graph_backed_asset_foo
diff --git a/examples/docs_snippets/docs_snippets/concepts/dagit/graphql/client_example.py b/examples/docs_snippets/docs_snippets/concepts/dagit/graphql/client_example.py
index 66a0ecc8e64fd..3e1b7ea8d3d19 100644
--- a/examples/docs_snippets/docs_snippets/concepts/dagit/graphql/client_example.py
+++ b/examples/docs_snippets/docs_snippets/concepts/dagit/graphql/client_example.py
@@ -89,7 +89,10 @@ def do_something_with_exc(some_exception): # pylint: disable=W0613
# end_reload_repo_location_marker
# start_shutdown_repo_location_marker
-from dagster_graphql import ShutdownRepositoryLocationInfo, ShutdownRepositoryLocationStatus
+from dagster_graphql import (
+ ShutdownRepositoryLocationInfo,
+ ShutdownRepositoryLocationStatus,
+)
shutdown_info: ShutdownRepositoryLocationInfo = client.shutdown_repository_location(
REPO_NAME
diff --git a/examples/docs_snippets/docs_snippets/concepts/ops_jobs_graphs/nested_graphs.py b/examples/docs_snippets/docs_snippets/concepts/ops_jobs_graphs/nested_graphs.py
index db1bd1f1a2d18..12d95e1da4e20 100644
--- a/examples/docs_snippets/docs_snippets/concepts/ops_jobs_graphs/nested_graphs.py
+++ b/examples/docs_snippets/docs_snippets/concepts/ops_jobs_graphs/nested_graphs.py
@@ -4,7 +4,12 @@
from dagster import graph, job, op
-from .unnested_ops import add_thirty_two, log_number, multiply_by_one_point_eight, return_fifty
+from .unnested_ops import (
+ add_thirty_two,
+ log_number,
+ multiply_by_one_point_eight,
+ return_fifty,
+)
# start_composite_solid_example_marker
diff --git a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_asset_job.py b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_asset_job.py
index 5996d11fe7d37..8d88e0778e381 100644
--- a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_asset_job.py
+++ b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_asset_job.py
@@ -1,4 +1,10 @@
-from dagster import AssetSelection, HourlyPartitionsDefinition, asset, define_asset_job, repository
+from dagster import (
+ AssetSelection,
+ HourlyPartitionsDefinition,
+ asset,
+ define_asset_job,
+ repository,
+)
hourly_partitions_def = HourlyPartitionsDefinition(start_date="2022-05-31-00:00")
diff --git a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_job_test.py b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_job_test.py
index 7ab6908632d9f..062e6ec90e798 100644
--- a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_job_test.py
+++ b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/partitioned_job_test.py
@@ -1,6 +1,8 @@
# isort: skip_file
-from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import do_stuff_partitioned
+from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import (
+ do_stuff_partitioned,
+)
# start
diff --git a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/run_status_run_requests.py b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/run_status_run_requests.py
index a5c128782c5c9..08c0de8713789 100644
--- a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/run_status_run_requests.py
+++ b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/run_status_run_requests.py
@@ -1,4 +1,10 @@
-from dagster import DagsterRunStatus, RunRequest, SkipReason, run_failure_sensor, run_status_sensor
+from dagster import (
+ DagsterRunStatus,
+ RunRequest,
+ SkipReason,
+ run_failure_sensor,
+ run_status_sensor,
+)
status_reporting_job = None
diff --git a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/sensor_alert.py b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/sensor_alert.py
index 223453f2e4fba..df8aace89b719 100644
--- a/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/sensor_alert.py
+++ b/examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/sensors/sensor_alert.py
@@ -13,7 +13,10 @@ def my_slack_on_run_failure(context: RunFailureSensorContext):
slack_client.chat_postMessage(
channel="#alert-channel",
- message=f'Job "{context.dagster_run.job_name}" failed. Error: {context.failure_event.message}',
+ message=(
+ f'Job "{context.dagster_run.job_name}" failed. Error:'
+ f" {context.failure_event.message}"
+ ),
)
@@ -29,7 +32,10 @@ def email_alert(_):
@run_failure_sensor
def my_email_failure_sensor(context: RunFailureSensorContext):
- message = f'Job "{context.dagster_run.job_name}" failed. Error: {context.failure_event.message}'
+ message = (
+ f'Job "{context.dagster_run.job_name}" failed. Error:'
+ f" {context.failure_event.message}"
+ )
email_alert(message)
diff --git a/examples/docs_snippets/docs_snippets/guides/dagster/development_to_production/branch_deployments/repository_v3.py b/examples/docs_snippets/docs_snippets/guides/dagster/development_to_production/branch_deployments/repository_v3.py
index 5d321e78a5ef7..a968390eecf29 100644
--- a/examples/docs_snippets/docs_snippets/guides/dagster/development_to_production/branch_deployments/repository_v3.py
+++ b/examples/docs_snippets/docs_snippets/guides/dagster/development_to_production/branch_deployments/repository_v3.py
@@ -3,7 +3,14 @@
from dagster import graph, repository, with_resources
from .clone_and_drop_db import drop_database_clone
-from .repository_v2 import clone_prod, comments, get_current_env, items, resource_defs, stories
+from .repository_v2 import (
+ clone_prod,
+ comments,
+ get_current_env,
+ items,
+ resource_defs,
+ stories,
+)
# start_drop_db
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_asset_dependency.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_asset_dependency.py
index 29c47e96dc952..3ea1c41e35593 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_asset_dependency.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_asset_dependency.py
@@ -1,4 +1,7 @@
-from docs_snippets.concepts.assets.asset_dependency import downstream_asset, upstream_asset
+from docs_snippets.concepts.assets.asset_dependency import (
+ downstream_asset,
+ upstream_asset,
+)
def test_asset_dependency():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_cross_repository_asset.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_cross_repository_asset.py
index 537b61e3dc68d..01976a1395e47 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_cross_repository_asset.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/assets_tests/test_cross_repository_asset.py
@@ -1,4 +1,7 @@
-from docs_snippets.concepts.assets.cross_repository_asset import repository_a, repository_b
+from docs_snippets.concepts.assets.cross_repository_asset import (
+ repository_a,
+ repository_b,
+)
def test_repository_asset_groups():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/configuration_tests/test_run_config.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/configuration_tests/test_run_config.py
index 4ba636766b588..f36941566d701 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/configuration_tests/test_run_config.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/configuration_tests/test_run_config.py
@@ -1,7 +1,9 @@
import tempfile
from docs_snippets.concepts.configuration.make_values_resource_any import file_dir_job
-from docs_snippets.concepts.configuration.make_values_resource_config_schema import file_dirs_job
+from docs_snippets.concepts.configuration.make_values_resource_config_schema import (
+ file_dirs_job,
+)
def test_make_values_resource_any():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_config_input_manager.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_config_input_manager.py
index 0d29bee5a9013..c5afac9878acb 100755
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_config_input_manager.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_config_input_manager.py
@@ -1,4 +1,6 @@
-from docs_snippets.concepts.io_management.config_input_manager import execute_with_config
+from docs_snippets.concepts.io_management.config_input_manager import (
+ execute_with_config,
+)
def test_execute_job():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_load_custom_type_from_config.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_load_custom_type_from_config.py
index 103f63148d21d..3e9c2cb8b724d 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_load_custom_type_from_config.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_load_custom_type_from_config.py
@@ -1,4 +1,6 @@
-from docs_snippets.concepts.io_management.load_custom_type_from_config import execute_with_config
+from docs_snippets.concepts.io_management.load_custom_type_from_config import (
+ execute_with_config,
+)
def test_execute_job():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_output_config.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_output_config.py
index 7b28824cf8f83..68d4bd4b61505 100755
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_output_config.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_output_config.py
@@ -1,4 +1,6 @@
-from docs_snippets.concepts.io_management.output_config import execute_my_job_with_config
+from docs_snippets.concepts.io_management.output_config import (
+ execute_my_job_with_config,
+)
def test_execute_job():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_subselection.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_subselection.py
index 59e86603d2bca..dd74be1482e92 100755
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_subselection.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/io_management_tests/test_subselection.py
@@ -1,4 +1,7 @@
-from docs_snippets.concepts.io_management.subselection import execute_full, execute_subselection
+from docs_snippets.concepts.io_management.subselection import (
+ execute_full,
+ execute_subselection,
+)
def test_execute_job():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/graph_tests/test_graphs.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/graph_tests/test_graphs.py
index 380165877db66..e5a4c88580c23 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/graph_tests/test_graphs.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/graph_tests/test_graphs.py
@@ -2,8 +2,12 @@
from docs_snippets.concepts.ops_jobs_graphs.graphs.fan_in_graph import fan_in
from docs_snippets.concepts.ops_jobs_graphs.graphs.graphs import alias, one_plus_one
from docs_snippets.concepts.ops_jobs_graphs.graphs.linear_graph import linear
-from docs_snippets.concepts.ops_jobs_graphs.graphs.multiple_io_graph import inputs_and_outputs
-from docs_snippets.concepts.ops_jobs_graphs.graphs.order_based_dependency import nothing_dependency
+from docs_snippets.concepts.ops_jobs_graphs.graphs.multiple_io_graph import (
+ inputs_and_outputs,
+)
+from docs_snippets.concepts.ops_jobs_graphs.graphs.order_based_dependency import (
+ nothing_dependency,
+)
def test_one_plus_one():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_jobs.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_jobs.py
index 5b8775293cc57..25f31491c16d2 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_jobs.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_jobs.py
@@ -21,8 +21,13 @@
from docs_snippets.concepts.ops_jobs_graphs.jobs_from_graphs import local_job, prod_job
from docs_snippets.concepts.ops_jobs_graphs.linear_job import linear
from docs_snippets.concepts.ops_jobs_graphs.multiple_io_job import inputs_and_outputs
-from docs_snippets.concepts.ops_jobs_graphs.order_based_dependency_job import nothing_dependency
-from docs_snippets.concepts.ops_jobs_graphs.retries import default_and_override_job, retry_job
+from docs_snippets.concepts.ops_jobs_graphs.order_based_dependency_job import (
+ nothing_dependency,
+)
+from docs_snippets.concepts.ops_jobs_graphs.retries import (
+ default_and_override_job,
+ retry_job,
+)
def test_one_plus_one():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_nested_graphs.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_nested_graphs.py
index cd77d534330a7..a421a0741bc3a 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_nested_graphs.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/ops_jobs_graphs_tests/test_nested_graphs.py
@@ -10,7 +10,10 @@
subgraph_config_job,
subgraph_multiple_outputs_job,
)
-from docs_snippets.concepts.ops_jobs_graphs.unnested_ops import all_together_unnested, return_fifty
+from docs_snippets.concepts.ops_jobs_graphs.unnested_ops import (
+ all_together_unnested,
+ return_fifty,
+)
from dagster import job
from dagster._utils import file_relative_path
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_asset_job.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_asset_job.py
index 38eae7c444cca..aea28f654adbb 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_asset_job.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_asset_job.py
@@ -1,4 +1,6 @@
-from docs_snippets.concepts.partitions_schedules_sensors.partitioned_asset_job import repo
+from docs_snippets.concepts.partitions_schedules_sensors.partitioned_asset_job import (
+ repo,
+)
def test():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_job.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_job.py
index 1f2ea2d404e54..64a154a7c99c9 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_job.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_partitioned_job.py
@@ -1,4 +1,6 @@
-from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import do_stuff_partitioned
+from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import (
+ do_stuff_partitioned,
+)
def test_do_stuff():
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_schedule_from_partitions.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_schedule_from_partitions.py
index a125dc5f35b96..d5c94ddc13b12 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_schedule_from_partitions.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_schedule_from_partitions.py
@@ -1,4 +1,6 @@
-from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import do_stuff_partitioned
+from docs_snippets.concepts.partitions_schedules_sensors.partitioned_job import (
+ do_stuff_partitioned,
+)
from docs_snippets.concepts.partitions_schedules_sensors.schedule_from_partitions import (
do_stuff_partitioned_schedule,
)
diff --git a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_static_partitioned_job.py b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_static_partitioned_job.py
index 1f8ecb1484a22..f4aa60693cace 100644
--- a/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_static_partitioned_job.py
+++ b/examples/docs_snippets/docs_snippets_tests/concepts_tests/partitions_schedules_sensors_tests/test_static_partitioned_job.py
@@ -1,4 +1,6 @@
-from docs_snippets.concepts.partitions_schedules_sensors.static_partitioned_job import continent_job
+from docs_snippets.concepts.partitions_schedules_sensors.static_partitioned_job import (
+ continent_job,
+)
def test_continent_job():
diff --git a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_graph.py b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_graph.py
index 9930fc53c6da0..ca7b409f1cb63 100644
--- a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_graph.py
+++ b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_graph.py
@@ -1,4 +1,6 @@
-from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_graph import repo
+from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_graph import (
+ repo,
+)
def test_sda_graph():
diff --git a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_io_manager.py b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_io_manager.py
index 4b42bb8e79dcd..3260032e1f030 100644
--- a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_io_manager.py
+++ b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_io_manager.py
@@ -1,4 +1,6 @@
-from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_io_manager import repo
+from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_io_manager import (
+ repo,
+)
def test_sda_nothing():
diff --git a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_nothing.py b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_nothing.py
index 5e64a92a1cd5b..12956d8fc44d0 100644
--- a/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_nothing.py
+++ b/examples/docs_snippets/docs_snippets_tests/guides_tests/enriching_with_software_defined_assets/test_sda_nothing.py
@@ -1,4 +1,6 @@
-from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_nothing import repo
+from docs_snippets.guides.dagster.enriching_with_software_defined_assets.sda_nothing import (
+ repo,
+)
def test_sda_nothing():
diff --git a/examples/docs_snippets/docs_snippets_tests/legacy_tests/dagster_pandas_guide_tests/test_pipelines.py b/examples/docs_snippets/docs_snippets_tests/legacy_tests/dagster_pandas_guide_tests/test_pipelines.py
index d0090ea4499b8..bcc7a30c2d9bc 100644
--- a/examples/docs_snippets/docs_snippets_tests/legacy_tests/dagster_pandas_guide_tests/test_pipelines.py
+++ b/examples/docs_snippets/docs_snippets_tests/legacy_tests/dagster_pandas_guide_tests/test_pipelines.py
@@ -3,7 +3,9 @@
from docs_snippets.legacy.dagster_pandas_guide.custom_column_constraint import (
custom_column_constraint_trip,
)
-from docs_snippets.legacy.dagster_pandas_guide.shape_constrained_trip import shape_constrained_trip
+from docs_snippets.legacy.dagster_pandas_guide.shape_constrained_trip import (
+ shape_constrained_trip,
+)
from docs_snippets.legacy.dagster_pandas_guide.summary_stats import summary_stats_trip
diff --git a/python_modules/dagster-graphql/dagster_graphql/schema/asset_graph.py b/python_modules/dagster-graphql/dagster_graphql/schema/asset_graph.py
index 8c9b783fa7508..0c982921ad504 100644
--- a/python_modules/dagster-graphql/dagster_graphql/schema/asset_graph.py
+++ b/python_modules/dagster-graphql/dagster_graphql/schema/asset_graph.py
@@ -281,8 +281,13 @@ def get_partition_keys(self) -> Sequence[str]:
# TODO: Add functionality for dynamic partitions definition
partitions_def_data = self._external_asset_node.partitions_def_data
if partitions_def_data:
- if (
- isinstance(partitions_def_data, (ExternalStaticPartitionsDefinitionData, ExternalTimeWindowPartitionsDefinitionData, ExternalMultiPartitionsDefinitionData))
+ if isinstance(
+ partitions_def_data,
+ (
+ ExternalStaticPartitionsDefinitionData,
+ ExternalTimeWindowPartitionsDefinitionData,
+ ExternalMultiPartitionsDefinitionData,
+ ),
):
return [
partition.name
diff --git a/python_modules/dagster/dagster/__init__.py b/python_modules/dagster/dagster/__init__.py
index 07232fff0d918..afa8b1abb6624 100644
--- a/python_modules/dagster/dagster/__init__.py
+++ b/python_modules/dagster/dagster/__init__.py
@@ -528,7 +528,6 @@
)
-
_DEPRECATED: Final[Mapping[str, TypingTuple[str, str, str]]] = {
"dagster_type_materializer": (
"dagster._core.types.config_schema",
diff --git a/python_modules/dagster/dagster/_core/definitions/decorators/repository_decorator.py b/python_modules/dagster/dagster/_core/definitions/decorators/repository_decorator.py
index a8e0b3f4f9e44..e09e5aabd6483 100644
--- a/python_modules/dagster/dagster/_core/definitions/decorators/repository_decorator.py
+++ b/python_modules/dagster/dagster/_core/definitions/decorators/repository_decorator.py
@@ -69,7 +69,20 @@ def __call__(
for i, definition in enumerate(_flatten(repository_definitions)):
if isinstance(definition, CacheableAssetsDefinition):
defer_repository_data = True
- elif not isinstance(definition, (PipelineDefinition, PartitionSetDefinition, ScheduleDefinition, SensorDefinition, GraphDefinition, AssetGroup, AssetsDefinition, SourceAsset, UnresolvedAssetJobDefinition)):
+ elif not isinstance(
+ definition,
+ (
+ PipelineDefinition,
+ PartitionSetDefinition,
+ ScheduleDefinition,
+ SensorDefinition,
+ GraphDefinition,
+ AssetGroup,
+ AssetsDefinition,
+ SourceAsset,
+ UnresolvedAssetJobDefinition,
+ ),
+ ):
bad_defns.append((i, type(definition)))
else:
repository_defns.append(definition)
diff --git a/python_modules/dagster/dagster/_core/host_representation/external_data.py b/python_modules/dagster/dagster/_core/host_representation/external_data.py
index 0886f9a9e0d9b..d1c0f2da5bad4 100644
--- a/python_modules/dagster/dagster/_core/host_representation/external_data.py
+++ b/python_modules/dagster/dagster/_core/host_representation/external_data.py
@@ -1235,7 +1235,10 @@ def external_multi_partitions_definition_from_def(
if any(
[
- not isinstance(dimension.partitions_def, (TimeWindowPartitionsDefinition, StaticPartitionsDefinition))
+ not isinstance(
+ dimension.partitions_def,
+ (TimeWindowPartitionsDefinition, StaticPartitionsDefinition),
+ )
for dimension in partitions_def.partitions_defs
]
):
diff --git a/python_modules/dagster/dagster/_core/storage/event_log/sql_event_log.py b/python_modules/dagster/dagster/_core/storage/event_log/sql_event_log.py
index 05ebbf5a992a5..10a56ff40a1dd 100644
--- a/python_modules/dagster/dagster/_core/storage/event_log/sql_event_log.py
+++ b/python_modules/dagster/dagster/_core/storage/event_log/sql_event_log.py
@@ -370,10 +370,7 @@ def get_records_for_run(
check.str_param(run_id, "run_id")
check.opt_str_param(cursor, "cursor")
- check.invariant(
- not of_type
- or isinstance(of_type, (DagsterEventType, frozenset, set))
- )
+ check.invariant(not of_type or isinstance(of_type, (DagsterEventType, frozenset, set)))
dagster_event_types = (
{of_type}
diff --git a/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_schedule_commands.py b/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_schedule_commands.py
index 9971b48ca8e7f..67a39fb857ce1 100644
--- a/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_schedule_commands.py
+++ b/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_schedule_commands.py
@@ -61,7 +61,7 @@ def test_schedules_start_and_stop(gen_schedule_args):
)
assert result.exit_code == 0
- assert result.output == 'Started schedule foo_schedule\n'
+ assert result.output == "Started schedule foo_schedule\n"
result = runner.invoke(
schedule_stop_command,
@@ -69,7 +69,7 @@ def test_schedules_start_and_stop(gen_schedule_args):
)
assert result.exit_code == 0
- assert result.output == 'Stopped schedule foo_schedule\n'
+ assert result.output == "Stopped schedule foo_schedule\n"
@pytest.mark.parametrize("gen_schedule_args", schedule_command_contexts())
diff --git a/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_sensor_commands.py b/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_sensor_commands.py
index eaac60dd4a327..3b6f414156700 100644
--- a/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_sensor_commands.py
+++ b/python_modules/dagster/dagster_tests/cli_tests/command_tests/test_sensor_commands.py
@@ -46,7 +46,7 @@ def test_sensors_start_and_stop(gen_sensor_args):
)
assert result.exit_code == 0
- assert result.output == 'Started sensor foo_sensor\n'
+ assert result.output == "Started sensor foo_sensor\n"
result = runner.invoke(
sensor_stop_command,
@@ -54,7 +54,7 @@ def test_sensors_start_and_stop(gen_sensor_args):
)
assert result.exit_code == 0
- assert result.output == 'Stopped sensor foo_sensor\n'
+ assert result.output == "Stopped sensor foo_sensor\n"
@pytest.mark.parametrize("gen_sensor_args", sensor_command_contexts())
diff --git a/python_modules/dagster/dagster_tests/core_tests/config_types_tests/evaluator_tests/test_config_mappings.py b/python_modules/dagster/dagster_tests/core_tests/config_types_tests/evaluator_tests/test_config_mappings.py
index ec1bf4c6e0dc6..f200e70e0a2bf 100644
--- a/python_modules/dagster/dagster_tests/core_tests/config_types_tests/evaluator_tests/test_config_mappings.py
+++ b/python_modules/dagster/dagster_tests/core_tests/config_types_tests/evaluator_tests/test_config_mappings.py
@@ -1,4 +1,3 @@
-
import pytest
from dagster import (
diff --git a/python_modules/dagster/dagster_tests/core_tests/definitions_tests/decorators_tests/test_solid.py b/python_modules/dagster/dagster_tests/core_tests/definitions_tests/decorators_tests/test_solid.py
index 165447d27b489..6fc01a9600401 100644
--- a/python_modules/dagster/dagster_tests/core_tests/definitions_tests/decorators_tests/test_solid.py
+++ b/python_modules/dagster/dagster_tests/core_tests/definitions_tests/decorators_tests/test_solid.py
@@ -241,9 +241,7 @@ def hello_world(context):
assert context.op_config == conf_value
called["yup"] = True
- execute_in_graph(
- hello_world, run_config={"solids": {"hello_world": {"config": conf_value}}}
- )
+ execute_in_graph(hello_world, run_config={"solids": {"hello_world": {"config": conf_value}}})
assert called["yup"]
diff --git a/python_modules/dagster/dagster_tests/core_tests/launcher_tests/test_default_run_launcher.py b/python_modules/dagster/dagster_tests/core_tests/launcher_tests/test_default_run_launcher.py
index e5a755445e6eb..29a6c224ca1f4 100644
--- a/python_modules/dagster/dagster_tests/core_tests/launcher_tests/test_default_run_launcher.py
+++ b/python_modules/dagster/dagster_tests/core_tests/launcher_tests/test_default_run_launcher.py
@@ -33,6 +33,7 @@
default_mode_def = ModeDefinition(resource_defs={"io_manager": fs_io_manager})
+
@solid
def noop_solid(_):
pass
diff --git a/python_modules/dagster/dagster_tests/core_tests/storage_tests/test_fs_io_manager.py b/python_modules/dagster/dagster_tests/core_tests/storage_tests/test_fs_io_manager.py
index 9d9ecc01ba3f3..c542210e8d74f 100644
--- a/python_modules/dagster/dagster_tests/core_tests/storage_tests/test_fs_io_manager.py
+++ b/python_modules/dagster/dagster_tests/core_tests/storage_tests/test_fs_io_manager.py
@@ -70,7 +70,7 @@ def test_fs_io_manager():
assert input_metadata_entry_a.label == "path"
assert input_metadata_entry_a.value == MetadataValue.path(filepath_a)
assert len(loaded_input_events) == 1
- assert loaded_input_events[0].event_specific_data.upstream_step_key == 'op_a'
+ assert loaded_input_events[0].event_specific_data.upstream_step_key == "op_a"
filepath_b = os.path.join(tmpdir_path, result.run_id, "op_b", "result")
result_metadata_entry_b = handled_output_events[1].event_specific_data.metadata_entries[0]
diff --git a/python_modules/dagster/dagster_tests/core_tests/test_output_materialization.py b/python_modules/dagster/dagster_tests/core_tests/test_output_materialization.py
index f81346cf83b26..fe4dd4e62dee1 100644
--- a/python_modules/dagster/dagster_tests/core_tests/test_output_materialization.py
+++ b/python_modules/dagster/dagster_tests/core_tests/test_output_materialization.py
@@ -361,7 +361,16 @@ def return_one():
assert result.success
event_types = [event.event_type_value for event in result.event_list]
- assert sum([True for event_type in event_types if event_type == DagsterEventType.ASSET_MATERIALIZATION.value]) == 2
+ assert (
+ sum(
+ [
+ True
+ for event_type in event_types
+ if event_type == DagsterEventType.ASSET_MATERIALIZATION.value
+ ]
+ )
+ == 2
+ )
@dagster_type_materializer(Int)
diff --git a/python_modules/libraries/dagstermill/dagstermill/asset_factory.py b/python_modules/libraries/dagstermill/dagstermill/asset_factory.py
index b76a967797e7e..a59e749f71f80 100644
--- a/python_modules/libraries/dagstermill/dagstermill/asset_factory.py
+++ b/python_modules/libraries/dagstermill/dagstermill/asset_factory.py
@@ -91,7 +91,8 @@ def _t_fn(context, **inputs):
)
except Exception as ex:
step_execution_context.log.warn(
- f"Error when attempting to materialize executed notebook: {serializable_error_info_from_exc_info(sys.exc_info())}"
+ "Error when attempting to materialize executed notebook:"
+ f" {serializable_error_info_from_exc_info(sys.exc_info())}"
)
# pylint: disable=no-member
# compat: