diff --git a/providers/databricks/tests/unit/databricks/hooks/test_databricks_sql.py b/providers/databricks/tests/unit/databricks/hooks/test_databricks_sql.py index 244ec605a1796..4fd5d80c6521e 100644 --- a/providers/databricks/tests/unit/databricks/hooks/test_databricks_sql.py +++ b/providers/databricks/tests/unit/databricks/hooks/test_databricks_sql.py @@ -35,8 +35,6 @@ from airflow.providers.databricks.hooks.databricks_sql import DatabricksSqlHook, create_timeout_thread from airflow.utils.session import provide_session -pytestmark = pytest.mark.db_test - TASK_ID = "databricks-sql-operator" DEFAULT_CONN_ID = "databricks_default" HOST = "xx.cloud.databricks.com" diff --git a/providers/databricks/tests/unit/databricks/operators/test_databricks.py b/providers/databricks/tests/unit/databricks/operators/test_databricks.py index 183f33a7f6629..7c221a4d8f61d 100644 --- a/providers/databricks/tests/unit/databricks/operators/test_databricks.py +++ b/providers/databricks/tests/unit/databricks/operators/test_databricks.py @@ -48,8 +48,6 @@ ) from airflow.providers.databricks.utils import databricks as utils -pytestmark = pytest.mark.db_test - DATE = "2017-04-20" TASK_ID = "databricks-operator" DEFAULT_CONN_ID = "databricks_default" @@ -812,7 +810,6 @@ def test_init_with_merging(self): ) assert expected == utils.normalise_json_content(op.json) - @pytest.mark.db_test def test_init_with_templating(self): json = { "new_cluster": NEW_CLUSTER, @@ -1250,7 +1247,6 @@ def test_init_with_merging(self): assert expected == op.json - @pytest.mark.db_test def test_init_with_templating(self): json = {"notebook_params": NOTEBOOK_PARAMS, "jar_params": TEMPLATED_JAR_PARAMS} diff --git a/providers/databricks/tests/unit/databricks/operators/test_databricks_workflow.py b/providers/databricks/tests/unit/databricks/operators/test_databricks_workflow.py index 60a6d343ba31e..0a4c3ee1f9438 100644 --- a/providers/databricks/tests/unit/databricks/operators/test_databricks_workflow.py +++ b/providers/databricks/tests/unit/databricks/operators/test_databricks_workflow.py @@ -34,8 +34,6 @@ from airflow.providers.standard.operators.empty import EmptyOperator from airflow.utils import timezone -pytestmark = pytest.mark.db_test - DEFAULT_DATE = timezone.datetime(2021, 1, 1) diff --git a/providers/databricks/tests/unit/databricks/plugins/test_databricks_workflow.py b/providers/databricks/tests/unit/databricks/plugins/test_databricks_workflow.py index 3c6f57c1d9d5d..f79dbcbcaeb12 100644 --- a/providers/databricks/tests/unit/databricks/plugins/test_databricks_workflow.py +++ b/providers/databricks/tests/unit/databricks/plugins/test_databricks_workflow.py @@ -152,7 +152,6 @@ def app(): yield app -@pytest.mark.db_test def test_get_task_instance(app): with app.app_context(): operator = Mock() @@ -170,7 +169,6 @@ def test_get_task_instance(app): assert result == dag_run -@pytest.mark.db_test def test_get_return_url_dag_id_run_id(app): dag_id = "example_dag" run_id = "example_run" @@ -182,7 +180,6 @@ def test_get_return_url_dag_id_run_id(app): assert actual_url == expected_url, f"Expected {expected_url}, got {actual_url}" -@pytest.mark.db_test def test_workflow_job_run_link(app): with app.app_context(): link = WorkflowJobRunLink() @@ -220,7 +217,6 @@ def test_workflow_job_run_link(app): @pytest.mark.skipif( RUNNING_TESTS_AGAINST_AIRFLOW_PACKAGES, reason="Web plugin test doesn't work when not against sources" ) -@pytest.mark.db_test def test_workflow_job_repair_single_failed_link(app): with app.app_context(): link = WorkflowJobRepairSingleTaskLink()