Skip to content

Export string sets in monitoring infos. (#31838) #28563

Export string sets in monitoring infos. (#31838)

Export string sets in monitoring infos. (#31838) #28563

GitHub Actions / Test Results failed Jul 12, 2024 in 0s

2 fail, 52 skipped, 84 pass in 1h 44m 25s

138 tests   84 ✅  1h 44m 25s ⏱️
  1 suites  52 💤
  1 files     2 ❌

Results for commit 50a3403.

Annotations

Check warning on line 0 in apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT

See this annotation in the file changed.

@github-actions github-actions / Test Results

test_streaming_inserts (apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT) failed

sdks/python/pytest_postCommitIT-df-py38.xml [took 12s]
Raw output
apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Fri, 12 Jul 2024 00:33:15 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '580', '-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "(88b4fd820646c1c9): The workflow could not be created. Causes: (88b4fd820646c8b6): Unable to get machine type information for machine type t2a-standard-1 in zone us-central1-c:  Please check that the machine type and zone are valid. Refer to https://cloud.google.com/compute/docs/regions-zones/regions-zones for the list of valid zones, and https://cloud.google.com/compute/docs/machine-types for the list of valid machine types. If you need assistance, please contact Google Cloud Support.",
    "status": "INVALID_ARGUMENT"
  }
}
>
self = <apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT testMethod=test_streaming_inserts>

    @pytest.mark.it_postcommit
    def test_streaming_inserts(self):
>     self._run_pubsub_bq_pipeline(WriteToBigQuery.Method.STREAMING_INSERTS)

apache_beam/io/gcp/bigquery_test.py:2534: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/io/gcp/bigquery_test.py:2526: in _run_pubsub_bq_pipeline
    _ = mesages | WriteToBigQuery(
apache_beam/pipeline.py:613: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:560: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:587: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/dataflow/test_dataflow_runner.py:53: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/dataflow/dataflow_runner.py:502: in run_pipeline
    self.dataflow_client.create_job(self.job), self)
apache_beam/utils/retry.py:298: in wrapper
    return fun(*args, **kwargs)
apache_beam/runners/dataflow/internal/apiclient.py:725: in create_job
    return self.submit_job_description(job)
apache_beam/utils/retry.py:298: in wrapper
    return fun(*args, **kwargs)
apache_beam/runners/dataflow/internal/apiclient.py:825: in submit_job_description
    response = self._client.projects_locations_jobs.Create(request)
apache_beam/runners/dataflow/internal/clients/dataflow/dataflow_v1b3_client.py:722: in Create
    return self._RunMethod(config, request, global_params=global_params)
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py:731: in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py:737: in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <apache_beam.runners.dataflow.internal.clients.dataflow.dataflow_v1b3_client.DataflowV1b3.ProjectsLocationsJobsService object at 0x7e1802334e80>
method_config = <ApiMethodInfo
 relative_path: 'v1b3/projects/{projectId}/locations/{location}/jobs'
 method_id: 'dataflow.projects.lo...DataflowProjectsLocationsJobsCreateRequest'
 response_type_name: 'Job'
 request_field: 'job'
 supports_download: False>
http_response = Response(info={'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Fri, 1...', request_url='https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json')
request = <DataflowProjectsLocationsJobsCreateRequest
 job: <Job
 clientRequestId: '20240712003314511353-9462'
 environment: <En...iles: []
 type: TypeValueValuesEnum(JOB_TYPE_STREAMING, 2)>
 location: 'us-central1'
 projectId: 'apache-beam-testing'>

    def __ProcessHttpResponse(self, method_config, http_response, request):
        """Process the given http response."""
        if http_response.status_code not in (http_client.OK,
                                             http_client.CREATED,
                                             http_client.NO_CONTENT):
>           raise exceptions.HttpError.FromResponse(
                http_response, method_config=method_config, request=request)
E           apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Fri, 12 Jul 2024 00:33:15 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '580', '-content-encoding': 'gzip'}>, content <{
E             "error": {
E               "code": 400,
E               "message": "(88b4fd820646c1c9): The workflow could not be created. Causes: (88b4fd820646c8b6): Unable to get machine type information for machine type t2a-standard-1 in zone us-central1-c:  Please check that the machine type and zone are valid. Refer to https://cloud.google.com/compute/docs/regions-zones/regions-zones for the list of valid zones, and https://cloud.google.com/compute/docs/machine-types for the list of valid machine types. If you need assistance, please contact Google Cloud Support.",
E               "status": "INVALID_ARGUMENT"
E             }
E           }
E           >

../../build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py:603: HttpBadRequestError

Check warning on line 0 in apache_beam.examples.streaming_wordcount_it_test.StreamingWordCountIT

See this annotation in the file changed.

@github-actions github-actions / Test Results

test_streaming_wordcount_it (apache_beam.examples.streaming_wordcount_it_test.StreamingWordCountIT) failed

sdks/python/pytest_postCommitIT-df-py38.xml [took 19s]
Raw output
apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Fri, 12 Jul 2024 01:06:06 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '580', '-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "(45fb1e9dc9d84f5d): The workflow could not be created. Causes: (45fb1e9dc9d842ba): Unable to get machine type information for machine type t2a-standard-1 in zone us-central1-c:  Please check that the machine type and zone are valid. Refer to https://cloud.google.com/compute/docs/regions-zones/regions-zones for the list of valid zones, and https://cloud.google.com/compute/docs/machine-types for the list of valid machine types. If you need assistance, please contact Google Cloud Support.",
    "status": "INVALID_ARGUMENT"
  }
}
>
self = <apache_beam.examples.streaming_wordcount_it_test.StreamingWordCountIT testMethod=test_streaming_wordcount_it>

    @pytest.mark.it_postcommit
    def test_streaming_wordcount_it(self):
      # Build expected dataset.
      expected_msg = [('%d: 1' % num).encode('utf-8')
                      for num in range(DEFAULT_INPUT_NUMBERS)]
    
      # Set extra options to the pipeline for test purpose
      state_verifier = PipelineStateMatcher(PipelineState.RUNNING)
      pubsub_msg_verifier = PubSubMessageMatcher(
          self.project, self.output_sub.name, expected_msg, timeout=400)
      extra_opts = {
          'input_subscription': self.input_sub.name,
          'output_topic': self.output_topic.name,
          'wait_until_finish_duration': WAIT_UNTIL_FINISH_DURATION,
          'on_success_matcher': all_of(state_verifier, pubsub_msg_verifier)
      }
    
      # Generate input data and inject to PubSub.
      self._inject_numbers(self.input_topic, DEFAULT_INPUT_NUMBERS)
    
      # Get pipeline options from command argument: --test-pipeline-options,
      # and start pipeline job by calling pipeline main function.
>     streaming_wordcount.run(
          self.test_pipeline.get_full_options_as_args(**extra_opts),
          save_main_session=False)

apache_beam/examples/streaming_wordcount_it_test.py:104: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/examples/streaming_wordcount.py:103: in run
    output | beam.io.WriteToPubSub(known_args.output_topic)
apache_beam/pipeline.py:613: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:587: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/dataflow/test_dataflow_runner.py:53: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/dataflow/dataflow_runner.py:502: in run_pipeline
    self.dataflow_client.create_job(self.job), self)
apache_beam/utils/retry.py:298: in wrapper
    return fun(*args, **kwargs)
apache_beam/runners/dataflow/internal/apiclient.py:725: in create_job
    return self.submit_job_description(job)
apache_beam/utils/retry.py:298: in wrapper
    return fun(*args, **kwargs)
apache_beam/runners/dataflow/internal/apiclient.py:825: in submit_job_description
    response = self._client.projects_locations_jobs.Create(request)
apache_beam/runners/dataflow/internal/clients/dataflow/dataflow_v1b3_client.py:722: in Create
    return self._RunMethod(config, request, global_params=global_params)
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py:731: in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py:737: in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <apache_beam.runners.dataflow.internal.clients.dataflow.dataflow_v1b3_client.DataflowV1b3.ProjectsLocationsJobsService object at 0x7ea1249397c0>
method_config = <ApiMethodInfo
 relative_path: 'v1b3/projects/{projectId}/locations/{location}/jobs'
 method_id: 'dataflow.projects.lo...DataflowProjectsLocationsJobsCreateRequest'
 response_type_name: 'Job'
 request_field: 'job'
 supports_download: False>
http_response = Response(info={'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Fri, 1...', request_url='https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json')
request = <DataflowProjectsLocationsJobsCreateRequest
 job: <Job
 clientRequestId: '20240712010604538546-8518'
 environment: <En...iles: []
 type: TypeValueValuesEnum(JOB_TYPE_STREAMING, 2)>
 location: 'us-central1'
 projectId: 'apache-beam-testing'>

    def __ProcessHttpResponse(self, method_config, http_response, request):
        """Process the given http response."""
        if http_response.status_code not in (http_client.OK,
                                             http_client.CREATED,
                                             http_client.NO_CONTENT):
>           raise exceptions.HttpError.FromResponse(
                http_response, method_config=method_config, request=request)
E           apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Fri, 12 Jul 2024 01:06:06 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '580', '-content-encoding': 'gzip'}>, content <{
E             "error": {
E               "code": 400,
E               "message": "(45fb1e9dc9d84f5d): The workflow could not be created. Causes: (45fb1e9dc9d842ba): Unable to get machine type information for machine type t2a-standard-1 in zone us-central1-c:  Please check that the machine type and zone are valid. Refer to https://cloud.google.com/compute/docs/regions-zones/regions-zones for the list of valid zones, and https://cloud.google.com/compute/docs/machine-types for the list of valid machine types. If you need assistance, please contact Google Cloud Support.",
E               "status": "INVALID_ARGUMENT"
E             }
E           }
E           >

../../build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py:603: HttpBadRequestError

Check notice on line 0 in .github

See this annotation in the file changed.

@github-actions github-actions / Test Results

52 skipped tests found

There are 52 skipped tests, see "Raw output" for the full list of skipped tests.
Raw output
apache_beam.examples.inference.tfx_bsl.tfx_bsl_inference_it_test.TFXRunInferenceTests ‑ test_tfx_run_inference_mobilenetv2
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.examples.streaming_wordcount_debugging_it_test.StreamingWordcountDebuggingIT ‑ test_streaming_wordcount_debugging_it
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_copy_jobs
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_dynamic_destinations
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_schema_autodetect
apache_beam.ml.gcp.naturallanguageml_test_it.NaturalLanguageMlTestIT ‑ test_analyzing_syntax
apache_beam.ml.gcp.videointelligenceml_test_it.VideoIntelligenceMlTestIT ‑ test_label_detection_with_video_context
apache_beam.ml.inference.base_test.RunInferenceBaseTest ‑ test_run_inference_with_side_inputin_streaming
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_it_test.OnnxInference ‑ test_onnx_run_inference_roberta_sentiment_classification
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_imagenet_mobilenetv2
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_imagenet_image_segmentation
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification_large_model
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_with_weights_classification
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch_large_model
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_single_batch
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.runners.dataflow.dataflow_exercise_metrics_pipeline_test.ExerciseMetricsPipelineTest ‑ test_metrics_it
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_basic_execution
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs_with_watermark_advancement
apache_beam.transforms.enrichment_handlers.feast_feature_store_it_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_test
apache_beam.transforms.periodicsequence_it_test.PeriodicSequenceIT ‑ test_periodicsequence_outputs_valid_watermarks_it
apache_beam.typehints.pytorch_type_compatibility_test
apache_beam.yaml.yaml_ml_test

Check notice on line 0 in .github

See this annotation in the file changed.

@github-actions github-actions / Test Results

138 tests found

There are 138 tests, see "Raw output" for the full list of tests.
Raw output
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_direct_read
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_direct_read_with_project
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_export_with_project
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_with_computation
apache_beam.examples.complete.autocomplete_test.AutocompleteTest ‑ test_autocomplete_it
apache_beam.examples.complete.game.game_stats_it_test.GameStatsIT ‑ test_game_stats_it
apache_beam.examples.complete.game.hourly_team_score_it_test.HourlyTeamScoreIT ‑ test_hourly_team_score_it
apache_beam.examples.complete.game.leader_board_it_test.LeaderBoardIT ‑ test_leader_board_it
apache_beam.examples.complete.game.user_score_it_test.UserScoreIT ‑ test_user_score_it
apache_beam.examples.complete.juliaset.juliaset.juliaset_test_it.JuliaSetTestIT ‑ test_run_example_with_setup_file
apache_beam.examples.cookbook.bigquery_tornadoes_it_test.BigqueryTornadoesIT ‑ test_bigquery_tornadoes_it
apache_beam.examples.cookbook.bigtableio_it_test.BigtableIOWriteTest ‑ test_bigtable_write
apache_beam.examples.cookbook.datastore_wordcount_it_test.DatastoreWordCountIT ‑ test_datastore_wordcount_it
apache_beam.examples.dataframe.flight_delays_it_test.FlightDelaysTest ‑ test_flight_delays
apache_beam.examples.dataframe.taxiride_it_test.TaxirideIT ‑ test_aggregation
apache_beam.examples.dataframe.taxiride_it_test.TaxirideIT ‑ test_enrich
apache_beam.examples.fastavro_it_test.FastavroIT ‑ test_avro_it
apache_beam.examples.inference.tfx_bsl.tfx_bsl_inference_it_test.TFXRunInferenceTests ‑ test_tfx_run_inference_mobilenetv2
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.examples.streaming_wordcount_debugging_it_test.StreamingWordcountDebuggingIT ‑ test_streaming_wordcount_debugging_it
apache_beam.examples.streaming_wordcount_it_test.StreamingWordCountIT ‑ test_streaming_wordcount_it
apache_beam.examples.wordcount_it_test.WordCountIT ‑ test_wordcount_impersonation_it
apache_beam.examples.wordcount_it_test.WordCountIT ‑ test_wordcount_it
apache_beam.io.fileio_test.MatchIntegrationTest ‑ test_transform_on_gcs
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_legacy_sql
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_new_types
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_new_types_avro
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_standard_sql
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_batch_copy_jobs_with_no_input_schema
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_copy_jobs
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_dynamic_destinations
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_multiple_destinations_transform
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_one_job_fails_all_jobs_fail
apache_beam.io.gcp.bigquery_io_read_it_test.BigqueryIOReadIT ‑ test_bigquery_read_1M_python
apache_beam.io.gcp.bigquery_io_read_it_test.BigqueryIOReadIT ‑ test_bigquery_read_custom_1M_python
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_direct_read
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_export_read
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_file_loads_write
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_query_read
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_streaming_inserts
apache_beam.io.gcp.bigquery_read_it_test.ReadAllBQTests ‑ test_read_queries
apache_beam.io.gcp.bigquery_read_it_test.ReadInteractiveRunnerTests ‑ test_read_in_interactive_runner
apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests ‑ test_iobase_source
apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests ‑ test_native_source
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_iobase_source
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_native_source
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_table_schema_retrieve
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_table_schema_retrieve_specifying_only_table
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_table_schema_retrieve_with_direct_read
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_column_selection
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_column_selection_and_row_restriction
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_native_datetime
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_query
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_query_and_filters
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_row_restriction
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_very_selective_filters
apache_beam.io.gcp.bigquery_test.BigQueryFileLoadsIntegrationTests ‑ test_avro_file_load
apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests ‑ test_multiple_destinations_transform
apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests ‑ test_value_provider_transform
apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT ‑ test_file_loads
apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT ‑ test_streaming_inserts
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_insert_errors_reporting
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_insert_non_transient_api_call_error
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_new_types
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_schema_autodetect
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_without_schema
apache_beam.io.gcp.datastore.v1new.datastore_write_it_test.DatastoreWriteIT ‑ test_datastore_write_limit
apache_beam.io.gcp.gcsfilesystem_integration_test.GcsFileSystemIntegrationTest ‑ test_copy
apache_beam.io.gcp.gcsfilesystem_integration_test.GcsFileSystemIntegrationTest ‑ test_rename
apache_beam.io.gcp.gcsfilesystem_integration_test.GcsFileSystemIntegrationTest ‑ test_rename_error
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_batch_copy_and_delete
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_create_default_bucket
apache_beam.io.gcp.healthcare.dicomio_integration_test.DICOMIoIntegrationTest ‑ test_dicom_search_instances
apache_beam.io.gcp.healthcare.dicomio_integration_test.DICOMIoIntegrationTest ‑ test_dicom_store_instance_from_gcs
apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest ‑ test_streaming_data_only
apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest ‑ test_streaming_with_attributes
apache_beam.io.parquetio_it_test.TestParquetIT ‑ test_parquetio_it
apache_beam.ml.gcp.cloud_dlp_it_test.CloudDLPIT ‑ test_deidentification
apache_beam.ml.gcp.cloud_dlp_it_test.CloudDLPIT ‑ test_inspection
apache_beam.ml.gcp.naturallanguageml_test_it.NaturalLanguageMlTestIT ‑ test_analyzing_syntax
apache_beam.ml.gcp.recommendations_ai_test_it.RecommendationAIIT ‑ test_create_catalog_item
apache_beam.ml.gcp.recommendations_ai_test_it.RecommendationAIIT ‑ test_create_user_event
apache_beam.ml.gcp.recommendations_ai_test_it.RecommendationAIIT ‑ test_predict
apache_beam.ml.gcp.videointelligenceml_test_it.VideoIntelligenceMlTestIT ‑ test_label_detection_with_video_context
apache_beam.ml.gcp.visionml_test_it.VisionMlTestIT ‑ test_text_detection_with_language_hint
apache_beam.ml.inference.base_test.RunInferenceBaseTest ‑ test_run_inference_with_side_inputin_streaming
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_it_test.OnnxInference ‑ test_onnx_run_inference_roberta_sentiment_classification
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_imagenet_mobilenetv2
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.sklearn_inference_it_test.SklearnInference ‑ test_sklearn_mnist_classification
apache_beam.ml.inference.sklearn_inference_it_test.SklearnInference ‑ test_sklearn_mnist_classification_large_model
apache_beam.ml.inference.sklearn_inference_it_test.SklearnInference ‑ test_sklearn_regression
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_imagenet_image_segmentation
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification_large_model
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_with_weights_classification
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch_large_model
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_single_batch
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.runners.dataflow.dataflow_exercise_metrics_pipeline_test.ExerciseMetricsPipelineTest ‑ test_metrics_it
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_basic_execution
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs_with_watermark_advancement
apache_beam.transforms.enrichment_handlers.feast_feature_store_it_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_test
apache_beam.transforms.external_it_test.ExternalTransformIT ‑ test_job_python_from_python_it
apache_beam.transforms.periodicsequence_it_test.PeriodicSequenceIT ‑ test_periodicsequence_outputs_valid_watermarks_it
apache_beam.typehints.pytorch_type_compatibility_test
apache_beam.yaml.yaml_ml_test