diff --git a/contributing-docs/testing/task_sdk_integration_tests.rst b/contributing-docs/testing/task_sdk_integration_tests.rst index 132b2822bafc6..f7efb2e6f403f 100644 --- a/contributing-docs/testing/task_sdk_integration_tests.rst +++ b/contributing-docs/testing/task_sdk_integration_tests.rst @@ -85,7 +85,7 @@ reproducibility: breeze testing task-sdk-integration-tests task_sdk_tests/test_task_sdk_health.py # Run with custom Docker image - DOCKER_IMAGE=my-custom-airflow-imahge:latest breeze testing task-sdk-integration-tests + DOCKER_IMAGE=my-custom-airflow-image:latest breeze testing task-sdk-integration-tests Running in Your Current Virtual Environment ........................................... diff --git a/dev/breeze/doc/images/output_ci-image.svg b/dev/breeze/doc/images/output_ci-image.svg index 0c614dfc94703..62ca31ecad645 100644 --- a/dev/breeze/doc/images/output_ci-image.svg +++ b/dev/breeze/doc/images/output_ci-image.svg @@ -118,8 +118,8 @@ verify               Verify CI image.                                                                              save                 Save CI image to a file.                                                                      load                 Load CI image from a file.                                                                    -export-mount-cache   Export content of the the mount cache to a directory.                                         -import-mount-cache   Export content of the the mount cache to a directory.                                         +export-mount-cache   Export content of the mount cache to a directory.                                             +import-mount-cache   Export content of the mount cache to a directory.                                             ╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ diff --git a/dev/breeze/doc/images/output_ci-image.txt b/dev/breeze/doc/images/output_ci-image.txt index 25b4a0db49abc..fcc45c9593176 100644 --- a/dev/breeze/doc/images/output_ci-image.txt +++ b/dev/breeze/doc/images/output_ci-image.txt @@ -1 +1 @@ -de6e1ee2eb602569a8cd1a3c6c4cbafe +7745b7d70e4a06b724ed92814d416670 diff --git a/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg b/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg index c1a516dbdefd7..4c56e91807a2a 100644 --- a/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg +++ b/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg @@ -101,7 +101,7 @@ Usage:breeze ci-image export-mount-cache[OPTIONS] -Export content of the the mount cache to a directory. +Export content of the mount cache to a directory. ╭─ Export flags ───────────────────────────────────────────────────────────────────────────────────────────────────────╮ *--cache-filePath to the file where cache is going to be exported(FILE)[required] diff --git a/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt b/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt index 6d1c1347207a2..32ccab745cc9c 100644 --- a/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt +++ b/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt @@ -1 +1 @@ -824109d9d232ba19a9aa0781ba8fccd9 +ebc1ae7a092c1e541957fb1e79829249 diff --git a/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg b/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg index 32eb4fdfebe6e..5b31140f32717 100644 --- a/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg +++ b/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg @@ -101,7 +101,7 @@ Usage:breeze ci-image import-mount-cache[OPTIONS] -Export content of the the mount cache to a directory. +Export content of the mount cache to a directory. ╭─ Import flags ───────────────────────────────────────────────────────────────────────────────────────────────────────╮ *--cache-filePath to the file where cache is stored(FILE)[required] diff --git a/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt b/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt index 00f5024812d07..7549b6730abea 100644 --- a/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt +++ b/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt @@ -1 +1 @@ -65be4537eea8d56d3ddc34d4ccab60c7 +669f8d8b36a1d9dde983862db6a69ce4 diff --git a/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py b/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py index 184562da62862..71655db360b67 100644 --- a/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py +++ b/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py @@ -930,7 +930,7 @@ def export_mount_cache( cache_file: Path, ): """ - Export content of the the mount cache to a directory. + Export content of the mount cache to a directory. """ perform_environment_checks() make_sure_builder_configured(params=BuildCiParams(builder=builder)) @@ -995,7 +995,7 @@ def import_mount_cache( cache_file: Path, ): """ - Export content of the the mount cache to a directory. + Export content of the mount cache to a directory. """ perform_environment_checks() make_sure_builder_configured(params=BuildCiParams(builder=builder)) diff --git a/docs/spelling_wordlist.txt b/docs/spelling_wordlist.txt index f9922cb4d8d6f..4d89962152fc2 100644 --- a/docs/spelling_wordlist.txt +++ b/docs/spelling_wordlist.txt @@ -253,6 +253,8 @@ chakra Changelog changelog charset +chatbot +chatbots Checklicence checklicence CheckOperator diff --git a/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py b/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py index ff71ad365a5d2..ef44b4b0dfa1d 100644 --- a/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py +++ b/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py @@ -790,7 +790,7 @@ def async_conn(self): async def get_async_conn(self): """Get an aiobotocore client to use for async operations.""" # We have to wrap the call `self.get_client_type` in another call `_get_async_conn`, - # because one of it's arguments `self.region_name` is a `@property` decorated function + # because one of its arguments `self.region_name` is a `@property` decorated function # calling the cached property `self.conn_config` at the end. return await sync_to_async(self._get_async_conn)() diff --git a/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py b/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py index 8201f32e76f01..bdf0f220a18f3 100644 --- a/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py +++ b/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py @@ -33,7 +33,7 @@ class ChimeWebhookHook(HttpHook): """ Interact with Amazon Chime Webhooks to create notifications. - .. warning:: This hook is only designed to work with web hooks and not chat bots. + .. warning:: This hook is only designed to work with web hooks and not chatbots. :param chime_conn_id: :ref:`Amazon Chime Connection ID ` with Endpoint as `https://hooks.chime.aws` and the webhook token diff --git a/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py b/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py index e4f3d211bea93..d4bfcc9532230 100644 --- a/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py +++ b/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py @@ -32,7 +32,7 @@ class CloudFormationCreateStackOperator(AwsBaseOperator[CloudFormationHook]): """ - An operator that creates a AWS CloudFormation stack. + An operator that creates an AWS CloudFormation stack. .. seealso:: For more information on how to use this operator, take a look at the guide: @@ -68,7 +68,7 @@ def execute(self, context: Context): class CloudFormationDeleteStackOperator(AwsBaseOperator[CloudFormationHook]): """ - An operator that deletes a AWS CloudFormation stack. + An operator that deletes an AWS CloudFormation stack. .. seealso:: For more information on how to use this operator, take a look at the guide: diff --git a/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py b/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py index f10ae9dbcc68a..52aa45703963e 100644 --- a/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py +++ b/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py @@ -49,12 +49,12 @@ class AzureBlobStorageToS3Operator(BaseOperator): :param prefix: Prefix string which filters objects whose name begin with this prefix. (templated) :param delimiter: The delimiter by which you want to filter the objects. (templated) - For e.g to lists the CSV files from in a directory in GCS you would use + For e.g. to lists the CSV files from in a directory in GCS you would use delimiter='.csv'. :param aws_conn_id: Connection id of the S3 connection to use :param dest_s3_key: The base S3 key to be used to store the files. (templated) :param dest_verify: Whether or not to verify SSL certificates for S3 connection. - By default SSL certificates are verified. + By default, SSL certificates are verified. You can provide the following values: - ``False``: do not validate SSL certificates. SSL will still be used diff --git a/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py b/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py index da189d66c4fec..0308b54988d8f 100644 --- a/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py +++ b/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py @@ -160,7 +160,7 @@ class BeamBasePipelineOperator(BaseOperator, BeamDataflowMixin, ABC): """ Abstract base class for Beam Pipeline Operators. - :param runner: Runner on which pipeline will be run. By default "DirectRunner" is being used. + :param runner: Runner on which pipeline will be run. By default, "DirectRunner" is being used. Other possible options: DataflowRunner, SparkRunner, FlinkRunner, PortableRunner. See: :class:`~providers.apache.beam.hooks.beam.BeamRunnerType` See: https://beam.apache.org/documentation/runners/capability-matrix/ diff --git a/providers/celery/src/airflow/providers/celery/executors/celery_executor.py b/providers/celery/src/airflow/providers/celery/executors/celery_executor.py index 13b5ce3237f4b..c8107b9160ebe 100644 --- a/providers/celery/src/airflow/providers/celery/executors/celery_executor.py +++ b/providers/celery/src/airflow/providers/celery/executors/celery_executor.py @@ -350,7 +350,7 @@ def _process_workloads(self, workloads: Sequence[workloads.All]) -> None: def _send_tasks(self, task_tuples_to_send: Sequence[TaskInstanceInCelery]): first_task = next(t[-1] for t in task_tuples_to_send) - # Celery state queries will stuck if we do not use one same backend + # Celery state queries will be stuck if we do not use one same backend # for all tasks. cached_celery_backend = first_task.backend diff --git a/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py b/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py index d618d87d429ec..8889771e16e71 100644 --- a/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py +++ b/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py @@ -112,7 +112,7 @@ def on_celery_import_modules(*args, **kwargs): """ Preload some "expensive" airflow modules once, so other task processes won't have to import it again. - Loading these for each task adds 0.3-0.5s *per task* before the task can run. For long running tasks this + Loading these for each task adds 0.3-0.5s *per task* before the task can run. For long-running tasks this doesn't matter, but for short tasks this starts to be a noticeable impact. """ import jinja2.ext # noqa: F401 diff --git a/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py b/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py index 47be19f5f6876..d266d025277d5 100644 --- a/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py +++ b/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py @@ -508,7 +508,7 @@ class CloudSqlProxyRunner(LoggingMixin): :param project_id: Optional id of the Google Cloud project to connect to - it overwrites default project id taken from the Google Cloud connection. :param sql_proxy_version: Specific version of SQL proxy to download - (for example 'v1.13'). By default latest version is downloaded. + (for example 'v1.13'). By default, latest version is downloaded. :param sql_proxy_binary_path: If specified, then proxy will be used from the path specified rather than dynamically generated. This means that if the binary is not present in that path it will also be downloaded. @@ -693,7 +693,7 @@ def stop_proxy(self) -> None: self.log.info("Skipped removing proxy - it was not downloaded: %s", self.sql_proxy_path) if os.path.isfile(self.credentials_path): self.log.info("Removing generated credentials file %s", self.credentials_path) - # Here file cannot be delete by concurrent task (each task has its own copy) + # Here file cannot be deleted by concurrent task (each task has its own copy) os.remove(self.credentials_path) def get_proxy_version(self) -> str | None: diff --git a/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py b/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py index 66a1e7ee0a6a7..61044e17b27ef 100644 --- a/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py +++ b/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py @@ -110,7 +110,7 @@ def create_batch_prediction_job( :param project_id: Required. Project to run training in. :param region: Required. Location to run training in. :param job_display_name: Required. The user-defined name of the BatchPredictionJob. The name can be - up to 128 characters long and can be consist of any UTF-8 characters. + up to 128 characters long and can consist of any UTF-8 characters. :param model_name: Required. A fully-qualified model resource name or model ID. :param instances_format: Required. The format in which instances are provided. Must be one of the formats listed in `Model.supported_input_storage_formats`. Default is "jsonl" when using @@ -267,7 +267,7 @@ def submit_batch_prediction_job( :param project_id: Required. Project to run training in. :param region: Required. Location to run training in. :param job_display_name: Required. The user-defined name of the BatchPredictionJob. The name can be - up to 128 characters long and can be consist of any UTF-8 characters. + up to 128 characters long and can consist of any UTF-8 characters. :param model_name: Required. A fully-qualified model resource name or model ID. :param instances_format: Required. The format in which instances are provided. Must be one of the formats listed in `Model.supported_input_storage_formats`. Default is "jsonl" when using diff --git a/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py b/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py index f3b858c63a47f..9d78152114f8d 100644 --- a/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py +++ b/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py @@ -244,7 +244,7 @@ class CloudBatchListTasksOperator(GoogleCloudBaseOperator): :param job_name: Required. The name of the job for which to list tasks. :param gcp_conn_id: The connection ID used to connect to Google Cloud. :param filter: The filter based on which to list the jobs. If left empty, all the jobs are listed. - :param group_name: The name of the group that owns the task. By default it's `group0`. + :param group_name: The name of the group that owns the task. By default, it's `group0`. :param limit: The number of tasks to list. If left empty, all the tasks matching the filter will be returned. :param impersonation_chain: Optional service account to impersonate using short-term diff --git a/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py b/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py index 3793b00905179..95f599189f29a 100644 --- a/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py +++ b/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py @@ -141,7 +141,7 @@ def execute(self, context: Context): class CloudRunDeleteJobOperator(GoogleCloudBaseOperator): """ - Deletes a job and wait for the the operation to be completed. Pushes the deleted job to xcom. + Deletes a job and wait for the operation to be completed. Pushes the deleted job to xcom. :param project_id: Required. The ID of the Google Cloud project that the service belongs to. :param region: Required. The ID of the Google Cloud region that the service belongs to. diff --git a/providers/google/src/airflow/providers/google/suite/hooks/calendar.py b/providers/google/src/airflow/providers/google/suite/hooks/calendar.py index ddc0078bba17d..163f117d74d82 100644 --- a/providers/google/src/airflow/providers/google/suite/hooks/calendar.py +++ b/providers/google/src/airflow/providers/google/suite/hooks/calendar.py @@ -104,7 +104,7 @@ def get_events( only the participant is returned. :param max_results: Optional. Maximum number of events returned on one result page. Incomplete pages can be detected by a non-empty ``nextPageToken`` field in the response. - By default the value is 250 events. The page size can never be larger than 2500 events + By default, the value is 250 events. The page size can never be larger than 2500 events :param order_by: Optional. Acceptable values are ``"startTime"`` or "updated" :param private_extended_property: Optional. Extended properties constraint specified as ``propertyName=value``. Matches only private properties. This parameter might be repeated diff --git a/providers/http/docs/changelog.rst b/providers/http/docs/changelog.rst index cf18187667572..24f43e223feea 100644 --- a/providers/http/docs/changelog.rst +++ b/providers/http/docs/changelog.rst @@ -284,7 +284,7 @@ Bug Fixes Misc ~~~~ -* ``add lowerbount to requests-toolbelt and replace requests_toolbelt with requests-toolbelt (#43020)`` +* ``add lower bound to requests-toolbelt and replace requests_toolbelt with requests-toolbelt (#43020)`` * ``Set lower bound to asgiref>=2.3.0 (#43001)`` @@ -583,7 +583,7 @@ Misc Features ~~~~~~~~ -* ``Add non login-password auth support for SimpleHttpOpeator (#29206)`` +* ``Add non login-password auth support for SimpleHttpOperator (#29206)`` 4.2.0 ..... diff --git a/providers/keycloak/docs/auth-manager/manage/permissions.rst b/providers/keycloak/docs/auth-manager/manage/permissions.rst index 94d665ac63dac..fc1e1e252bf0a 100644 --- a/providers/keycloak/docs/auth-manager/manage/permissions.rst +++ b/providers/keycloak/docs/auth-manager/manage/permissions.rst @@ -86,7 +86,7 @@ This will create More resources about permissions can be found in the official documentation of Keycloak: -1- `Keylcloak Authorization Process `_ +1- `Keycloak Authorization Process `_ 2- `Keycloak Permission Overview `_ diff --git a/providers/ssh/docs/changelog.rst b/providers/ssh/docs/changelog.rst index 697e636f7a318..046bb47fca2e6 100644 --- a/providers/ssh/docs/changelog.rst +++ b/providers/ssh/docs/changelog.rst @@ -713,7 +713,7 @@ Bug Fixes * ``Updated documentation for June 2021 provider release (#16294)`` * ``Add Connection Documentation to more Providers (#15408)`` * ``More documentation update for June providers release (#16405)`` - * ``Synchronizes updated changelog after buggfix release (#16464)`` + * ``Synchronizes updated changelog after bugfix release (#16464)`` 1.3.0 ..... diff --git a/task-sdk/src/airflow/sdk/api/client.py b/task-sdk/src/airflow/sdk/api/client.py index 0a3de02894150..c70cec0e89577 100644 --- a/task-sdk/src/airflow/sdk/api/client.py +++ b/task-sdk/src/airflow/sdk/api/client.py @@ -807,7 +807,7 @@ def auth_flow(self, request: httpx.Request): yield request -# This exists as a aid for debugging or local running via the `dry_run` argument to Client. It doesn't make +# This exists as an aid for debugging or local running via the `dry_run` argument to Client. It doesn't make # sense for returning connections etc. def noop_handler(request: httpx.Request) -> httpx.Response: path = request.url.path