diff --git a/contributing-docs/testing/task_sdk_integration_tests.rst b/contributing-docs/testing/task_sdk_integration_tests.rst
index 132b2822bafc6..f7efb2e6f403f 100644
--- a/contributing-docs/testing/task_sdk_integration_tests.rst
+++ b/contributing-docs/testing/task_sdk_integration_tests.rst
@@ -85,7 +85,7 @@ reproducibility:
breeze testing task-sdk-integration-tests task_sdk_tests/test_task_sdk_health.py
# Run with custom Docker image
- DOCKER_IMAGE=my-custom-airflow-imahge:latest breeze testing task-sdk-integration-tests
+ DOCKER_IMAGE=my-custom-airflow-image:latest breeze testing task-sdk-integration-tests
Running in Your Current Virtual Environment
...........................................
diff --git a/dev/breeze/doc/images/output_ci-image.svg b/dev/breeze/doc/images/output_ci-image.svg
index 0c614dfc94703..62ca31ecad645 100644
--- a/dev/breeze/doc/images/output_ci-image.svg
+++ b/dev/breeze/doc/images/output_ci-image.svg
@@ -118,8 +118,8 @@
│verify Verify CI image. │
│save Save CI image to a file. │
│load Load CI image from a file. │
-│export-mount-cache Export content of the the mount cache to a directory. │
-│import-mount-cache Export content of the the mount cache to a directory. │
+│export-mount-cache Export content of the mount cache to a directory. │
+│import-mount-cache Export content of the mount cache to a directory. │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
diff --git a/dev/breeze/doc/images/output_ci-image.txt b/dev/breeze/doc/images/output_ci-image.txt
index 25b4a0db49abc..fcc45c9593176 100644
--- a/dev/breeze/doc/images/output_ci-image.txt
+++ b/dev/breeze/doc/images/output_ci-image.txt
@@ -1 +1 @@
-de6e1ee2eb602569a8cd1a3c6c4cbafe
+7745b7d70e4a06b724ed92814d416670
diff --git a/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg b/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg
index c1a516dbdefd7..4c56e91807a2a 100644
--- a/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg
+++ b/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg
@@ -101,7 +101,7 @@
Usage:breeze ci-image export-mount-cache[OPTIONS]
-Export content of the the mount cache to a directory.
+Export content of the mount cache to a directory.
╭─ Export flags ───────────────────────────────────────────────────────────────────────────────────────────────────────╮
│*--cache-filePath to the file where cache is going to be exported(FILE)[required]│
diff --git a/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt b/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt
index 6d1c1347207a2..32ccab745cc9c 100644
--- a/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt
+++ b/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt
@@ -1 +1 @@
-824109d9d232ba19a9aa0781ba8fccd9
+ebc1ae7a092c1e541957fb1e79829249
diff --git a/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg b/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg
index 32eb4fdfebe6e..5b31140f32717 100644
--- a/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg
+++ b/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg
@@ -101,7 +101,7 @@
Usage:breeze ci-image import-mount-cache[OPTIONS]
-Export content of the the mount cache to a directory.
+Export content of the mount cache to a directory.
╭─ Import flags ───────────────────────────────────────────────────────────────────────────────────────────────────────╮
│*--cache-filePath to the file where cache is stored(FILE)[required]│
diff --git a/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt b/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt
index 00f5024812d07..7549b6730abea 100644
--- a/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt
+++ b/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt
@@ -1 +1 @@
-65be4537eea8d56d3ddc34d4ccab60c7
+669f8d8b36a1d9dde983862db6a69ce4
diff --git a/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py b/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py
index 184562da62862..71655db360b67 100644
--- a/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py
+++ b/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py
@@ -930,7 +930,7 @@ def export_mount_cache(
cache_file: Path,
):
"""
- Export content of the the mount cache to a directory.
+ Export content of the mount cache to a directory.
"""
perform_environment_checks()
make_sure_builder_configured(params=BuildCiParams(builder=builder))
@@ -995,7 +995,7 @@ def import_mount_cache(
cache_file: Path,
):
"""
- Export content of the the mount cache to a directory.
+ Export content of the mount cache to a directory.
"""
perform_environment_checks()
make_sure_builder_configured(params=BuildCiParams(builder=builder))
diff --git a/docs/spelling_wordlist.txt b/docs/spelling_wordlist.txt
index f9922cb4d8d6f..4d89962152fc2 100644
--- a/docs/spelling_wordlist.txt
+++ b/docs/spelling_wordlist.txt
@@ -253,6 +253,8 @@ chakra
Changelog
changelog
charset
+chatbot
+chatbots
Checklicence
checklicence
CheckOperator
diff --git a/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py b/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py
index ff71ad365a5d2..ef44b4b0dfa1d 100644
--- a/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py
+++ b/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py
@@ -790,7 +790,7 @@ def async_conn(self):
async def get_async_conn(self):
"""Get an aiobotocore client to use for async operations."""
# We have to wrap the call `self.get_client_type` in another call `_get_async_conn`,
- # because one of it's arguments `self.region_name` is a `@property` decorated function
+ # because one of its arguments `self.region_name` is a `@property` decorated function
# calling the cached property `self.conn_config` at the end.
return await sync_to_async(self._get_async_conn)()
diff --git a/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py b/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py
index 8201f32e76f01..bdf0f220a18f3 100644
--- a/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py
+++ b/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py
@@ -33,7 +33,7 @@ class ChimeWebhookHook(HttpHook):
"""
Interact with Amazon Chime Webhooks to create notifications.
- .. warning:: This hook is only designed to work with web hooks and not chat bots.
+ .. warning:: This hook is only designed to work with web hooks and not chatbots.
:param chime_conn_id: :ref:`Amazon Chime Connection ID `
with Endpoint as `https://hooks.chime.aws` and the webhook token
diff --git a/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py b/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py
index e4f3d211bea93..d4bfcc9532230 100644
--- a/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py
+++ b/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py
@@ -32,7 +32,7 @@
class CloudFormationCreateStackOperator(AwsBaseOperator[CloudFormationHook]):
"""
- An operator that creates a AWS CloudFormation stack.
+ An operator that creates an AWS CloudFormation stack.
.. seealso::
For more information on how to use this operator, take a look at the guide:
@@ -68,7 +68,7 @@ def execute(self, context: Context):
class CloudFormationDeleteStackOperator(AwsBaseOperator[CloudFormationHook]):
"""
- An operator that deletes a AWS CloudFormation stack.
+ An operator that deletes an AWS CloudFormation stack.
.. seealso::
For more information on how to use this operator, take a look at the guide:
diff --git a/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py b/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py
index f10ae9dbcc68a..52aa45703963e 100644
--- a/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py
+++ b/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py
@@ -49,12 +49,12 @@ class AzureBlobStorageToS3Operator(BaseOperator):
:param prefix: Prefix string which filters objects whose name begin with
this prefix. (templated)
:param delimiter: The delimiter by which you want to filter the objects. (templated)
- For e.g to lists the CSV files from in a directory in GCS you would use
+ For e.g. to lists the CSV files from in a directory in GCS you would use
delimiter='.csv'.
:param aws_conn_id: Connection id of the S3 connection to use
:param dest_s3_key: The base S3 key to be used to store the files. (templated)
:param dest_verify: Whether or not to verify SSL certificates for S3 connection.
- By default SSL certificates are verified.
+ By default, SSL certificates are verified.
You can provide the following values:
- ``False``: do not validate SSL certificates. SSL will still be used
diff --git a/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py b/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py
index da189d66c4fec..0308b54988d8f 100644
--- a/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py
+++ b/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py
@@ -160,7 +160,7 @@ class BeamBasePipelineOperator(BaseOperator, BeamDataflowMixin, ABC):
"""
Abstract base class for Beam Pipeline Operators.
- :param runner: Runner on which pipeline will be run. By default "DirectRunner" is being used.
+ :param runner: Runner on which pipeline will be run. By default, "DirectRunner" is being used.
Other possible options: DataflowRunner, SparkRunner, FlinkRunner, PortableRunner.
See: :class:`~providers.apache.beam.hooks.beam.BeamRunnerType`
See: https://beam.apache.org/documentation/runners/capability-matrix/
diff --git a/providers/celery/src/airflow/providers/celery/executors/celery_executor.py b/providers/celery/src/airflow/providers/celery/executors/celery_executor.py
index 13b5ce3237f4b..c8107b9160ebe 100644
--- a/providers/celery/src/airflow/providers/celery/executors/celery_executor.py
+++ b/providers/celery/src/airflow/providers/celery/executors/celery_executor.py
@@ -350,7 +350,7 @@ def _process_workloads(self, workloads: Sequence[workloads.All]) -> None:
def _send_tasks(self, task_tuples_to_send: Sequence[TaskInstanceInCelery]):
first_task = next(t[-1] for t in task_tuples_to_send)
- # Celery state queries will stuck if we do not use one same backend
+ # Celery state queries will be stuck if we do not use one same backend
# for all tasks.
cached_celery_backend = first_task.backend
diff --git a/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py b/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py
index d618d87d429ec..8889771e16e71 100644
--- a/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py
+++ b/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py
@@ -112,7 +112,7 @@ def on_celery_import_modules(*args, **kwargs):
"""
Preload some "expensive" airflow modules once, so other task processes won't have to import it again.
- Loading these for each task adds 0.3-0.5s *per task* before the task can run. For long running tasks this
+ Loading these for each task adds 0.3-0.5s *per task* before the task can run. For long-running tasks this
doesn't matter, but for short tasks this starts to be a noticeable impact.
"""
import jinja2.ext # noqa: F401
diff --git a/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py b/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py
index 47be19f5f6876..d266d025277d5 100644
--- a/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py
+++ b/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py
@@ -508,7 +508,7 @@ class CloudSqlProxyRunner(LoggingMixin):
:param project_id: Optional id of the Google Cloud project to connect to - it overwrites
default project id taken from the Google Cloud connection.
:param sql_proxy_version: Specific version of SQL proxy to download
- (for example 'v1.13'). By default latest version is downloaded.
+ (for example 'v1.13'). By default, latest version is downloaded.
:param sql_proxy_binary_path: If specified, then proxy will be
used from the path specified rather than dynamically generated. This means
that if the binary is not present in that path it will also be downloaded.
@@ -693,7 +693,7 @@ def stop_proxy(self) -> None:
self.log.info("Skipped removing proxy - it was not downloaded: %s", self.sql_proxy_path)
if os.path.isfile(self.credentials_path):
self.log.info("Removing generated credentials file %s", self.credentials_path)
- # Here file cannot be delete by concurrent task (each task has its own copy)
+ # Here file cannot be deleted by concurrent task (each task has its own copy)
os.remove(self.credentials_path)
def get_proxy_version(self) -> str | None:
diff --git a/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py b/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py
index 66a1e7ee0a6a7..61044e17b27ef 100644
--- a/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py
+++ b/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py
@@ -110,7 +110,7 @@ def create_batch_prediction_job(
:param project_id: Required. Project to run training in.
:param region: Required. Location to run training in.
:param job_display_name: Required. The user-defined name of the BatchPredictionJob. The name can be
- up to 128 characters long and can be consist of any UTF-8 characters.
+ up to 128 characters long and can consist of any UTF-8 characters.
:param model_name: Required. A fully-qualified model resource name or model ID.
:param instances_format: Required. The format in which instances are provided. Must be one of the
formats listed in `Model.supported_input_storage_formats`. Default is "jsonl" when using
@@ -267,7 +267,7 @@ def submit_batch_prediction_job(
:param project_id: Required. Project to run training in.
:param region: Required. Location to run training in.
:param job_display_name: Required. The user-defined name of the BatchPredictionJob. The name can be
- up to 128 characters long and can be consist of any UTF-8 characters.
+ up to 128 characters long and can consist of any UTF-8 characters.
:param model_name: Required. A fully-qualified model resource name or model ID.
:param instances_format: Required. The format in which instances are provided. Must be one of the
formats listed in `Model.supported_input_storage_formats`. Default is "jsonl" when using
diff --git a/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py b/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py
index f3b858c63a47f..9d78152114f8d 100644
--- a/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py
+++ b/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py
@@ -244,7 +244,7 @@ class CloudBatchListTasksOperator(GoogleCloudBaseOperator):
:param job_name: Required. The name of the job for which to list tasks.
:param gcp_conn_id: The connection ID used to connect to Google Cloud.
:param filter: The filter based on which to list the jobs. If left empty, all the jobs are listed.
- :param group_name: The name of the group that owns the task. By default it's `group0`.
+ :param group_name: The name of the group that owns the task. By default, it's `group0`.
:param limit: The number of tasks to list.
If left empty, all the tasks matching the filter will be returned.
:param impersonation_chain: Optional service account to impersonate using short-term
diff --git a/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py b/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py
index 3793b00905179..95f599189f29a 100644
--- a/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py
+++ b/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py
@@ -141,7 +141,7 @@ def execute(self, context: Context):
class CloudRunDeleteJobOperator(GoogleCloudBaseOperator):
"""
- Deletes a job and wait for the the operation to be completed. Pushes the deleted job to xcom.
+ Deletes a job and wait for the operation to be completed. Pushes the deleted job to xcom.
:param project_id: Required. The ID of the Google Cloud project that the service belongs to.
:param region: Required. The ID of the Google Cloud region that the service belongs to.
diff --git a/providers/google/src/airflow/providers/google/suite/hooks/calendar.py b/providers/google/src/airflow/providers/google/suite/hooks/calendar.py
index ddc0078bba17d..163f117d74d82 100644
--- a/providers/google/src/airflow/providers/google/suite/hooks/calendar.py
+++ b/providers/google/src/airflow/providers/google/suite/hooks/calendar.py
@@ -104,7 +104,7 @@ def get_events(
only the participant is returned.
:param max_results: Optional. Maximum number of events returned on one result page.
Incomplete pages can be detected by a non-empty ``nextPageToken`` field in the response.
- By default the value is 250 events. The page size can never be larger than 2500 events
+ By default, the value is 250 events. The page size can never be larger than 2500 events
:param order_by: Optional. Acceptable values are ``"startTime"`` or "updated"
:param private_extended_property: Optional. Extended properties constraint specified as
``propertyName=value``. Matches only private properties. This parameter might be repeated
diff --git a/providers/http/docs/changelog.rst b/providers/http/docs/changelog.rst
index cf18187667572..24f43e223feea 100644
--- a/providers/http/docs/changelog.rst
+++ b/providers/http/docs/changelog.rst
@@ -284,7 +284,7 @@ Bug Fixes
Misc
~~~~
-* ``add lowerbount to requests-toolbelt and replace requests_toolbelt with requests-toolbelt (#43020)``
+* ``add lower bound to requests-toolbelt and replace requests_toolbelt with requests-toolbelt (#43020)``
* ``Set lower bound to asgiref>=2.3.0 (#43001)``
@@ -583,7 +583,7 @@ Misc
Features
~~~~~~~~
-* ``Add non login-password auth support for SimpleHttpOpeator (#29206)``
+* ``Add non login-password auth support for SimpleHttpOperator (#29206)``
4.2.0
.....
diff --git a/providers/keycloak/docs/auth-manager/manage/permissions.rst b/providers/keycloak/docs/auth-manager/manage/permissions.rst
index 94d665ac63dac..fc1e1e252bf0a 100644
--- a/providers/keycloak/docs/auth-manager/manage/permissions.rst
+++ b/providers/keycloak/docs/auth-manager/manage/permissions.rst
@@ -86,7 +86,7 @@ This will create
More resources about permissions can be found in the official documentation of Keycloak:
-1- `Keylcloak Authorization Process `_
+1- `Keycloak Authorization Process `_
2- `Keycloak Permission Overview `_
diff --git a/providers/ssh/docs/changelog.rst b/providers/ssh/docs/changelog.rst
index 697e636f7a318..046bb47fca2e6 100644
--- a/providers/ssh/docs/changelog.rst
+++ b/providers/ssh/docs/changelog.rst
@@ -713,7 +713,7 @@ Bug Fixes
* ``Updated documentation for June 2021 provider release (#16294)``
* ``Add Connection Documentation to more Providers (#15408)``
* ``More documentation update for June providers release (#16405)``
- * ``Synchronizes updated changelog after buggfix release (#16464)``
+ * ``Synchronizes updated changelog after bugfix release (#16464)``
1.3.0
.....
diff --git a/task-sdk/src/airflow/sdk/api/client.py b/task-sdk/src/airflow/sdk/api/client.py
index 0a3de02894150..c70cec0e89577 100644
--- a/task-sdk/src/airflow/sdk/api/client.py
+++ b/task-sdk/src/airflow/sdk/api/client.py
@@ -807,7 +807,7 @@ def auth_flow(self, request: httpx.Request):
yield request
-# This exists as a aid for debugging or local running via the `dry_run` argument to Client. It doesn't make
+# This exists as an aid for debugging or local running via the `dry_run` argument to Client. It doesn't make
# sense for returning connections etc.
def noop_handler(request: httpx.Request) -> httpx.Response:
path = request.url.path