From b6323edde05168ae61044c7547e90b67e8120751 Mon Sep 17 00:00:00 2001
From: Didier Durand <2927957+didier-durand@users.noreply.github.com>
Date: Wed, 29 Oct 2025 23:26:51 +0100
Subject: [PATCH] =?UTF-8?q?[v3-1-test]=20[Doc]=C2=A0Fixing=20some=20typos?=
=?UTF-8?q?=20and=20spelling=20errors=20(#57225)?=
MIME-Version: 1.0
Content-Type: text/plain; charset=UTF-8
Content-Transfer-Encoding: 8bit
* Fixing some typos and spelling errors
* adding 'chatbot' to spelling_wordlist.txt
* adding 'chatbots' to spelling_wordlist.txt
* removed 'Chatbots' from spelling_wordlist.txt
* Committing new versions of somes files generated by prek
(cherry picked from commit a04136fa81f34c81214252bd80675598ff686027)
Co-authored-by: Didier Durand <2927957+didier-durand@users.noreply.github.com>
---
contributing-docs/testing/task_sdk_integration_tests.rst | 2 +-
dev/breeze/doc/images/output_ci-image.svg | 4 ++--
dev/breeze/doc/images/output_ci-image.txt | 2 +-
dev/breeze/doc/images/output_ci-image_export-mount-cache.svg | 2 +-
dev/breeze/doc/images/output_ci-image_export-mount-cache.txt | 2 +-
dev/breeze/doc/images/output_ci-image_import-mount-cache.svg | 2 +-
dev/breeze/doc/images/output_ci-image_import-mount-cache.txt | 2 +-
dev/breeze/src/airflow_breeze/commands/ci_image_commands.py | 4 ++--
docs/spelling_wordlist.txt | 2 ++
.../amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py | 2 +-
.../amazon/src/airflow/providers/amazon/aws/hooks/chime.py | 2 +-
.../airflow/providers/amazon/aws/operators/cloud_formation.py | 4 ++--
.../providers/amazon/aws/transfers/azure_blob_to_s3.py | 4 ++--
.../beam/src/airflow/providers/apache/beam/operators/beam.py | 2 +-
.../src/airflow/providers/celery/executors/celery_executor.py | 2 +-
.../providers/celery/executors/celery_executor_utils.py | 2 +-
.../src/airflow/providers/google/cloud/hooks/cloud_sql.py | 4 ++--
.../google/cloud/hooks/vertex_ai/batch_prediction_job.py | 4 ++--
.../airflow/providers/google/cloud/operators/cloud_batch.py | 2 +-
.../src/airflow/providers/google/cloud/operators/cloud_run.py | 2 +-
.../src/airflow/providers/google/suite/hooks/calendar.py | 2 +-
providers/http/docs/changelog.rst | 4 ++--
providers/keycloak/docs/auth-manager/manage/permissions.rst | 2 +-
providers/ssh/docs/changelog.rst | 2 +-
task-sdk/src/airflow/sdk/api/client.py | 2 +-
25 files changed, 33 insertions(+), 31 deletions(-)
diff --git a/contributing-docs/testing/task_sdk_integration_tests.rst b/contributing-docs/testing/task_sdk_integration_tests.rst
index 132b2822bafc6..f7efb2e6f403f 100644
--- a/contributing-docs/testing/task_sdk_integration_tests.rst
+++ b/contributing-docs/testing/task_sdk_integration_tests.rst
@@ -85,7 +85,7 @@ reproducibility:
breeze testing task-sdk-integration-tests task_sdk_tests/test_task_sdk_health.py
# Run with custom Docker image
- DOCKER_IMAGE=my-custom-airflow-imahge:latest breeze testing task-sdk-integration-tests
+ DOCKER_IMAGE=my-custom-airflow-image:latest breeze testing task-sdk-integration-tests
Running in Your Current Virtual Environment
...........................................
diff --git a/dev/breeze/doc/images/output_ci-image.svg b/dev/breeze/doc/images/output_ci-image.svg
index 0c614dfc94703..62ca31ecad645 100644
--- a/dev/breeze/doc/images/output_ci-image.svg
+++ b/dev/breeze/doc/images/output_ci-image.svg
@@ -118,8 +118,8 @@
│verify Verify CI image. │
│save Save CI image to a file. │
│load Load CI image from a file. │
-│export-mount-cache Export content of the the mount cache to a directory. │
-│import-mount-cache Export content of the the mount cache to a directory. │
+│export-mount-cache Export content of the mount cache to a directory. │
+│import-mount-cache Export content of the mount cache to a directory. │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
diff --git a/dev/breeze/doc/images/output_ci-image.txt b/dev/breeze/doc/images/output_ci-image.txt
index 25b4a0db49abc..fcc45c9593176 100644
--- a/dev/breeze/doc/images/output_ci-image.txt
+++ b/dev/breeze/doc/images/output_ci-image.txt
@@ -1 +1 @@
-de6e1ee2eb602569a8cd1a3c6c4cbafe
+7745b7d70e4a06b724ed92814d416670
diff --git a/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg b/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg
index c1a516dbdefd7..4c56e91807a2a 100644
--- a/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg
+++ b/dev/breeze/doc/images/output_ci-image_export-mount-cache.svg
@@ -101,7 +101,7 @@
Usage:breeze ci-image export-mount-cache[OPTIONS]
-Export content of the the mount cache to a directory.
+Export content of the mount cache to a directory.
╭─ Export flags ───────────────────────────────────────────────────────────────────────────────────────────────────────╮
│*--cache-filePath to the file where cache is going to be exported(FILE)[required]│
diff --git a/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt b/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt
index 6d1c1347207a2..32ccab745cc9c 100644
--- a/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt
+++ b/dev/breeze/doc/images/output_ci-image_export-mount-cache.txt
@@ -1 +1 @@
-824109d9d232ba19a9aa0781ba8fccd9
+ebc1ae7a092c1e541957fb1e79829249
diff --git a/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg b/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg
index 32eb4fdfebe6e..5b31140f32717 100644
--- a/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg
+++ b/dev/breeze/doc/images/output_ci-image_import-mount-cache.svg
@@ -101,7 +101,7 @@
Usage:breeze ci-image import-mount-cache[OPTIONS]
-Export content of the the mount cache to a directory.
+Export content of the mount cache to a directory.
╭─ Import flags ───────────────────────────────────────────────────────────────────────────────────────────────────────╮
│*--cache-filePath to the file where cache is stored(FILE)[required]│
diff --git a/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt b/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt
index 00f5024812d07..7549b6730abea 100644
--- a/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt
+++ b/dev/breeze/doc/images/output_ci-image_import-mount-cache.txt
@@ -1 +1 @@
-65be4537eea8d56d3ddc34d4ccab60c7
+669f8d8b36a1d9dde983862db6a69ce4
diff --git a/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py b/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py
index 184562da62862..71655db360b67 100644
--- a/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py
+++ b/dev/breeze/src/airflow_breeze/commands/ci_image_commands.py
@@ -930,7 +930,7 @@ def export_mount_cache(
cache_file: Path,
):
"""
- Export content of the the mount cache to a directory.
+ Export content of the mount cache to a directory.
"""
perform_environment_checks()
make_sure_builder_configured(params=BuildCiParams(builder=builder))
@@ -995,7 +995,7 @@ def import_mount_cache(
cache_file: Path,
):
"""
- Export content of the the mount cache to a directory.
+ Export content of the mount cache to a directory.
"""
perform_environment_checks()
make_sure_builder_configured(params=BuildCiParams(builder=builder))
diff --git a/docs/spelling_wordlist.txt b/docs/spelling_wordlist.txt
index a4dbbdf3202aa..5d673b7b86029 100644
--- a/docs/spelling_wordlist.txt
+++ b/docs/spelling_wordlist.txt
@@ -253,6 +253,8 @@ chakra
Changelog
changelog
charset
+chatbot
+chatbots
Checklicence
checklicence
CheckOperator
diff --git a/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py b/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py
index 979cb4bc345b9..7097807b78af9 100644
--- a/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py
+++ b/providers/amazon/src/airflow/providers/amazon/aws/hooks/base_aws.py
@@ -790,7 +790,7 @@ def async_conn(self):
async def get_async_conn(self):
"""Get an aiobotocore client to use for async operations."""
# We have to wrap the call `self.get_client_type` in another call `_get_async_conn`,
- # because one of it's arguments `self.region_name` is a `@property` decorated function
+ # because one of its arguments `self.region_name` is a `@property` decorated function
# calling the cached property `self.conn_config` at the end.
return await sync_to_async(self._get_async_conn)()
diff --git a/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py b/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py
index 8201f32e76f01..bdf0f220a18f3 100644
--- a/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py
+++ b/providers/amazon/src/airflow/providers/amazon/aws/hooks/chime.py
@@ -33,7 +33,7 @@ class ChimeWebhookHook(HttpHook):
"""
Interact with Amazon Chime Webhooks to create notifications.
- .. warning:: This hook is only designed to work with web hooks and not chat bots.
+ .. warning:: This hook is only designed to work with web hooks and not chatbots.
:param chime_conn_id: :ref:`Amazon Chime Connection ID `
with Endpoint as `https://hooks.chime.aws` and the webhook token
diff --git a/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py b/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py
index e4f3d211bea93..d4bfcc9532230 100644
--- a/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py
+++ b/providers/amazon/src/airflow/providers/amazon/aws/operators/cloud_formation.py
@@ -32,7 +32,7 @@
class CloudFormationCreateStackOperator(AwsBaseOperator[CloudFormationHook]):
"""
- An operator that creates a AWS CloudFormation stack.
+ An operator that creates an AWS CloudFormation stack.
.. seealso::
For more information on how to use this operator, take a look at the guide:
@@ -68,7 +68,7 @@ def execute(self, context: Context):
class CloudFormationDeleteStackOperator(AwsBaseOperator[CloudFormationHook]):
"""
- An operator that deletes a AWS CloudFormation stack.
+ An operator that deletes an AWS CloudFormation stack.
.. seealso::
For more information on how to use this operator, take a look at the guide:
diff --git a/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py b/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py
index e58d43393c85a..9328379703438 100644
--- a/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py
+++ b/providers/amazon/src/airflow/providers/amazon/aws/transfers/azure_blob_to_s3.py
@@ -49,12 +49,12 @@ class AzureBlobStorageToS3Operator(BaseOperator):
:param prefix: Prefix string which filters objects whose name begin with
this prefix. (templated)
:param delimiter: The delimiter by which you want to filter the objects. (templated)
- For e.g to lists the CSV files from in a directory in GCS you would use
+ For e.g. to lists the CSV files from in a directory in GCS you would use
delimiter='.csv'.
:param aws_conn_id: Connection id of the S3 connection to use
:param dest_s3_key: The base S3 key to be used to store the files. (templated)
:param dest_verify: Whether or not to verify SSL certificates for S3 connection.
- By default SSL certificates are verified.
+ By default, SSL certificates are verified.
You can provide the following values:
- ``False``: do not validate SSL certificates. SSL will still be used
diff --git a/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py b/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py
index 3ac04cd4c57aa..389013866950b 100644
--- a/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py
+++ b/providers/apache/beam/src/airflow/providers/apache/beam/operators/beam.py
@@ -161,7 +161,7 @@ class BeamBasePipelineOperator(BaseOperator, BeamDataflowMixin, ABC):
"""
Abstract base class for Beam Pipeline Operators.
- :param runner: Runner on which pipeline will be run. By default "DirectRunner" is being used.
+ :param runner: Runner on which pipeline will be run. By default, "DirectRunner" is being used.
Other possible options: DataflowRunner, SparkRunner, FlinkRunner, PortableRunner.
See: :class:`~providers.apache.beam.hooks.beam.BeamRunnerType`
See: https://beam.apache.org/documentation/runners/capability-matrix/
diff --git a/providers/celery/src/airflow/providers/celery/executors/celery_executor.py b/providers/celery/src/airflow/providers/celery/executors/celery_executor.py
index 95c4c97f12ce6..d48eb1464c44d 100644
--- a/providers/celery/src/airflow/providers/celery/executors/celery_executor.py
+++ b/providers/celery/src/airflow/providers/celery/executors/celery_executor.py
@@ -344,7 +344,7 @@ def _process_workloads(self, workloads: Sequence[workloads.All]) -> None:
def _send_tasks(self, task_tuples_to_send: Sequence[TaskInstanceInCelery]):
first_task = next(t[-1] for t in task_tuples_to_send)
- # Celery state queries will stuck if we do not use one same backend
+ # Celery state queries will be stuck if we do not use one same backend
# for all tasks.
cached_celery_backend = first_task.backend
diff --git a/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py b/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py
index 76202dd139f33..fe9323fedbfd9 100644
--- a/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py
+++ b/providers/celery/src/airflow/providers/celery/executors/celery_executor_utils.py
@@ -122,7 +122,7 @@ def on_celery_import_modules(*args, **kwargs):
"""
Preload some "expensive" airflow modules once, so other task processes won't have to import it again.
- Loading these for each task adds 0.3-0.5s *per task* before the task can run. For long running tasks this
+ Loading these for each task adds 0.3-0.5s *per task* before the task can run. For long-running tasks this
doesn't matter, but for short tasks this starts to be a noticeable impact.
"""
import jinja2.ext # noqa: F401
diff --git a/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py b/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py
index 32b86cd9c0203..409bb2cce30ba 100644
--- a/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py
+++ b/providers/google/src/airflow/providers/google/cloud/hooks/cloud_sql.py
@@ -502,7 +502,7 @@ class CloudSqlProxyRunner(LoggingMixin):
:param project_id: Optional id of the Google Cloud project to connect to - it overwrites
default project id taken from the Google Cloud connection.
:param sql_proxy_version: Specific version of SQL proxy to download
- (for example 'v1.13'). By default latest version is downloaded.
+ (for example 'v1.13'). By default, latest version is downloaded.
:param sql_proxy_binary_path: If specified, then proxy will be
used from the path specified rather than dynamically generated. This means
that if the binary is not present in that path it will also be downloaded.
@@ -687,7 +687,7 @@ def stop_proxy(self) -> None:
self.log.info("Skipped removing proxy - it was not downloaded: %s", self.sql_proxy_path)
if os.path.isfile(self.credentials_path):
self.log.info("Removing generated credentials file %s", self.credentials_path)
- # Here file cannot be delete by concurrent task (each task has its own copy)
+ # Here file cannot be deleted by concurrent task (each task has its own copy)
os.remove(self.credentials_path)
def get_proxy_version(self) -> str | None:
diff --git a/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py b/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py
index 66a1e7ee0a6a7..61044e17b27ef 100644
--- a/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py
+++ b/providers/google/src/airflow/providers/google/cloud/hooks/vertex_ai/batch_prediction_job.py
@@ -110,7 +110,7 @@ def create_batch_prediction_job(
:param project_id: Required. Project to run training in.
:param region: Required. Location to run training in.
:param job_display_name: Required. The user-defined name of the BatchPredictionJob. The name can be
- up to 128 characters long and can be consist of any UTF-8 characters.
+ up to 128 characters long and can consist of any UTF-8 characters.
:param model_name: Required. A fully-qualified model resource name or model ID.
:param instances_format: Required. The format in which instances are provided. Must be one of the
formats listed in `Model.supported_input_storage_formats`. Default is "jsonl" when using
@@ -267,7 +267,7 @@ def submit_batch_prediction_job(
:param project_id: Required. Project to run training in.
:param region: Required. Location to run training in.
:param job_display_name: Required. The user-defined name of the BatchPredictionJob. The name can be
- up to 128 characters long and can be consist of any UTF-8 characters.
+ up to 128 characters long and can consist of any UTF-8 characters.
:param model_name: Required. A fully-qualified model resource name or model ID.
:param instances_format: Required. The format in which instances are provided. Must be one of the
formats listed in `Model.supported_input_storage_formats`. Default is "jsonl" when using
diff --git a/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py b/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py
index bc6da9ab502ec..e1cb8d9b3d763 100644
--- a/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py
+++ b/providers/google/src/airflow/providers/google/cloud/operators/cloud_batch.py
@@ -244,7 +244,7 @@ class CloudBatchListTasksOperator(GoogleCloudBaseOperator):
:param job_name: Required. The name of the job for which to list tasks.
:param gcp_conn_id: The connection ID used to connect to Google Cloud.
:param filter: The filter based on which to list the jobs. If left empty, all the jobs are listed.
- :param group_name: The name of the group that owns the task. By default it's `group0`.
+ :param group_name: The name of the group that owns the task. By default, it's `group0`.
:param limit: The number of tasks to list.
If left empty, all the tasks matching the filter will be returned.
:param impersonation_chain: Optional service account to impersonate using short-term
diff --git a/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py b/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py
index b0da1debc3667..4542e643f750d 100644
--- a/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py
+++ b/providers/google/src/airflow/providers/google/cloud/operators/cloud_run.py
@@ -141,7 +141,7 @@ def execute(self, context: Context):
class CloudRunDeleteJobOperator(GoogleCloudBaseOperator):
"""
- Deletes a job and wait for the the operation to be completed. Pushes the deleted job to xcom.
+ Deletes a job and wait for the operation to be completed. Pushes the deleted job to xcom.
:param project_id: Required. The ID of the Google Cloud project that the service belongs to.
:param region: Required. The ID of the Google Cloud region that the service belongs to.
diff --git a/providers/google/src/airflow/providers/google/suite/hooks/calendar.py b/providers/google/src/airflow/providers/google/suite/hooks/calendar.py
index ddc0078bba17d..163f117d74d82 100644
--- a/providers/google/src/airflow/providers/google/suite/hooks/calendar.py
+++ b/providers/google/src/airflow/providers/google/suite/hooks/calendar.py
@@ -104,7 +104,7 @@ def get_events(
only the participant is returned.
:param max_results: Optional. Maximum number of events returned on one result page.
Incomplete pages can be detected by a non-empty ``nextPageToken`` field in the response.
- By default the value is 250 events. The page size can never be larger than 2500 events
+ By default, the value is 250 events. The page size can never be larger than 2500 events
:param order_by: Optional. Acceptable values are ``"startTime"`` or "updated"
:param private_extended_property: Optional. Extended properties constraint specified as
``propertyName=value``. Matches only private properties. This parameter might be repeated
diff --git a/providers/http/docs/changelog.rst b/providers/http/docs/changelog.rst
index 2dbe2cc30ab96..a59310198289f 100644
--- a/providers/http/docs/changelog.rst
+++ b/providers/http/docs/changelog.rst
@@ -270,7 +270,7 @@ Bug Fixes
Misc
~~~~
-* ``add lowerbount to requests-toolbelt and replace requests_toolbelt with requests-toolbelt (#43020)``
+* ``add lower bound to requests-toolbelt and replace requests_toolbelt with requests-toolbelt (#43020)``
* ``Set lower bound to asgiref>=2.3.0 (#43001)``
@@ -569,7 +569,7 @@ Misc
Features
~~~~~~~~
-* ``Add non login-password auth support for SimpleHttpOpeator (#29206)``
+* ``Add non login-password auth support for SimpleHttpOperator (#29206)``
4.2.0
.....
diff --git a/providers/keycloak/docs/auth-manager/manage/permissions.rst b/providers/keycloak/docs/auth-manager/manage/permissions.rst
index 94d665ac63dac..fc1e1e252bf0a 100644
--- a/providers/keycloak/docs/auth-manager/manage/permissions.rst
+++ b/providers/keycloak/docs/auth-manager/manage/permissions.rst
@@ -86,7 +86,7 @@ This will create
More resources about permissions can be found in the official documentation of Keycloak:
-1- `Keylcloak Authorization Process `_
+1- `Keycloak Authorization Process `_
2- `Keycloak Permission Overview `_
diff --git a/providers/ssh/docs/changelog.rst b/providers/ssh/docs/changelog.rst
index 358835f82e0cc..b722b19e47b92 100644
--- a/providers/ssh/docs/changelog.rst
+++ b/providers/ssh/docs/changelog.rst
@@ -674,7 +674,7 @@ Bug Fixes
* ``Updated documentation for June 2021 provider release (#16294)``
* ``Add Connection Documentation to more Providers (#15408)``
* ``More documentation update for June providers release (#16405)``
- * ``Synchronizes updated changelog after buggfix release (#16464)``
+ * ``Synchronizes updated changelog after bugfix release (#16464)``
1.3.0
.....
diff --git a/task-sdk/src/airflow/sdk/api/client.py b/task-sdk/src/airflow/sdk/api/client.py
index 6945c2dfbcf96..bc735f684747e 100644
--- a/task-sdk/src/airflow/sdk/api/client.py
+++ b/task-sdk/src/airflow/sdk/api/client.py
@@ -778,7 +778,7 @@ def auth_flow(self, request: httpx.Request):
yield request
-# This exists as a aid for debugging or local running via the `dry_run` argument to Client. It doesn't make
+# This exists as an aid for debugging or local running via the `dry_run` argument to Client. It doesn't make
# sense for returning connections etc.
def noop_handler(request: httpx.Request) -> httpx.Response:
path = request.url.path