diff --git a/tests/system/providers/google/ads/example_ads.py b/tests/system/providers/google/ads/example_ads.py index 239894310195e..5dd336e7df449 100644 --- a/tests/system/providers/google/ads/example_ads.py +++ b/tests/system/providers/google/ads/example_ads.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.ads.operators.ads import GoogleAdsListAccountsOperator from airflow.providers.google.ads.transfers.ads_to_gcs import GoogleAdsToGcsOperator from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator @@ -72,7 +72,7 @@ ] # [END howto_google_ads_env_variables] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_dataset.py b/tests/system/providers/google/cloud/automl/example_automl_dataset.py index 1d7da91c8f6f1..1b0c687a2d56c 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_dataset.py +++ b/tests/system/providers/google/cloud/automl/example_automl_dataset.py @@ -25,7 +25,7 @@ from copy import deepcopy from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.automl import ( AutoMLCreateDatasetOperator, @@ -72,7 +72,7 @@ def get_target_column_spec(columns_specs: list[dict], column_name: str) -> str: raise Exception(f"Unknown target column: {column_name}") -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_model.py b/tests/system/providers/google/cloud/automl/example_automl_model.py index 274dd83e95f76..fbcbcb1898cb2 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_model.py +++ b/tests/system/providers/google/cloud/automl/example_automl_model.py @@ -27,7 +27,7 @@ from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.automl import ( AutoMLBatchPredictOperator, @@ -110,7 +110,7 @@ def get_target_column_spec(columns_specs: list[dict], column_name: str) -> str: raise Exception(f"Unknown target column: {column_name}") -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_nl_text_classification.py b/tests/system/providers/google/cloud/automl/example_automl_nl_text_classification.py index 753c91dfd02ec..e43463063cf8e 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_nl_text_classification.py +++ b/tests/system/providers/google/cloud/automl/example_automl_nl_text_classification.py @@ -27,7 +27,7 @@ from google.cloud.aiplatform import schema from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.gcs import ( @@ -75,7 +75,7 @@ extract_object_id = CloudAutoMLHook.extract_object_id # Example DAG for AutoML Natural Language Text Classification -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_nl_text_extraction.py b/tests/system/providers/google/cloud/automl/example_automl_nl_text_extraction.py index 06b22779a85b0..3575eb6b38a7f 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_nl_text_extraction.py +++ b/tests/system/providers/google/cloud/automl/example_automl_nl_text_extraction.py @@ -27,7 +27,7 @@ from google.cloud.aiplatform import schema from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.gcs import ( @@ -74,7 +74,7 @@ extract_object_id = CloudAutoMLHook.extract_object_id # Example DAG for AutoML Natural Language Entities Extraction -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_nl_text_sentiment.py b/tests/system/providers/google/cloud/automl/example_automl_nl_text_sentiment.py index 1529f07bc678c..f4e16c39088f9 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_nl_text_sentiment.py +++ b/tests/system/providers/google/cloud/automl/example_automl_nl_text_sentiment.py @@ -27,7 +27,7 @@ from google.cloud.aiplatform import schema from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.gcs import ( @@ -75,7 +75,7 @@ extract_object_id = CloudAutoMLHook.extract_object_id # Example DAG for AutoML Natural Language Text Sentiment -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_translation.py b/tests/system/providers/google/cloud/automl/example_automl_translation.py index e03453f03a27e..2814da0cfe8f1 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_translation.py +++ b/tests/system/providers/google/cloud/automl/example_automl_translation.py @@ -26,8 +26,8 @@ from google.cloud import storage -from airflow import models from airflow.decorators import task +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.automl import ( @@ -74,7 +74,7 @@ # Example DAG for AutoML Translation -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_video_intelligence_classification.py b/tests/system/providers/google/cloud/automl/example_automl_video_intelligence_classification.py index 7993494ffffd9..b88dcebe108fd 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_video_intelligence_classification.py +++ b/tests/system/providers/google/cloud/automl/example_automl_video_intelligence_classification.py @@ -24,7 +24,7 @@ from datetime import datetime from typing import cast -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.automl import ( @@ -68,7 +68,7 @@ # Example DAG for AutoML Video Intelligence Classification -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_video_intelligence_tracking.py b/tests/system/providers/google/cloud/automl/example_automl_video_intelligence_tracking.py index 61802077b3230..be0d07c7c45f9 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_video_intelligence_tracking.py +++ b/tests/system/providers/google/cloud/automl/example_automl_video_intelligence_tracking.py @@ -24,7 +24,7 @@ from datetime import datetime from typing import cast -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.automl import ( @@ -68,7 +68,7 @@ # Example DAG for AutoML Video Intelligence Object Tracking -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_vision_classification.py b/tests/system/providers/google/cloud/automl/example_automl_vision_classification.py index a85078779b3ef..9ff541893d51f 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_vision_classification.py +++ b/tests/system/providers/google/cloud/automl/example_automl_vision_classification.py @@ -26,8 +26,8 @@ from google.cloud import storage -from airflow import models from airflow.decorators import task +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.automl import ( @@ -73,7 +73,7 @@ extract_object_id = CloudAutoMLHook.extract_object_id # Example DAG for AutoML Vision Classification -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/automl/example_automl_vision_object_detection.py b/tests/system/providers/google/cloud/automl/example_automl_vision_object_detection.py index 894b467dd372b..e9a49c4b2403b 100644 --- a/tests/system/providers/google/cloud/automl/example_automl_vision_object_detection.py +++ b/tests/system/providers/google/cloud/automl/example_automl_vision_object_detection.py @@ -24,7 +24,7 @@ from datetime import datetime from typing import cast -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.hooks.automl import CloudAutoMLHook from airflow.providers.google.cloud.operators.automl import ( @@ -67,7 +67,7 @@ # Example DAG for AutoML Vision Object Detection -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/azure/example_azure_blob_to_gcs.py b/tests/system/providers/google/cloud/azure/example_azure_blob_to_gcs.py index 5d7649479df60..0e07256fdbf94 100644 --- a/tests/system/providers/google/cloud/azure/example_azure_blob_to_gcs.py +++ b/tests/system/providers/google/cloud/azure/example_azure_blob_to_gcs.py @@ -20,7 +20,7 @@ import os from datetime import datetime -from airflow import DAG +from airflow.models.dag import DAG from airflow.providers.google.cloud.transfers.azure_blob_to_gcs import AzureBlobStorageToGCSOperator from airflow.providers.microsoft.azure.sensors.wasb import ( WasbBlobSensor, diff --git a/tests/system/providers/google/cloud/azure/example_azure_fileshare_to_gcs.py b/tests/system/providers/google/cloud/azure/example_azure_fileshare_to_gcs.py index b61011f1e502b..c46287db08897 100644 --- a/tests/system/providers/google/cloud/azure/example_azure_fileshare_to_gcs.py +++ b/tests/system/providers/google/cloud/azure/example_azure_fileshare_to_gcs.py @@ -19,7 +19,7 @@ import os from datetime import datetime, timedelta -from airflow import DAG +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.azure_fileshare_to_gcs import AzureFileShareToGCSOperator from airflow.utils.trigger_rule import TriggerRule diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_dataset.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_dataset.py index fadc85d9086ac..2f3357cafaf64 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_dataset.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_dataset.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, @@ -39,7 +39,7 @@ DATASET_NAME = f"dataset_{DAG_ID}_{ENV_ID}" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_dts.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_dts.py index b6a31b19eb83b..e5b3ae389433a 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_dts.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_dts.py @@ -26,8 +26,8 @@ from pathlib import Path from typing import cast -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, @@ -81,7 +81,7 @@ # [END howto_bigquery_dts_create_args] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_operations.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_operations.py index 8a5f101fc3773..b4e6306170bb0 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_operations.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_operations.py @@ -24,7 +24,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateExternalTableOperator, @@ -43,7 +43,7 @@ CSV_FILE_LOCAL_PATH = str(Path(__file__).parent / "resources" / "us-states.csv") -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_operations_location.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_operations_location.py index 04b8f3187d3f1..7c7b72643531f 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_operations_location.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_operations_location.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -38,7 +38,7 @@ DATASET_NAME = f"dataset_{DAG_ID}_{ENV_ID}" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_queries.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_queries.py index 3ce1bc2801a77..7898088046651 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_queries.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_queries.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCheckOperator, @@ -68,7 +68,7 @@ ) # [END howto_operator_bigquery_query] - with models.DAG( + with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_queries_async.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_queries_async.py index cdedfb0bc2cdf..4ec675de38b98 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_queries_async.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_queries_async.py @@ -23,7 +23,7 @@ import os from datetime import datetime, timedelta -from airflow import DAG +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCheckOperator, diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_sensors.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_sensors.py index 4f8d695787e20..d6a466a26f8c2 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_sensors.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_sensors.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -56,7 +56,7 @@ ] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_tables.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_tables.py index c5ba17f9edfef..83815c3fd2da5 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_tables.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_tables.py @@ -25,7 +25,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -52,7 +52,7 @@ GCS_PATH_TO_SCHEMA_JSON = f"gs://{BUCKET_NAME}/{SCHEMA_JSON_DESTINATION}" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_bigquery.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_bigquery.py index 9770412de078a..716c89cd51948 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_bigquery.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_bigquery.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -41,7 +41,7 @@ LOCATION = "US" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_gcs.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_gcs.py index 7fe6ad0f6dfe8..bc86701a4cf22 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_gcs.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_gcs.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -43,7 +43,7 @@ TABLE = "test" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_gcs_async.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_gcs_async.py index cccec95831f5d..e8a9f942d4df4 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_gcs_async.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_gcs_async.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -43,7 +43,7 @@ TABLE = "test" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_mssql.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_mssql.py index b91e5881ac93b..c02c7a36cbc19 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_mssql.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_mssql.py @@ -25,7 +25,7 @@ import pytest -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -46,7 +46,7 @@ TABLE = "table_42" destination_table = "mssql_table_test" -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_mysql.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_mysql.py index 11b898a569d1d..c5e7fd2e218ae 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_mysql.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_mysql.py @@ -25,7 +25,7 @@ import pytest -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -45,7 +45,7 @@ TABLE = "table_42" destination_table = "mysql_table_test" -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_postgres.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_postgres.py index 80aa19f99cf99..c0d365091a178 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_to_postgres.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_to_postgres.py @@ -25,7 +25,7 @@ import pytest -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -45,7 +45,7 @@ TABLE = "table_42" destination_table = "postgres_table_test" -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigquery/example_bigquery_transfer.py b/tests/system/providers/google/cloud/bigquery/example_bigquery_transfer.py index adc12131a7ff7..acbc312189375 100644 --- a/tests/system/providers/google/cloud/bigquery/example_bigquery_transfer.py +++ b/tests/system/providers/google/cloud/bigquery/example_bigquery_transfer.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -44,7 +44,7 @@ ORIGIN = "origin" TARGET = "target" -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/bigtable/example_bigtable.py b/tests/system/providers/google/cloud/bigtable/example_bigtable.py index 4f06c9c2fa724..16b63ee2a7e19 100644 --- a/tests/system/providers/google/cloud/bigtable/example_bigtable.py +++ b/tests/system/providers/google/cloud/bigtable/example_bigtable.py @@ -46,8 +46,8 @@ import os from datetime import datetime -from airflow import models from airflow.decorators import task_group +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigtable import ( BigtableCreateInstanceOperator, BigtableCreateTableOperator, @@ -80,7 +80,7 @@ CBT_POKE_INTERVAL = 60 -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/cloud_batch/example_cloud_batch.py b/tests/system/providers/google/cloud/cloud_batch/example_cloud_batch.py index d3f3d752a8a1e..5c5d0ac68947a 100644 --- a/tests/system/providers/google/cloud/cloud_batch/example_cloud_batch.py +++ b/tests/system/providers/google/cloud/cloud_batch/example_cloud_batch.py @@ -25,7 +25,7 @@ from google.cloud import batch_v1 -from airflow import models +from airflow.models.dag import DAG from airflow.operators.python import PythonOperator from airflow.providers.google.cloud.operators.cloud_batch import ( CloudBatchDeleteJobOperator, @@ -115,7 +115,7 @@ def _create_job(): # [END howto_operator_batch_job_creation] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/cloud_build/example_cloud_build.py b/tests/system/providers/google/cloud/cloud_build/example_cloud_build.py index 8b0eed2c4c5ad..47ce74a7afdee 100644 --- a/tests/system/providers/google/cloud/cloud_build/example_cloud_build.py +++ b/tests/system/providers/google/cloud/cloud_build/example_cloud_build.py @@ -27,8 +27,8 @@ import yaml -from airflow import models from airflow.decorators import task_group +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.cloud_build import ( @@ -69,7 +69,7 @@ # [END howto_operator_create_build_from_repo_body] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/cloud_build/example_cloud_build_trigger.py b/tests/system/providers/google/cloud/cloud_build/example_cloud_build_trigger.py index 7828c33129803..7e35466a68bbd 100644 --- a/tests/system/providers/google/cloud/cloud_build/example_cloud_build_trigger.py +++ b/tests/system/providers/google/cloud/cloud_build/example_cloud_build_trigger.py @@ -24,8 +24,8 @@ from datetime import datetime from typing import Any, cast -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.operators.cloud_build import ( CloudBuildCreateBuildTriggerOperator, @@ -83,7 +83,7 @@ # [END howto_operator_create_build_from_repo_body] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/cloud_functions/example_functions.py b/tests/system/providers/google/cloud/cloud_functions/example_functions.py index c4041666f43bd..c990717ef21de 100644 --- a/tests/system/providers/google/cloud/cloud_functions/example_functions.py +++ b/tests/system/providers/google/cloud/cloud_functions/example_functions.py @@ -27,8 +27,8 @@ from datetime import datetime from typing import Any -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.functions import ( CloudFunctionDeleteFunctionOperator, CloudFunctionDeployFunctionOperator, @@ -76,7 +76,7 @@ # [END howto_operator_gcf_deploy_variants] -with models.DAG( +with DAG( DAG_ID, default_args=default_args, start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/cloud_memorystore/example_cloud_memorystore_memcached.py b/tests/system/providers/google/cloud/cloud_memorystore/example_cloud_memorystore_memcached.py index 1ee65ede8f710..9e12caf1c431d 100644 --- a/tests/system/providers/google/cloud/cloud_memorystore/example_cloud_memorystore_memcached.py +++ b/tests/system/providers/google/cloud/cloud_memorystore/example_cloud_memorystore_memcached.py @@ -25,7 +25,7 @@ from google.protobuf.field_mask_pb2 import FieldMask -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.cloud_memorystore import ( CloudMemorystoreMemcachedApplyParametersOperator, CloudMemorystoreMemcachedCreateInstanceOperator, @@ -55,7 +55,7 @@ # [END howto_operator_memcached_instance] -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/cloud_memorystore/example_cloud_memorystore_redis.py b/tests/system/providers/google/cloud/cloud_memorystore/example_cloud_memorystore_redis.py index 0aa2800473b4c..3b6c1f46415d7 100644 --- a/tests/system/providers/google/cloud/cloud_memorystore/example_cloud_memorystore_redis.py +++ b/tests/system/providers/google/cloud/cloud_memorystore/example_cloud_memorystore_redis.py @@ -25,7 +25,7 @@ from google.cloud.redis_v1 import FailoverInstanceRequest, Instance -from airflow import models +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.cloud_memorystore import ( CloudMemorystoreCreateInstanceAndImportOperator, @@ -68,7 +68,7 @@ SECOND_INSTANCE = {"tier": Instance.Tier.STANDARD_HA, "memory_size_gb": 3} -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/cloud_run/example_cloud_run.py b/tests/system/providers/google/cloud/cloud_run/example_cloud_run.py index e5f83578cd59d..279db3714f26d 100644 --- a/tests/system/providers/google/cloud/cloud_run/example_cloud_run.py +++ b/tests/system/providers/google/cloud/cloud_run/example_cloud_run.py @@ -26,7 +26,7 @@ from google.cloud.run_v2 import Job from google.cloud.run_v2.types import k8s_min -from airflow import models +from airflow.models.dag import DAG from airflow.operators.python import PythonOperator from airflow.providers.google.cloud.operators.cloud_run import ( CloudRunCreateJobOperator, @@ -128,7 +128,7 @@ def _create_job_with_label(): return job -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql.py b/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql.py index f5ad5066fc253..b11208c022fdd 100644 --- a/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql.py +++ b/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql.py @@ -26,7 +26,7 @@ from datetime import datetime from urllib.parse import urlsplit -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.operators.cloud_sql import ( CloudSQLCloneInstanceOperator, @@ -146,7 +146,7 @@ db_patch_body = {"charset": "utf16", "collation": "utf16_general_ci"} # [END howto_operator_cloudsql_db_patch_body] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql_query_mysql.py b/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql_query_mysql.py index 8ac3608286d1d..f705a93eac2b5 100644 --- a/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql_query_mysql.py +++ b/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql_query_mysql.py @@ -29,9 +29,10 @@ from googleapiclient import discovery -from airflow import models, settings +from airflow import settings from airflow.decorators import task, task_group -from airflow.models import Connection +from airflow.models.connection import Connection +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.cloud_sql import ( CloudSQLCreateInstanceDatabaseOperator, @@ -162,7 +163,7 @@ log = logging.getLogger(__name__) -with models.DAG( +with DAG( dag_id=DAG_ID, start_date=datetime(2021, 1, 1), catchup=False, diff --git a/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql_query_postgres.py b/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql_query_postgres.py index 1de7387f6f0f1..b826302bc29c0 100644 --- a/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql_query_postgres.py +++ b/tests/system/providers/google/cloud/cloud_sql/example_cloud_sql_query_postgres.py @@ -29,9 +29,10 @@ from googleapiclient import discovery -from airflow import models, settings +from airflow import settings from airflow.decorators import task, task_group -from airflow.models import Connection +from airflow.models.connection import Connection +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.cloud_sql import ( CloudSQLCreateInstanceDatabaseOperator, @@ -167,7 +168,7 @@ log = logging.getLogger(__name__) -with models.DAG( +with DAG( dag_id=DAG_ID, start_date=datetime(2021, 1, 1), catchup=False, diff --git a/tests/system/providers/google/cloud/composer/example_cloud_composer.py b/tests/system/providers/google/cloud/composer/example_cloud_composer.py index a6de4bc26be55..e2dd19179b21c 100644 --- a/tests/system/providers/google/cloud/composer/example_cloud_composer.py +++ b/tests/system/providers/google/cloud/composer/example_cloud_composer.py @@ -20,8 +20,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.cloud_composer import ( CloudComposerCreateEnvironmentOperator, CloudComposerDeleteEnvironmentOperator, @@ -60,7 +60,7 @@ # [END howto_operator_composer_update_environment] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/composer/example_cloud_composer_deferrable.py b/tests/system/providers/google/cloud/composer/example_cloud_composer_deferrable.py index 7ccafe98e997e..7a2ebc43c51ed 100644 --- a/tests/system/providers/google/cloud/composer/example_cloud_composer_deferrable.py +++ b/tests/system/providers/google/cloud/composer/example_cloud_composer_deferrable.py @@ -20,8 +20,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.cloud_composer import ( CloudComposerCreateEnvironmentOperator, CloudComposerDeleteEnvironmentOperator, @@ -56,7 +56,7 @@ # [END howto_operator_composer_update_environment] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/compute/example_compute.py b/tests/system/providers/google/cloud/compute/example_compute.py index 13e25ae794848..eebc6f71e03b1 100644 --- a/tests/system/providers/google/cloud/compute/example_compute.py +++ b/tests/system/providers/google/cloud/compute/example_compute.py @@ -27,8 +27,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.compute import ( ComputeEngineDeleteInstanceOperator, ComputeEngineDeleteInstanceTemplateOperator, @@ -98,7 +98,7 @@ } # [END howto_operator_gce_args_common] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/compute/example_compute_igm.py b/tests/system/providers/google/cloud/compute/example_compute_igm.py index 45449cf12574f..68f4bd37f434c 100644 --- a/tests/system/providers/google/cloud/compute/example_compute_igm.py +++ b/tests/system/providers/google/cloud/compute/example_compute_igm.py @@ -27,8 +27,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.compute import ( ComputeEngineCopyInstanceTemplateOperator, ComputeEngineDeleteInstanceGroupManagerOperator, @@ -108,7 +108,7 @@ # [END howto_operator_compute_igm_update_template_args] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/compute/example_compute_ssh.py b/tests/system/providers/google/cloud/compute/example_compute_ssh.py index 532cabc374335..26f6e89ff17ba 100644 --- a/tests/system/providers/google/cloud/compute/example_compute_ssh.py +++ b/tests/system/providers/google/cloud/compute/example_compute_ssh.py @@ -26,8 +26,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.hooks.compute_ssh import ComputeEngineSSHHook from airflow.providers.google.cloud.operators.compute import ( ComputeEngineDeleteInstanceOperator, @@ -69,7 +69,7 @@ } # [END howto_operator_gce_args_common] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/compute/example_compute_ssh_os_login.py b/tests/system/providers/google/cloud/compute/example_compute_ssh_os_login.py index 6677550e81e13..88cdc8bcca082 100644 --- a/tests/system/providers/google/cloud/compute/example_compute_ssh_os_login.py +++ b/tests/system/providers/google/cloud/compute/example_compute_ssh_os_login.py @@ -26,8 +26,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.hooks.compute_ssh import ComputeEngineSSHHook from airflow.providers.google.cloud.operators.compute import ( ComputeEngineDeleteInstanceOperator, @@ -77,7 +77,7 @@ } # [END howto_operator_gce_args_common] -with models.DAG( +with DAG( DAG_ID, schedule_interval="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/compute/example_compute_ssh_parallel.py b/tests/system/providers/google/cloud/compute/example_compute_ssh_parallel.py index 0aea75f80b722..60efada741ba3 100644 --- a/tests/system/providers/google/cloud/compute/example_compute_ssh_parallel.py +++ b/tests/system/providers/google/cloud/compute/example_compute_ssh_parallel.py @@ -26,8 +26,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.hooks.compute_ssh import ComputeEngineSSHHook from airflow.providers.google.cloud.operators.compute import ( ComputeEngineDeleteInstanceOperator, @@ -69,7 +69,7 @@ } # [END howto_operator_gce_args_common] -with models.DAG( +with DAG( DAG_ID, schedule_interval="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_deidentify_content.py b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_deidentify_content.py index ce560edde199e..6c3d2cee9c6dc 100644 --- a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_deidentify_content.py +++ b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_deidentify_content.py @@ -27,7 +27,7 @@ from google.cloud.dlp_v2.types import ContentItem, DeidentifyTemplate, InspectConfig -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dlp import ( CloudDLPCreateDeidentifyTemplateOperator, CloudDLPDeidentifyContentOperator, @@ -84,7 +84,7 @@ TEMPLATE_ID = f"template_{DAG_ID}_{ENV_ID}" DEIDENTIFY_TEMPLATE = DeidentifyTemplate(deidentify_config=DEIDENTIFY_CONFIG) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_info_types.py b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_info_types.py index f0094b6544bbc..ae8f7f66469f8 100644 --- a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_info_types.py +++ b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_info_types.py @@ -28,7 +28,7 @@ from google.cloud.dlp_v2.types import ContentItem, InspectConfig, InspectTemplate -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dlp import ( CloudDLPCreateStoredInfoTypeOperator, CloudDLPDeleteStoredInfoTypeOperator, @@ -77,7 +77,7 @@ } } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_inspect_template.py b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_inspect_template.py index a440f3b2142bd..23d09aae03285 100644 --- a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_inspect_template.py +++ b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_inspect_template.py @@ -27,7 +27,7 @@ from google.cloud.dlp_v2.types import ContentItem, InspectConfig, InspectTemplate -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dlp import ( CloudDLPCreateInspectTemplateOperator, CloudDLPDeleteInspectTemplateOperator, @@ -53,7 +53,7 @@ INSPECT_TEMPLATE = InspectTemplate(inspect_config=INSPECT_CONFIG) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_job.py b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_job.py index 5004cedca3d0d..86906c144505c 100644 --- a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_job.py +++ b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_job.py @@ -27,7 +27,7 @@ from google.cloud.dlp_v2.types import InspectConfig, InspectJobConfig -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dlp import ( CloudDLPCancelDLPJobOperator, CloudDLPCreateDLPJobOperator, @@ -52,7 +52,7 @@ ) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_job_trigger.py b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_job_trigger.py index cfebe63bf869e..73308f115c6d6 100644 --- a/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_job_trigger.py +++ b/tests/system/providers/google/cloud/data_loss_prevention/example_dlp_job_trigger.py @@ -25,7 +25,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dlp import ( CloudDLPCreateJobTriggerOperator, CloudDLPDeleteJobTriggerOperator, @@ -51,7 +51,7 @@ TRIGGER_ID = f"trigger_{ENV_ID}" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataflow/example_dataflow_go.py b/tests/system/providers/google/cloud/dataflow/example_dataflow_go.py index 548a6902b7197..71985a2f835f7 100644 --- a/tests/system/providers/google/cloud/dataflow/example_dataflow_go.py +++ b/tests/system/providers/google/cloud/dataflow/example_dataflow_go.py @@ -27,7 +27,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.apache.beam.hooks.beam import BeamRunnerType from airflow.providers.apache.beam.operators.beam import BeamRunGoPipelineOperator from airflow.providers.google.cloud.hooks.dataflow import DataflowJobStatus @@ -61,7 +61,7 @@ } } -with models.DAG( +with DAG( "example_beam_native_go", start_date=datetime(2021, 1, 1), schedule="@once", diff --git a/tests/system/providers/google/cloud/dataflow/example_dataflow_native_java.py b/tests/system/providers/google/cloud/dataflow/example_dataflow_native_java.py index e3dbe69217090..53b33b89e9571 100644 --- a/tests/system/providers/google/cloud/dataflow/example_dataflow_native_java.py +++ b/tests/system/providers/google/cloud/dataflow/example_dataflow_native_java.py @@ -36,7 +36,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.apache.beam.hooks.beam import BeamRunnerType from airflow.providers.apache.beam.operators.beam import BeamRunJavaPipelineOperator from airflow.providers.google.cloud.operators.dataflow import CheckJobRunning @@ -56,7 +56,7 @@ GCS_JAR = f"gs://{PUBLIC_BUCKET}/{REMOTE_JAR_FILE_PATH}" LOCATION = "europe-west3" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataflow/example_dataflow_native_python.py b/tests/system/providers/google/cloud/dataflow/example_dataflow_native_python.py index 5b95bc8ffa663..957fe9afbe85a 100644 --- a/tests/system/providers/google/cloud/dataflow/example_dataflow_native_python.py +++ b/tests/system/providers/google/cloud/dataflow/example_dataflow_native_python.py @@ -25,7 +25,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.apache.beam.hooks.beam import BeamRunnerType from airflow.providers.apache.beam.operators.beam import BeamRunPythonPipelineOperator from airflow.providers.google.cloud.operators.dataflow import DataflowStopJobOperator @@ -53,7 +53,7 @@ } } -with models.DAG( +with DAG( DAG_ID, default_args=default_args, schedule="@once", diff --git a/tests/system/providers/google/cloud/dataflow/example_dataflow_native_python_async.py b/tests/system/providers/google/cloud/dataflow/example_dataflow_native_python_async.py index 9d03e851a3cdc..05695eab31c31 100644 --- a/tests/system/providers/google/cloud/dataflow/example_dataflow_native_python_async.py +++ b/tests/system/providers/google/cloud/dataflow/example_dataflow_native_python_async.py @@ -26,8 +26,8 @@ from pathlib import Path from typing import Callable -from airflow import models from airflow.exceptions import AirflowException +from airflow.models.dag import DAG from airflow.providers.apache.beam.hooks.beam import BeamRunnerType from airflow.providers.apache.beam.operators.beam import BeamRunPythonPipelineOperator from airflow.providers.google.cloud.hooks.dataflow import DataflowJobStatus @@ -61,7 +61,7 @@ } } -with models.DAG( +with DAG( DAG_ID, default_args=default_args, schedule="@once", diff --git a/tests/system/providers/google/cloud/dataflow/example_dataflow_sql.py b/tests/system/providers/google/cloud/dataflow/example_dataflow_sql.py index f9c4b9a776e93..86afc9fa38b7d 100644 --- a/tests/system/providers/google/cloud/dataflow/example_dataflow_sql.py +++ b/tests/system/providers/google/cloud/dataflow/example_dataflow_sql.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -52,7 +52,7 @@ ) -with models.DAG( +with DAG( dag_id=DAG_ID, start_date=datetime(2021, 1, 1), catchup=False, diff --git a/tests/system/providers/google/cloud/dataflow/example_dataflow_template.py b/tests/system/providers/google/cloud/dataflow/example_dataflow_template.py index 94af20ac666e7..9182834fbb5c4 100644 --- a/tests/system/providers/google/cloud/dataflow/example_dataflow_template.py +++ b/tests/system/providers/google/cloud/dataflow/example_dataflow_template.py @@ -26,7 +26,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataflow import ( DataflowStartFlexTemplateOperator, DataflowTemplatedJobStartOperator, @@ -72,7 +72,7 @@ }, } -with models.DAG( +with DAG( DAG_ID, default_args=default_args, schedule="@once", diff --git a/tests/system/providers/google/cloud/dataform/example_dataform.py b/tests/system/providers/google/cloud/dataform/example_dataform.py index f47f8579c501e..4e5053868b9f3 100644 --- a/tests/system/providers/google/cloud/dataform/example_dataform.py +++ b/tests/system/providers/google/cloud/dataform/example_dataform.py @@ -25,7 +25,7 @@ from google.cloud.dataform_v1beta1 import WorkflowInvocation -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import BigQueryDeleteDatasetOperator from airflow.providers.google.cloud.operators.dataform import ( DataformCancelWorkflowInvocationOperator, @@ -58,7 +58,7 @@ DATAFORM_SCHEMA_NAME = f"schema_{DAG_ID}_{ENV_ID}" # This DAG is not self-run we need to do some extra configuration to execute it in automation process -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/datafusion/example_datafusion.py b/tests/system/providers/google/cloud/datafusion/example_datafusion.py index c4926e3780690..71eb4aba6d9bf 100644 --- a/tests/system/providers/google/cloud/datafusion/example_datafusion.py +++ b/tests/system/providers/google/cloud/datafusion/example_datafusion.py @@ -22,8 +22,8 @@ import os from datetime import datetime -from airflow import models from airflow.decorators import task +from airflow.models.dag import DAG from airflow.providers.google.cloud.hooks.datafusion import DataFusionHook from airflow.providers.google.cloud.operators.datafusion import ( CloudDataFusionCreateInstanceOperator, @@ -162,7 +162,7 @@ CloudDataFusionCreatePipelineOperator.template_fields += ("pipeline",) -with models.DAG( +with DAG( DAG_ID, start_date=datetime(2021, 1, 1), catchup=False, diff --git a/tests/system/providers/google/cloud/datapipelines/example_datapipeline.py b/tests/system/providers/google/cloud/datapipelines/example_datapipeline.py index e4b82705a72f3..519087734a646 100644 --- a/tests/system/providers/google/cloud/datapipelines/example_datapipeline.py +++ b/tests/system/providers/google/cloud/datapipelines/example_datapipeline.py @@ -25,7 +25,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.datapipeline import ( CreateDataPipelineOperator, RunDataPipelineOperator, @@ -56,7 +56,7 @@ FILE_LOCAL_PATH = str(Path(__file__).parent / "resources" / FILE_NAME) TEMPLATE_LOCAL_PATH = str(Path(__file__).parent / "resources" / TEMPLATE_FILE) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataplex/example_dataplex.py b/tests/system/providers/google/cloud/dataplex/example_dataplex.py index 11c9edb507ec5..6cb15e396afc4 100644 --- a/tests/system/providers/google/cloud/dataplex/example_dataplex.py +++ b/tests/system/providers/google/cloud/dataplex/example_dataplex.py @@ -22,8 +22,8 @@ import datetime import os -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataplex import ( DataplexCreateLakeOperator, DataplexCreateTaskOperator, @@ -77,7 +77,7 @@ # [END howto_dataplex_lake_configuration] -with models.DAG( +with DAG( DAG_ID, start_date=datetime.datetime(2021, 1, 1), schedule="@once", diff --git a/tests/system/providers/google/cloud/dataplex/example_dataplex_dq.py b/tests/system/providers/google/cloud/dataplex/example_dataplex_dq.py index 1290698c0e97f..df503d531edae 100644 --- a/tests/system/providers/google/cloud/dataplex/example_dataplex_dq.py +++ b/tests/system/providers/google/cloud/dataplex/example_dataplex_dq.py @@ -25,8 +25,8 @@ from google.cloud import dataplex_v1 from google.protobuf.field_mask_pb2 import FieldMask -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateEmptyTableOperator, @@ -144,7 +144,7 @@ } -with models.DAG( +with DAG( DAG_ID, start_date=datetime(2021, 1, 1), schedule="@once", diff --git a/tests/system/providers/google/cloud/dataprep/example_dataprep.py b/tests/system/providers/google/cloud/dataprep/example_dataprep.py index a192f7fec5149..c07cd5a4562df 100644 --- a/tests/system/providers/google/cloud/dataprep/example_dataprep.py +++ b/tests/system/providers/google/cloud/dataprep/example_dataprep.py @@ -22,7 +22,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataprep import ( DataprepCopyFlowOperator, DataprepDeleteFlowOperator, @@ -57,7 +57,7 @@ }, ) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), # Override to match your needs diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_batch.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_batch.py index 3e9aed3502926..7dd5eff73aa73 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_batch.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_batch.py @@ -24,7 +24,7 @@ from google.api_core.retry import Retry -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCancelOperationOperator, DataprocCreateBatchOperator, @@ -53,7 +53,7 @@ } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_batch_deferrable.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_batch_deferrable.py index 8ed1893a5b3fd..4306219ee8fd9 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_batch_deferrable.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_batch_deferrable.py @@ -24,7 +24,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateBatchOperator, DataprocDeleteBatchOperator, @@ -45,7 +45,7 @@ } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_batch_persistent.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_batch_persistent.py index 2a8aaf4975360..ed96efa4b0d27 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_batch_persistent.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_batch_persistent.py @@ -22,7 +22,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( ClusterGenerator, DataprocCreateBatchOperator, @@ -66,7 +66,7 @@ } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_deferrable.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_deferrable.py index 3c39a99ae59a2..a0c427b568757 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_deferrable.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_deferrable.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -76,7 +76,7 @@ TIMEOUT = {"seconds": 1 * 24 * 60 * 60} -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_generator.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_generator.py index 9267dbf7ebe13..9761326c4cc5e 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_generator.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_generator.py @@ -25,7 +25,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( ClusterGenerator, DataprocCreateClusterOperator, @@ -68,7 +68,7 @@ # [END how_to_cloud_dataproc_create_cluster_generate_cluster_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_update.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_update.py index ca7dff04af6bd..643ac7bf82317 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_update.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_cluster_update.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -66,7 +66,7 @@ TIMEOUT = {"seconds": 1 * 24 * 60 * 60} -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_gke.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_gke.py index 2958229a15a02..c839136528863 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_gke.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_gke.py @@ -30,7 +30,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -83,7 +83,7 @@ # [END how_to_cloud_dataproc_create_cluster_in_gke_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_hadoop.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_hadoop.py index 690561b66bee2..734dd5830b8da 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_hadoop.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_hadoop.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -70,7 +70,7 @@ # [END how_to_cloud_dataproc_hadoop_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_hive.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_hive.py index 1bc4e18a4b5c2..af3c54636b141 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_hive.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_hive.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -75,7 +75,7 @@ # [END how_to_cloud_dataproc_hive_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_pig.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_pig.py index c0157e5f9691a..a7d27efdc8174 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_pig.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_pig.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -63,7 +63,7 @@ # [END how_to_cloud_dataproc_pig_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_presto.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_presto.py index a41b5077f59b8..bd4503183053c 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_presto.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_presto.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -70,7 +70,7 @@ # [END how_to_cloud_dataproc_presto_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_pyspark.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_pyspark.py index b73ae2742fedb..237d0197c3a5b 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_pyspark.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_pyspark.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -63,7 +63,7 @@ # [END how_to_cloud_dataproc_pyspark_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_spark.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_spark.py index f30c7224f67e0..3acda02d3cb09 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_spark.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_spark.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -65,7 +65,7 @@ # [END how_to_cloud_dataproc_spark_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_async.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_async.py index 63485d201403a..80ef2388f3a33 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_async.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_async.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -64,7 +64,7 @@ } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_deferrable.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_deferrable.py index baa7875df2f69..336def4696b34 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_deferrable.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_deferrable.py @@ -24,7 +24,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -66,7 +66,7 @@ # [END how_to_cloud_dataproc_spark_deferrable_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_sql.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_sql.py index ebf3333d25f4b..f50cc81b92dcb 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_sql.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_spark_sql.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -62,7 +62,7 @@ # [END how_to_cloud_dataproc_sparksql_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_sparkr.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_sparkr.py index b98fcde8c2388..8f76bf6dde80e 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_sparkr.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_sparkr.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -63,7 +63,7 @@ # [END how_to_cloud_dataproc_sparkr_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_trino.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_trino.py index 4d60157c26048..04032f351330b 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_trino.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_trino.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, DataprocDeleteClusterOperator, @@ -72,7 +72,7 @@ # [END how_to_cloud_dataproc_trino_config] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_workflow.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_workflow.py index dce59531826c1..0736d16bdab82 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_workflow.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_workflow.py @@ -22,7 +22,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateWorkflowTemplateOperator, DataprocInstantiateInlineWorkflowTemplateOperator, @@ -61,7 +61,7 @@ } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc/example_dataproc_workflow_deferrable.py b/tests/system/providers/google/cloud/dataproc/example_dataproc_workflow_deferrable.py index c6f9909d92db5..6500af4366b70 100644 --- a/tests/system/providers/google/cloud/dataproc/example_dataproc_workflow_deferrable.py +++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_workflow_deferrable.py @@ -22,7 +22,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateWorkflowTemplateOperator, DataprocInstantiateInlineWorkflowTemplateOperator, @@ -61,7 +61,7 @@ } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore.py b/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore.py index e3c6d2a3b5a43..8b7208dd1e6c7 100644 --- a/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore.py +++ b/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore.py @@ -27,7 +27,7 @@ from google.protobuf.field_mask_pb2 import FieldMask -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc_metastore import ( DataprocMetastoreCreateMetadataImportOperator, DataprocMetastoreCreateServiceOperator, @@ -86,7 +86,7 @@ UPDATE_MASK = FieldMask(paths=["labels"]) # [END how_to_cloud_dataproc_metastore_update_service] -with models.DAG( +with DAG( DAG_ID, start_date=datetime.datetime(2021, 1, 1), schedule="@once", diff --git a/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore_backup.py b/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore_backup.py index 5351d9df7d618..c2423feaae294 100644 --- a/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore_backup.py +++ b/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore_backup.py @@ -24,7 +24,7 @@ import datetime import os -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.dataproc_metastore import ( DataprocMetastoreCreateBackupOperator, DataprocMetastoreCreateServiceOperator, @@ -55,7 +55,7 @@ } # [END how_to_cloud_dataproc_metastore_create_backup] -with models.DAG( +with DAG( DAG_ID, start_date=datetime.datetime(2021, 1, 1), schedule="@once", diff --git a/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore_hive_partition_sensor.py b/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore_hive_partition_sensor.py index 134caff6dbe00..8b83b1f07b10a 100644 --- a/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore_hive_partition_sensor.py +++ b/tests/system/providers/google/cloud/dataproc_metastore/example_dataproc_metastore_hive_partition_sensor.py @@ -26,8 +26,8 @@ import datetime import os -from airflow import models from airflow.decorators import task +from airflow.models.dag import DAG from airflow.providers.google.cloud.hooks.gcs import _parse_gcs_url from airflow.providers.google.cloud.operators.dataproc import ( DataprocCreateClusterOperator, @@ -106,7 +106,7 @@ SELECT SubmissionDate,TransactionAmount,TransactionType FROM transactions; """ -with models.DAG( +with DAG( DAG_ID, start_date=datetime.datetime(2021, 1, 1), schedule="@once", diff --git a/tests/system/providers/google/cloud/datastore/example_datastore_commit.py b/tests/system/providers/google/cloud/datastore/example_datastore_commit.py index 0c5cf5cda3934..6dba0709b1409 100644 --- a/tests/system/providers/google/cloud/datastore/example_datastore_commit.py +++ b/tests/system/providers/google/cloud/datastore/example_datastore_commit.py @@ -24,8 +24,8 @@ from datetime import datetime from typing import Any -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.datastore import ( CloudDatastoreAllocateIdsOperator, CloudDatastoreBeginTransactionOperator, @@ -57,7 +57,7 @@ # [END how_to_transaction_def] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/datastore/example_datastore_query.py b/tests/system/providers/google/cloud/datastore/example_datastore_query.py index 2ac98ac85dfa2..cc461da927e45 100644 --- a/tests/system/providers/google/cloud/datastore/example_datastore_query.py +++ b/tests/system/providers/google/cloud/datastore/example_datastore_query.py @@ -24,7 +24,7 @@ from datetime import datetime from typing import Any -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.datastore import ( CloudDatastoreAllocateIdsOperator, CloudDatastoreBeginTransactionOperator, @@ -46,7 +46,7 @@ TRANSACTION_OPTIONS: dict[str, Any] = {"readWrite": {}} -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/datastore/example_datastore_rollback.py b/tests/system/providers/google/cloud/datastore/example_datastore_rollback.py index c8d06cab18c68..57d33879c7869 100644 --- a/tests/system/providers/google/cloud/datastore/example_datastore_rollback.py +++ b/tests/system/providers/google/cloud/datastore/example_datastore_rollback.py @@ -24,7 +24,7 @@ from datetime import datetime from typing import Any -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.datastore import ( CloudDatastoreBeginTransactionOperator, CloudDatastoreRollbackOperator, @@ -38,7 +38,7 @@ TRANSACTION_OPTIONS: dict[str, Any] = {"readWrite": {}} -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_calendar_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_calendar_to_gcs.py index 76a9781670307..d0d1c96f28ef6 100644 --- a/tests/system/providers/google/cloud/gcs/example_calendar_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_calendar_to_gcs.py @@ -20,7 +20,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.calendar_to_gcs import GoogleCalendarToGCSOperator from airflow.utils.trigger_rule import TriggerRule @@ -33,7 +33,7 @@ CALENDAR_ID = os.environ.get("CALENDAR_ID", "1234567890qwerty") API_VERSION = "v3" -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_firestore.py b/tests/system/providers/google/cloud/gcs/example_firestore.py index ae904cf8dccfa..acc67d80edb1b 100644 --- a/tests/system/providers/google/cloud/gcs/example_firestore.py +++ b/tests/system/providers/google/cloud/gcs/example_firestore.py @@ -27,7 +27,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateExternalTableOperator, @@ -63,7 +63,7 @@ } -with models.DAG( +with DAG( DAG_ID, start_date=datetime(2021, 1, 1), schedule="@once", diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_acl.py b/tests/system/providers/google/cloud/gcs/example_gcs_acl.py index 4d635758ef74f..01ab83e51ae35 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_acl.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_acl.py @@ -24,7 +24,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSBucketCreateAclEntryOperator, GCSCreateBucketOperator, @@ -48,7 +48,7 @@ GCS_ACL_OBJECT_ROLE = "OWNER" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_copy_delete.py b/tests/system/providers/google/cloud/gcs/example_gcs_copy_delete.py index 9d1a3f5162275..f89464b829ca5 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_copy_delete.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_copy_delete.py @@ -25,8 +25,8 @@ from datetime import datetime from pathlib import Path -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, @@ -49,7 +49,7 @@ UPLOAD_FILE_PATH = str(Path(__file__).parent / "resources" / FILE_NAME) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_sensor.py b/tests/system/providers/google/cloud/gcs/example_gcs_sensor.py index 5313c4847dc17..607c670cc2908 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_sensor.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_sensor.py @@ -24,8 +24,8 @@ from datetime import datetime from pathlib import Path -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.sensors.gcs import ( GCSObjectExistenceAsyncSensor, @@ -67,7 +67,7 @@ def mode_setter(self, value): return property(mode_getter, mode_setter) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_to_bigquery.py b/tests/system/providers/google/cloud/gcs/example_gcs_to_bigquery.py index 9d04992fa4861..732034f699766 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_to_bigquery.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_to_bigquery.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryDeleteDatasetOperator, @@ -38,7 +38,7 @@ TABLE_NAME = "test" PROJECT_ID = os.environ.get("SYSTEM_TESTS_GCP_PROJECT") -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_to_bigquery_async.py b/tests/system/providers/google/cloud/gcs/example_gcs_to_bigquery_async.py index 0bda9be4c3f3e..92308f4a3987f 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_to_bigquery_async.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_to_bigquery_async.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryDeleteDatasetOperator, @@ -46,7 +46,7 @@ MAX_ID_STR = "name" MAX_ID_DATE = "date" -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_gcs_to_gcs.py index 86b417c600e6c..46077894031c6 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_to_gcs.py @@ -25,9 +25,9 @@ import shutil from datetime import datetime -from airflow import models from airflow.decorators import task from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, @@ -50,7 +50,7 @@ PREFIX = f"{HOME}/data/{DAG_ID}_{ENV_ID}/" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_to_gdrive.py b/tests/system/providers/google/cloud/gcs/example_gcs_to_gdrive.py index c3371a2c3c111..e9cd16d8f97b0 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_to_gdrive.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_to_gdrive.py @@ -27,7 +27,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.local_to_gcs import LocalFilesystemToGCSOperator from airflow.providers.google.suite.transfers.gcs_to_gdrive import GCSToGoogleDriveOperator @@ -50,7 +50,7 @@ FILE_NAME = "example_upload.txt" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_to_sheets.py b/tests/system/providers/google/cloud/gcs/example_gcs_to_sheets.py index e1dbb1b7b67f6..e2166fd78871d 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_to_sheets.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_to_sheets.py @@ -20,7 +20,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.sheets_to_gcs import GoogleSheetsToGCSOperator from airflow.providers.google.suite.transfers.gcs_to_sheets import GCSToGoogleSheetsOperator @@ -34,7 +34,7 @@ SPREADSHEET_ID = os.environ.get("SPREADSHEET_ID", "example-spreadsheetID") NEW_SPREADSHEET_ID = os.environ.get("NEW_SPREADSHEET_ID", "1234567890qwerty") -with models.DAG( +with DAG( DAG_ID, start_date=datetime(2021, 1, 1), schedule="@once", # Override to match your needs diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_transform.py b/tests/system/providers/google/cloud/gcs/example_gcs_transform.py index b6af144dcd772..355fa1e99b293 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_transform.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_transform.py @@ -24,7 +24,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -46,7 +46,7 @@ TRANSFORM_SCRIPT_PATH = str(Path(__file__).parent / "resources" / "transform_script.py") -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_transform_timespan.py b/tests/system/providers/google/cloud/gcs/example_gcs_transform_timespan.py index 3e08fb87ff1b4..75812e63bb9aa 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_transform_timespan.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_transform_timespan.py @@ -24,8 +24,8 @@ from datetime import datetime from pathlib import Path -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -52,7 +52,7 @@ TRANSFORM_SCRIPT_PATH = str(Path(__file__).parent / "resources" / "transform_timespan.py") -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gcs_upload_download.py b/tests/system/providers/google/cloud/gcs/example_gcs_upload_download.py index b1a0fd68df027..437cb4f87f619 100644 --- a/tests/system/providers/google/cloud/gcs/example_gcs_upload_download.py +++ b/tests/system/providers/google/cloud/gcs/example_gcs_upload_download.py @@ -24,7 +24,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.gcs_to_local import GCSToLocalFilesystemOperator from airflow.providers.google.cloud.transfers.local_to_gcs import LocalFilesystemToGCSOperator @@ -41,7 +41,7 @@ UPLOAD_FILE_PATH = str(Path(__file__).parent / "resources" / FILE_NAME) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_gdrive_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_gdrive_to_gcs.py index 7a26e630bbb09..b3debb12ddff1 100644 --- a/tests/system/providers/google/cloud/gcs/example_gdrive_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_gdrive_to_gcs.py @@ -21,7 +21,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.gdrive_to_gcs import GoogleDriveToGCSOperator from airflow.providers.google.cloud.transfers.local_to_gcs import LocalFilesystemToGCSOperator @@ -41,7 +41,7 @@ FILE_NAME = "example_upload.txt" LOCAL_PATH = str(Path(__file__).parent / "resources" / FILE_NAME) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_mssql_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_mssql_to_gcs.py index 76b9a680959f2..f7d7dd4a82deb 100644 --- a/tests/system/providers/google/cloud/gcs/example_mssql_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_mssql_to_gcs.py @@ -21,7 +21,7 @@ import pytest -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator try: @@ -41,7 +41,7 @@ SQL_QUERY = "USE airflow SELECT * FROM Country;" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_mysql_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_mysql_to_gcs.py index 97404f1a8aa47..818004fc613ce 100644 --- a/tests/system/providers/google/cloud/gcs/example_mysql_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_mysql_to_gcs.py @@ -21,7 +21,7 @@ import pytest -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator try: @@ -40,7 +40,7 @@ SQL_QUERY = "SELECT * from test_table" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_oracle_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_oracle_to_gcs.py index ead7f5c86ba80..571478c8333c0 100644 --- a/tests/system/providers/google/cloud/gcs/example_oracle_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_oracle_to_gcs.py @@ -19,7 +19,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.oracle_to_gcs import OracleToGCSOperator from airflow.utils.trigger_rule import TriggerRule @@ -32,7 +32,7 @@ FILENAME = "test_file" SQL_QUERY = "SELECT * from test_table" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_s3_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_s3_to_gcs.py index f735b3ea55a3f..437f09a7e6613 100644 --- a/tests/system/providers/google/cloud/gcs/example_s3_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_s3_to_gcs.py @@ -20,8 +20,8 @@ from datetime import datetime from pathlib import Path -from airflow import models from airflow.decorators import task +from airflow.models.dag import DAG from airflow.providers.amazon.aws.hooks.s3 import S3Hook from airflow.providers.amazon.aws.operators.s3 import S3CreateBucketOperator, S3DeleteBucketOperator from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator @@ -46,7 +46,7 @@ def upload_file(): s3_hook.load_file(filename=UPLOAD_FILE, key=PREFIX, bucket_name=BUCKET_NAME) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_s3_to_gcs_async.py b/tests/system/providers/google/cloud/gcs/example_s3_to_gcs_async.py index ccecc5db4349c..4c3832a96ad85 100644 --- a/tests/system/providers/google/cloud/gcs/example_s3_to_gcs_async.py +++ b/tests/system/providers/google/cloud/gcs/example_s3_to_gcs_async.py @@ -19,8 +19,8 @@ import os from datetime import datetime -from airflow import models from airflow.decorators import task +from airflow.models.dag import DAG from airflow.providers.amazon.aws.hooks.s3 import S3Hook from airflow.providers.amazon.aws.operators.s3 import S3CreateBucketOperator, S3DeleteBucketOperator from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator @@ -44,7 +44,7 @@ def upload_file(): s3_hook.load_file(filename=UPLOAD_FILE, key=PREFIX, bucket_name=BUCKET_NAME) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_sftp_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_sftp_to_gcs.py index 80b53774d55f2..a7b4274c78789 100644 --- a/tests/system/providers/google/cloud/gcs/example_sftp_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_sftp_to_gcs.py @@ -24,8 +24,8 @@ from datetime import datetime from pathlib import Path -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.sftp_to_gcs import SFTPToGCSOperator @@ -51,7 +51,7 @@ FILE_NAME = "tmp.tar.gz" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_sheets.py b/tests/system/providers/google/cloud/gcs/example_sheets.py index aef54aedbcc4f..ef43fd03fc1e5 100644 --- a/tests/system/providers/google/cloud/gcs/example_sheets.py +++ b/tests/system/providers/google/cloud/gcs/example_sheets.py @@ -20,7 +20,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator @@ -42,7 +42,7 @@ "sheets": [{"properties": {"title": "Sheet1"}}], } -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs, start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_sheets_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_sheets_to_gcs.py index fb414461b6f72..ad06b8e6f37af 100644 --- a/tests/system/providers/google/cloud/gcs/example_sheets_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_sheets_to_gcs.py @@ -20,7 +20,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.sheets_to_gcs import GoogleSheetsToGCSOperator from airflow.utils.trigger_rule import TriggerRule @@ -32,7 +32,7 @@ BUCKET_NAME = f"bucket_{DAG_ID}_{ENV_ID}" SPREADSHEET_ID = os.environ.get("SPREADSHEET_ID", "1234567890qwerty") -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/gcs/example_trino_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_trino_to_gcs.py index 96690f71471a7..d2b661a643971 100644 --- a/tests/system/providers/google/cloud/gcs/example_trino_to_gcs.py +++ b/tests/system/providers/google/cloud/gcs/example_trino_to_gcs.py @@ -24,7 +24,7 @@ import re from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.bigquery import ( BigQueryCreateEmptyDatasetOperator, BigQueryCreateExternalTableOperator, @@ -53,7 +53,7 @@ def safe_name(s: str) -> str: return re.sub("[^0-9a-zA-Z_]+", "_", s) -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/kubernetes_engine/example_kubernetes_engine.py b/tests/system/providers/google/cloud/kubernetes_engine/example_kubernetes_engine.py index bf155499f2949..4f53fe5183a5c 100644 --- a/tests/system/providers/google/cloud/kubernetes_engine/example_kubernetes_engine.py +++ b/tests/system/providers/google/cloud/kubernetes_engine/example_kubernetes_engine.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.kubernetes_engine import ( GKECreateClusterOperator, @@ -42,7 +42,7 @@ CLUSTER = {"name": CLUSTER_NAME, "initial_node_count": 1} # [END howto_operator_gcp_gke_create_cluster_definition] -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/kubernetes_engine/example_kubernetes_engine_async.py b/tests/system/providers/google/cloud/kubernetes_engine/example_kubernetes_engine_async.py index 73430a39385af..70c465a0c1992 100644 --- a/tests/system/providers/google/cloud/kubernetes_engine/example_kubernetes_engine_async.py +++ b/tests/system/providers/google/cloud/kubernetes_engine/example_kubernetes_engine_async.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.kubernetes_engine import ( GKECreateClusterOperator, @@ -40,7 +40,7 @@ CLUSTER = {"name": CLUSTER_NAME, "initial_node_count": 1} -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/life_sciences/example_life_sciences.py b/tests/system/providers/google/cloud/life_sciences/example_life_sciences.py index 73c4d701b7a14..e2361df2f858a 100644 --- a/tests/system/providers/google/cloud/life_sciences/example_life_sciences.py +++ b/tests/system/providers/google/cloud/life_sciences/example_life_sciences.py @@ -21,8 +21,8 @@ from datetime import datetime from pathlib import Path -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.life_sciences import LifeSciencesRunPipelineOperator from airflow.providers.google.cloud.transfers.local_to_gcs import LocalFilesystemToGCSOperator @@ -85,7 +85,7 @@ } # [END howto_configure_multiple_action_pipeline] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/ml_engine/example_mlengine.py b/tests/system/providers/google/cloud/ml_engine/example_mlengine.py index 51bf14fb5e88e..8fac67aab5403 100644 --- a/tests/system/providers/google/cloud/ml_engine/example_mlengine.py +++ b/tests/system/providers/google/cloud/ml_engine/example_mlengine.py @@ -25,8 +25,8 @@ from datetime import datetime from math import ceil -from airflow import models from airflow.decorators import task +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.mlengine import ( @@ -69,7 +69,7 @@ def generate_model_predict_input_data() -> list[int]: return [1, 4, 9, 16, 25, 36] -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/ml_engine/example_mlengine_async.py b/tests/system/providers/google/cloud/ml_engine/example_mlengine_async.py index 7c2caef846cfc..301fb1ffa2b1a 100644 --- a/tests/system/providers/google/cloud/ml_engine/example_mlengine_async.py +++ b/tests/system/providers/google/cloud/ml_engine/example_mlengine_async.py @@ -25,8 +25,8 @@ from datetime import datetime from math import ceil -from airflow import models from airflow.decorators import task +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.mlengine import ( @@ -69,7 +69,7 @@ def generate_model_predict_input_data() -> list[int]: return [1, 4, 9, 16, 25, 36] -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/natural_language/example_natural_language.py b/tests/system/providers/google/cloud/natural_language/example_natural_language.py index 9c9cae26b86d6..ad8fb72014f10 100644 --- a/tests/system/providers/google/cloud/natural_language/example_natural_language.py +++ b/tests/system/providers/google/cloud/natural_language/example_natural_language.py @@ -25,7 +25,7 @@ from google.cloud.language_v1 import Document -from airflow import models +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.natural_language import ( CloudNaturalLanguageAnalyzeEntitiesOperator, @@ -54,7 +54,7 @@ # [END howto_operator_gcp_natural_language_document_gcs] -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/pubsub/example_pubsub.py b/tests/system/providers/google/cloud/pubsub/example_pubsub.py index 81687f97bb90c..167cbf04ac0a1 100644 --- a/tests/system/providers/google/cloud/pubsub/example_pubsub.py +++ b/tests/system/providers/google/cloud/pubsub/example_pubsub.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.pubsub import ( PubSubCreateSubscriptionOperator, @@ -53,7 +53,7 @@ """ # [END howto_operator_gcp_pubsub_pull_messages_result_cmd] -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/pubsub/example_pubsub_deferrable.py b/tests/system/providers/google/cloud/pubsub/example_pubsub_deferrable.py index 3f5b9a2b5dd2e..f40b03b698d57 100644 --- a/tests/system/providers/google/cloud/pubsub/example_pubsub_deferrable.py +++ b/tests/system/providers/google/cloud/pubsub/example_pubsub_deferrable.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.pubsub import ( PubSubCreateSubscriptionOperator, PubSubCreateTopicOperator, @@ -43,7 +43,7 @@ MESSAGE = {"data": b"Tool", "attributes": {"name": "wrench", "mass": "1.3kg", "count": "3"}} -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/spanner/example_spanner.py b/tests/system/providers/google/cloud/spanner/example_spanner.py index f3228f95001d6..cb945b7d05b8f 100644 --- a/tests/system/providers/google/cloud/spanner/example_spanner.py +++ b/tests/system/providers/google/cloud/spanner/example_spanner.py @@ -23,7 +23,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.spanner import ( SpannerDeleteDatabaseInstanceOperator, SpannerDeleteInstanceOperator, @@ -48,7 +48,7 @@ OPERATION_ID = "unique_operation_id" -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/speech_to_text/example_speech_to_text.py b/tests/system/providers/google/cloud/speech_to_text/example_speech_to_text.py index c18ba81964d54..5afc86e64c799 100644 --- a/tests/system/providers/google/cloud/speech_to_text/example_speech_to_text.py +++ b/tests/system/providers/google/cloud/speech_to_text/example_speech_to_text.py @@ -20,7 +20,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.speech_to_text import CloudSpeechToTextRecognizeSpeechOperator from airflow.providers.google.cloud.operators.text_to_speech import CloudTextToSpeechSynthesizeOperator @@ -47,7 +47,7 @@ AUDIO = {"uri": f"gs://{BUCKET_NAME}/{FILE_NAME}"} # [END howto_operator_speech_to_text_api_arguments] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/sql_to_sheets/example_sql_to_sheets.py b/tests/system/providers/google/cloud/sql_to_sheets/example_sql_to_sheets.py index e3d2dc711ee4e..a360a7de42089 100644 --- a/tests/system/providers/google/cloud/sql_to_sheets/example_sql_to_sheets.py +++ b/tests/system/providers/google/cloud/sql_to_sheets/example_sql_to_sheets.py @@ -69,7 +69,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.suite.transfers.sql_to_sheets import SQLToGoogleSheetsOperator ENV_ID = os.environ.get("SYSTEM_TESTS_ENV_ID") @@ -79,7 +79,7 @@ DAG_ID = "example_sql_to_sheets" SQL = "select col2 from test_table" -with models.DAG( +with DAG( DAG_ID, start_date=datetime(2021, 1, 1), schedule="@once", # Override to match your needs diff --git a/tests/system/providers/google/cloud/stackdriver/example_stackdriver.py b/tests/system/providers/google/cloud/stackdriver/example_stackdriver.py index 12c68a4afd392..d4c6fcfbf4107 100644 --- a/tests/system/providers/google/cloud/stackdriver/example_stackdriver.py +++ b/tests/system/providers/google/cloud/stackdriver/example_stackdriver.py @@ -24,8 +24,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.stackdriver import ( StackdriverDeleteAlertOperator, StackdriverDeleteNotificationChannelOperator, @@ -123,7 +123,7 @@ "type": "pubsub", } -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_aws.py b/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_aws.py index 9855f1ceb5ea2..c2fb4e59457e8 100644 --- a/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_aws.py +++ b/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_aws.py @@ -25,7 +25,7 @@ from pydantic.main import deepcopy -from airflow import models +from airflow.models.dag import DAG from airflow.providers.amazon.aws.operators.s3 import S3CreateBucketOperator, S3DeleteBucketOperator from airflow.providers.amazon.aws.transfers.gcs_to_s3 import GCSToS3Operator from airflow.providers.google.cloud.hooks.cloud_storage_transfer_service import ( @@ -100,7 +100,7 @@ aws_to_gcs_transfer_body_2 = deepcopy(aws_to_gcs_transfer_body) aws_to_gcs_transfer_body_2[JOB_NAME] = GCP_TRANSFER_JOB_2_NAME -with models.DAG( +with DAG( dag_id=DAG_ID, start_date=datetime(2021, 1, 1), catchup=False, diff --git a/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_gcp.py b/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_gcp.py index 3bf1f43a20806..fe3ef438686ce 100644 --- a/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_gcp.py +++ b/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_gcp.py @@ -26,7 +26,7 @@ from datetime import datetime, timedelta from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.hooks.cloud_storage_transfer_service import ( ALREADY_EXISTING_IN_SINK, BUCKET_NAME, @@ -101,7 +101,7 @@ } # [END howto_operator_gcp_transfer_update_job_body] -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_gcs_to_gcs.py b/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_gcs_to_gcs.py index 93f7547a0b40f..8a312e042f99c 100644 --- a/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_gcs_to_gcs.py +++ b/tests/system/providers/google/cloud/storage_transfer/example_cloud_storage_transfer_service_gcs_to_gcs.py @@ -26,7 +26,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.cloud_storage_transfer_service import ( CloudDataTransferServiceGCSToGCSOperator, ) @@ -50,7 +50,7 @@ FILE_LOCAL_PATH = str(Path(CURRENT_FOLDER) / "resources" / FILE_NAME) -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/tasks/example_queue.py b/tests/system/providers/google/cloud/tasks/example_queue.py index 6fd2647b02b4c..6030fc182ed04 100644 --- a/tests/system/providers/google/cloud/tasks/example_queue.py +++ b/tests/system/providers/google/cloud/tasks/example_queue.py @@ -31,9 +31,9 @@ from google.cloud.tasks_v2.types import Queue from google.protobuf.field_mask_pb2 import FieldMask -from airflow import models from airflow.decorators import task from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.tasks import ( CloudTasksQueueCreateOperator, @@ -54,7 +54,7 @@ QUEUE_ID = f"queue-{ENV_ID}-{DAG_ID.replace('_', '-')}" -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/tasks/example_tasks.py b/tests/system/providers/google/cloud/tasks/example_tasks.py index 1a6acdb250020..24d3aff7f5cb2 100644 --- a/tests/system/providers/google/cloud/tasks/example_tasks.py +++ b/tests/system/providers/google/cloud/tasks/example_tasks.py @@ -28,9 +28,9 @@ from google.cloud.tasks_v2.types import Queue from google.protobuf import timestamp_pb2 -from airflow import models from airflow.decorators import task from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.tasks import ( CloudTasksQueueCreateOperator, CloudTasksQueueDeleteOperator, @@ -61,7 +61,7 @@ "schedule_time": timestamp, } -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/text_to_speech/example_text_to_speech.py b/tests/system/providers/google/cloud/text_to_speech/example_text_to_speech.py index 0e414ae5a58e9..b185787ac04c6 100644 --- a/tests/system/providers/google/cloud/text_to_speech/example_text_to_speech.py +++ b/tests/system/providers/google/cloud/text_to_speech/example_text_to_speech.py @@ -20,7 +20,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.text_to_speech import CloudTextToSpeechSynthesizeOperator from airflow.utils.trigger_rule import TriggerRule @@ -41,7 +41,7 @@ AUDIO_CONFIG = {"audio_encoding": "LINEAR16"} # [END howto_operator_text_to_speech_api_arguments] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/transfers/example_gcs_to_sftp.py b/tests/system/providers/google/cloud/transfers/example_gcs_to_sftp.py index a4f2427fd1e43..a1a58e0f3aa20 100644 --- a/tests/system/providers/google/cloud/transfers/example_gcs_to_sftp.py +++ b/tests/system/providers/google/cloud/transfers/example_gcs_to_sftp.py @@ -24,7 +24,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.gcs_to_sftp import GCSToSFTPOperator from airflow.providers.google.cloud.transfers.local_to_gcs import LocalFilesystemToGCSOperator @@ -47,7 +47,7 @@ UPLOAD_IN_DIR_DST = f"dir-2/{FILE_NAME}" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/transfers/example_gdrive_to_local.py b/tests/system/providers/google/cloud/transfers/example_gdrive_to_local.py index b31406fdba639..a7a0b6ac3d5a5 100644 --- a/tests/system/providers/google/cloud/transfers/example_gdrive_to_local.py +++ b/tests/system/providers/google/cloud/transfers/example_gdrive_to_local.py @@ -27,8 +27,8 @@ from datetime import datetime from pathlib import Path -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.transfers.gdrive_to_local import GoogleDriveToLocalOperator from airflow.providers.google.cloud.transfers.local_to_gcs import LocalFilesystemToGCSOperator @@ -47,7 +47,7 @@ LOCAL_PATH = str(Path(__file__).parent / "resources" / FILE_NAME) -with models.DAG( +with DAG( DAG_ID, start_date=datetime(2021, 1, 1), schedule="@once", diff --git a/tests/system/providers/google/cloud/transfers/example_postgres_to_gcs.py b/tests/system/providers/google/cloud/transfers/example_postgres_to_gcs.py index 64e1e1574d5ad..45d52ddbe160a 100644 --- a/tests/system/providers/google/cloud/transfers/example_postgres_to_gcs.py +++ b/tests/system/providers/google/cloud/transfers/example_postgres_to_gcs.py @@ -26,9 +26,9 @@ from googleapiclient import discovery -from airflow import models from airflow.decorators import task -from airflow.models import Connection +from airflow.models.connection import Connection +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.common.sql.operators.sql import SQLExecuteQueryOperator from airflow.providers.google.cloud.operators.cloud_sql import ( @@ -86,7 +86,7 @@ log = logging.getLogger(__name__) -with models.DAG( +with DAG( dag_id=DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/translate/example_translate.py b/tests/system/providers/google/cloud/translate/example_translate.py index dad18b074ab75..5a5ad1e943317 100644 --- a/tests/system/providers/google/cloud/translate/example_translate.py +++ b/tests/system/providers/google/cloud/translate/example_translate.py @@ -23,13 +23,13 @@ from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.translate import CloudTranslateTextOperator DAG_ID = "example_gcp_translate" -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/translate_speech/example_translate_speech.py b/tests/system/providers/google/cloud/translate_speech/example_translate_speech.py index 3675ff765da9f..ba58833079926 100644 --- a/tests/system/providers/google/cloud/translate_speech/example_translate_speech.py +++ b/tests/system/providers/google/cloud/translate_speech/example_translate_speech.py @@ -20,7 +20,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.text_to_speech import CloudTextToSpeechSynthesizeOperator from airflow.providers.google.cloud.operators.translate_speech import CloudTranslateSpeechOperator @@ -53,7 +53,7 @@ # [END howto_operator_translate_speech_arguments] -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_forecasting_training.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_forecasting_training.py index fb6a7e441e42b..5f0ad8be24590 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_forecasting_training.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_forecasting_training.py @@ -31,7 +31,7 @@ from google.protobuf.json_format import ParseDict from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -83,7 +83,7 @@ } -with models.DAG( +with DAG( f"{DAG_ID}_forecasting_training_job", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_image_training.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_image_training.py index b7f3bb15b9a70..9ec4049479676 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_image_training.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_image_training.py @@ -30,7 +30,7 @@ from google.cloud.aiplatform import schema from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -70,7 +70,7 @@ ] -with models.DAG( +with DAG( f"{DAG_ID}_image_training_job", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_list_training.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_list_training.py index b8f5cdae68b6e..381b52ebe88e7 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_list_training.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_list_training.py @@ -27,7 +27,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.vertex_ai.auto_ml import ListAutoMLTrainingJobOperator ENV_ID = os.environ.get("SYSTEM_TESTS_ENV_ID") @@ -36,7 +36,7 @@ REGION = "us-central1" -with models.DAG( +with DAG( f"{DAG_ID}_list_training_job", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_tabular_training.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_tabular_training.py index 36ed1ff12d1bf..633a357a290f5 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_tabular_training.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_tabular_training.py @@ -31,7 +31,7 @@ from google.protobuf.json_format import ParseDict from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -85,7 +85,7 @@ {"numeric": {"column_name": "PhotoAmt"}}, ] -with models.DAG( +with DAG( f"{DAG_ID}_tabular_training_job", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_text_training.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_text_training.py index 1780f1cfa6f7c..64cd2a54ea9da 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_text_training.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_text_training.py @@ -30,7 +30,7 @@ from google.cloud.aiplatform import schema from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -69,7 +69,7 @@ }, ] -with models.DAG( +with DAG( f"{DAG_ID}_text_training_job", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_video_training.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_video_training.py index d921d16d53877..c6810ed36ec3c 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_video_training.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_auto_ml_video_training.py @@ -30,7 +30,7 @@ from google.cloud.aiplatform import schema from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -69,7 +69,7 @@ }, ] -with models.DAG( +with DAG( f"{DAG_ID}_video_training_job", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_batch_prediction_job.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_batch_prediction_job.py index 5d696f34578a8..1b2c3e873305d 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_batch_prediction_job.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_batch_prediction_job.py @@ -31,7 +31,7 @@ from google.protobuf.json_format import ParseDict from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -94,7 +94,7 @@ MODEL_PARAMETERS = ParseDict({}, Value()) -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_container.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_container.py index 612e3e4f9a006..50f62fbf3c44a 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_container.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_container.py @@ -31,7 +31,7 @@ from google.protobuf.json_format import ParseDict from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -83,7 +83,7 @@ def TABULAR_DATASET(bucket_name): VALIDATION_FRACTION_SPLIT = 0.15 -with models.DAG( +with DAG( f"{DAG_ID}_custom_container", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_job.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_job.py index ad611eb67a3f3..364cbad67334f 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_job.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_job.py @@ -31,7 +31,7 @@ from google.protobuf.json_format import ParseDict from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -79,7 +79,7 @@ def TABULAR_DATASET(bucket_name): LOCAL_TRAINING_SCRIPT_PATH = "california_housing_training_script.py" -with models.DAG( +with DAG( f"{DAG_ID}_custom", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_job_python_package.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_job_python_package.py index 99a1a734eb66a..e0e677c16a95f 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_job_python_package.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_custom_job_python_package.py @@ -31,7 +31,7 @@ from google.protobuf.json_format import ParseDict from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -85,7 +85,7 @@ def TABULAR_DATASET(bucket_name): PYTHON_MODULE_NAME = "aiplatform_custom_trainer_script.task" -with models.DAG( +with DAG( f"{DAG_ID}_python_package", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_dataset.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_dataset.py index 153788835be61..71717bcf03fa7 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_dataset.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_dataset.py @@ -31,7 +31,7 @@ from google.protobuf.json_format import ParseDict from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -111,7 +111,7 @@ TEST_UPDATE_MASK = {"paths": ["displayName"]} -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_endpoint.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_endpoint.py index 8be2357dd870a..b3a471776a715 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_endpoint.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_endpoint.py @@ -30,7 +30,7 @@ from google.cloud.aiplatform import schema from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -81,7 +81,7 @@ } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_hyperparameter_tuning_job.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_hyperparameter_tuning_job.py index 6239f263eec4e..9430eaa0b876e 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_hyperparameter_tuning_job.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_hyperparameter_tuning_job.py @@ -29,7 +29,7 @@ from google.cloud import aiplatform -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.vertex_ai.hyperparameter_tuning_job import ( CreateHyperparameterTuningJobOperator, @@ -76,7 +76,7 @@ } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_list_custom_jobs.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_list_custom_jobs.py index 1876139efc8bc..ab093392e3d37 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_list_custom_jobs.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_list_custom_jobs.py @@ -27,7 +27,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.vertex_ai.custom_job import ListCustomTrainingJobOperator ENV_ID = os.environ.get("SYSTEM_TESTS_ENV_ID") @@ -35,7 +35,7 @@ DAG_ID = "vertex_ai_custom_job_operations" REGION = "us-central1" -with models.DAG( +with DAG( f"{DAG_ID}_list_custom_job", schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_model_service.py b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_model_service.py index 98004acac25c8..fccc3a610fded 100644 --- a/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_model_service.py +++ b/tests/system/providers/google/cloud/vertex_ai/example_vertex_ai_model_service.py @@ -31,7 +31,7 @@ from google.protobuf.json_format import ParseDict from google.protobuf.struct_pb2 import Value -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import ( GCSCreateBucketOperator, GCSDeleteBucketOperator, @@ -106,7 +106,7 @@ } -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/video_intelligence/example_video_intelligence.py b/tests/system/providers/google/cloud/video_intelligence/example_video_intelligence.py index ab2289412a257..c76e1d3e723a8 100644 --- a/tests/system/providers/google/cloud/video_intelligence/example_video_intelligence.py +++ b/tests/system/providers/google/cloud/video_intelligence/example_video_intelligence.py @@ -31,8 +31,8 @@ from google.api_core.retry import Retry -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.video_intelligence import ( @@ -62,7 +62,7 @@ INPUT_URI = f"gs://{BUCKET_NAME_DST}/{FILE_NAME}" # [END howto_operator_video_intelligence_other_args] -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vision/example_vision_annotate_image.py b/tests/system/providers/google/cloud/vision/example_vision_annotate_image.py index 61d9fb2eca49c..69ec87bac4d8b 100644 --- a/tests/system/providers/google/cloud/vision/example_vision_annotate_image.py +++ b/tests/system/providers/google/cloud/vision/example_vision_annotate_image.py @@ -20,8 +20,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.vision import ( @@ -74,7 +74,7 @@ PATH_SRC = "vision/ocr/sign.jpg" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vision/example_vision_autogenerated.py b/tests/system/providers/google/cloud/vision/example_vision_autogenerated.py index e16f8dacbed37..35c5ee38dfb79 100644 --- a/tests/system/providers/google/cloud/vision/example_vision_autogenerated.py +++ b/tests/system/providers/google/cloud/vision/example_vision_autogenerated.py @@ -21,8 +21,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.vision import ( CloudVisionAddProductToProductSetOperator, @@ -106,7 +106,7 @@ # Path to the data inside the public bucket PATH_SRC = "vision/ocr/sign.jpg" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/vision/example_vision_explicit.py b/tests/system/providers/google/cloud/vision/example_vision_explicit.py index 134bf86f28388..40acf76745415 100644 --- a/tests/system/providers/google/cloud/vision/example_vision_explicit.py +++ b/tests/system/providers/google/cloud/vision/example_vision_explicit.py @@ -20,8 +20,8 @@ import os from datetime import datetime -from airflow import models from airflow.models.baseoperator import chain +from airflow.models.dag import DAG from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.cloud.operators.vision import ( CloudVisionAddProductToProductSetOperator, @@ -92,7 +92,7 @@ PATH_SRC = "vision/ocr/sign.jpg" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/cloud/workflows/example_workflows.py b/tests/system/providers/google/cloud/workflows/example_workflows.py index e1a0fd0f18b9e..c2c313d363272 100644 --- a/tests/system/providers/google/cloud/workflows/example_workflows.py +++ b/tests/system/providers/google/cloud/workflows/example_workflows.py @@ -22,7 +22,7 @@ from google.protobuf.field_mask_pb2 import FieldMask -from airflow import DAG +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.operators.workflows import ( WorkflowsCancelExecutionOperator, diff --git a/tests/system/providers/google/datacatalog/example_datacatalog_entries.py b/tests/system/providers/google/datacatalog/example_datacatalog_entries.py index cc194dce2bee6..cf4276d3a2dd5 100644 --- a/tests/system/providers/google/datacatalog/example_datacatalog_entries.py +++ b/tests/system/providers/google/datacatalog/example_datacatalog_entries.py @@ -22,7 +22,7 @@ from google.protobuf.field_mask_pb2 import FieldMask -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.datacatalog import ( @@ -50,7 +50,7 @@ ENTRY_ID = "python_files" ENTRY_NAME = "Wizard" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/datacatalog/example_datacatalog_search_catalog.py b/tests/system/providers/google/datacatalog/example_datacatalog_search_catalog.py index 762511e565f71..2beeebe54f568 100644 --- a/tests/system/providers/google/datacatalog/example_datacatalog_search_catalog.py +++ b/tests/system/providers/google/datacatalog/example_datacatalog_search_catalog.py @@ -23,7 +23,7 @@ from google.cloud.datacatalog import TagField, TagTemplateField -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.datacatalog import ( @@ -55,7 +55,7 @@ TAG_TEMPLATE_DISPLAY_NAME = f"Data Catalog {DAG_ID} {ENV_ID}" FIELD_NAME_1 = "first" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/datacatalog/example_datacatalog_tag_templates.py b/tests/system/providers/google/datacatalog/example_datacatalog_tag_templates.py index 7e8e336005612..24ba00b630fe4 100644 --- a/tests/system/providers/google/datacatalog/example_datacatalog_tag_templates.py +++ b/tests/system/providers/google/datacatalog/example_datacatalog_tag_templates.py @@ -22,7 +22,7 @@ from google.cloud.datacatalog import FieldType, TagTemplateField -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.datacatalog import ( @@ -49,7 +49,7 @@ FIELD_NAME_2 = "second" FIELD_NAME_3 = "first-rename" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/datacatalog/example_datacatalog_tags.py b/tests/system/providers/google/datacatalog/example_datacatalog_tags.py index dd8e2e1896be0..5d7f1cff3ed0c 100644 --- a/tests/system/providers/google/datacatalog/example_datacatalog_tags.py +++ b/tests/system/providers/google/datacatalog/example_datacatalog_tags.py @@ -23,7 +23,7 @@ from google.cloud.datacatalog import TagField, TagTemplateField -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.operators.bash import BashOperator from airflow.providers.google.cloud.operators.datacatalog import ( @@ -56,7 +56,7 @@ TAG_TEMPLATE_DISPLAY_NAME = f"Data Catalog {DAG_ID} {ENV_ID}" FIELD_NAME_1 = "first" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/leveldb/example_leveldb.py b/tests/system/providers/google/leveldb/example_leveldb.py index 2662c3894084e..31dd363f2c955 100644 --- a/tests/system/providers/google/leveldb/example_leveldb.py +++ b/tests/system/providers/google/leveldb/example_leveldb.py @@ -25,8 +25,8 @@ import pytest -from airflow import models from airflow.exceptions import AirflowOptionalProviderFeatureException +from airflow.models.dag import DAG try: from airflow.providers.google.leveldb.operators.leveldb import LevelDBOperator @@ -38,7 +38,7 @@ ENV_ID = os.environ.get("SYSTEM_TESTS_ENV_ID") DAG_ID = "example_leveldb" -with models.DAG( +with DAG( DAG_ID, start_date=datetime(2021, 1, 1), schedule="@once", diff --git a/tests/system/providers/google/marketing_platform/example_analytics.py b/tests/system/providers/google/marketing_platform/example_analytics.py index 88b2346c2d900..ce95f27297fe7 100644 --- a/tests/system/providers/google/marketing_platform/example_analytics.py +++ b/tests/system/providers/google/marketing_platform/example_analytics.py @@ -22,7 +22,7 @@ import os from datetime import datetime -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.marketing_platform.operators.analytics import ( GoogleAnalyticsDataImportUploadOperator, GoogleAnalyticsDeletePreviousDataUploadsOperator, @@ -43,7 +43,7 @@ WEB_PROPERTY_AD_WORDS_LINK_ID = os.environ.get("GA_WEB_PROPERTY_AD_WORDS_LINK_ID", "rQafFTPOQdmkx4U-fxUfhj") DATA_ID = "kjdDu3_tQa6n8Q1kXFtSmg" -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs, start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/marketing_platform/example_campaign_manager.py b/tests/system/providers/google/marketing_platform/example_campaign_manager.py index 9c8fd0a5b4867..61e4253d75ac4 100644 --- a/tests/system/providers/google/marketing_platform/example_campaign_manager.py +++ b/tests/system/providers/google/marketing_platform/example_campaign_manager.py @@ -26,7 +26,7 @@ from datetime import datetime from typing import cast -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator from airflow.providers.google.marketing_platform.operators.campaign_manager import ( @@ -107,7 +107,7 @@ "value": 123.4, } -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs, start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/marketing_platform/example_search_ads.py b/tests/system/providers/google/marketing_platform/example_search_ads.py index 1e4a6705a43e4..0f16e816c9cdc 100644 --- a/tests/system/providers/google/marketing_platform/example_search_ads.py +++ b/tests/system/providers/google/marketing_platform/example_search_ads.py @@ -24,7 +24,7 @@ from datetime import datetime from typing import cast -from airflow import models +from airflow.models.dag import DAG from airflow.models.xcom_arg import XComArg from airflow.providers.google.marketing_platform.operators.search_ads import ( GoogleSearchAdsDownloadReportOperator, @@ -51,7 +51,7 @@ } # [END howto_search_ads_env_variables] -with models.DAG( +with DAG( DAG_ID, schedule="@once", # Override to match your needs, start_date=datetime(2021, 1, 1), diff --git a/tests/system/providers/google/suite/example_local_to_drive.py b/tests/system/providers/google/suite/example_local_to_drive.py index 5a1f44e3fd751..7e5130f68bcde 100644 --- a/tests/system/providers/google/suite/example_local_to_drive.py +++ b/tests/system/providers/google/suite/example_local_to_drive.py @@ -26,7 +26,7 @@ from datetime import datetime from pathlib import Path -from airflow import models +from airflow.models.dag import DAG from airflow.providers.google.suite.transfers.local_to_drive import LocalFilesystemToGoogleDriveOperator DAG_ID = "example_local_to_drive" @@ -42,7 +42,7 @@ DRIVE_FOLDER = "test-folder" -with models.DAG( +with DAG( DAG_ID, schedule="@once", start_date=datetime(2021, 1, 1),