This file documents any backwards-incompatible changes in Airflow and assists users migrating to a new version.
Table of contents
- Airflow Master
- Airflow 1.10.5
- Airflow 1.10.4
- Airflow 1.10.3
- Airflow 1.10.2
- Airflow 1.10.1
- Airflow 1.10
- Airflow 1.9
- Airflow 1.8.1
- Airflow 1.8
- Airflow 1.7.1.2
The TriggerDagRunOperator now takes a conf
argument to which a dict can be provided as conf for the DagRun.
As a result, the python_callable
argument was removed. PR: apache#6317.
The change in GCP operators implies that GCP Hooks for those operators require now keyword parameters rather
than positional ones in all methods where project_id
is used. The methods throw an explanatory exception
in case they are called using positional parameters.
Hooks involved:
- PubSubHook
- MLEngineHook
Other GCP hooks are unaffected.
The fernet mechanism is enabled by default to increase the security of the default installation. In order to
restore the previous behavior, the user must consciously set an empty key in the fernet_key
option of
section [core]
in the airflow.cfg
file.
At the same time, this means that the apache-airflow[crypto]
extra-packages are always installed.
However, this requires that your operating system has libffi-dev
installed.
In the PubSubPublishOperator
and PubSubHook.publsh
method the data field in a message should be bytestring (utf-8 encoded) rather than base64 encoded string.
Due to the normalization of the parameters within GCP operators and hooks a parameters like project
or topic_project
are deprecated and will be substituted by parameter project_id
.
In PubSubHook.create_subscription
hook method in the parameter subscription_project
is replaced by subscription_project_id
.
Template fields are updated accordingly and old ones may not work.
It is required now to pass key-word only arguments to PubSub
hook.
These changes are not backward compatible.
Affected components:
- airflow.gcp.hooks.pubsub.PubSubHook
- airflow.gcp.operators.pubsub.PubSubTopicCreateOperator
- airflow.gcp.operators.pubsub.PubSubSubscriptionCreateOperator
- airflow.gcp.operators.pubsub.PubSubTopicDeleteOperator
- airflow.gcp.operators.pubsub.PubSubSubscriptionDeleteOperator
- airflow.gcp.operators.pubsub.PubSubPublishOperator
- airflow.gcp.sensors.pubsub.PubSubPullSensor
The region of Airflow's default connection to AWS (aws_default
) was previously
set to us-east-1
during installation.
The region now needs to be set manually, either in the connection screens in
Airflow, via the ~/.aws
config files, or via the AWS_DEFAULT_REGION
environment
variable.
Hipchat has reached end of life and is no longer available.
For more information please see https://community.atlassian.com/t5/Stride-articles/Stride-and-Hipchat-Cloud-have-reached-End-of-Life-updated/ba-p/940248
The following metrics are deprecated and won't be emitted in Airflow 2.0:
scheduler.dagbag.errors
anddagbag_import_errors
-- usedag_processing.import_errors
insteaddag_file_processor_timeouts
-- usedag_processing.processor_timeouts
insteadcollect_dags
-- usedag_processing.total_parse_time
insteaddag.loading-duration.<basename>
-- usedag_processing.last_duration.<basename>
insteaddag_processing.last_runtime.<basename>
-- usedag_processing.last_duration.<basename>
instead
In previous versions, it was possible to pass the None
value to the gcp_conn_id
in the GKEPodOperator
operator, which resulted in credentials being determined according to the
Application Default Credentials strategy.
Now this parameter requires a value. To restore the previous behavior, configure the connection without specifying the service account.
Detailed information about connection management is available: Google Cloud Platform Connection.
Previously not all hooks and operators related to Google Cloud Platform use
gcp_conn_id
as parameter for GCP connection. There is currently one parameter
which apply to most services. Parameters like datastore_conn_id
, bigquery_conn_id
,
google_cloud_storage_conn_id
and similar have been deprecated. Operators that require two connections are not changed.
Following components were affected by normalization:
- airflow.gcp.hooks.datastore.DatastoreHook
- airflow.gcp.hooks.bigquery.BigQueryHook
- airflow.gcp.hooks.gcs.GoogleCloudStorageHook
- airflow.gcp.operators.bigquery.BigQueryCheckOperator
- airflow.gcp.operators.bigquery.BigQueryValueCheckOperator
- airflow.gcp.operators.bigquery.BigQueryIntervalCheckOperator
- airflow.gcp.operators.bigquery.BigQueryGetDataOperator
- airflow.gcp.operators.bigquery.BigQueryOperator
- airflow.gcp.operators.bigquery.BigQueryDeleteDatasetOperator
- airflow.gcp.operators.bigquery.BigQueryCreateEmptyDatasetOperator
- airflow.gcp.operators.bigquery.BigQueryTableDeleteOperator
- airflow.gcp.operators.gcs.GoogleCloudStorageCreateBucketOperator
- airflow.gcp.operators.gcs.GoogleCloudStorageListOperator
- airflow.gcp.operators.gcs.GoogleCloudStorageDownloadOperator
- airflow.gcp.operators.gcs.GoogleCloudStorageDeleteOperator
- airflow.gcp.operators.gcs.GoogleCloudStorageBucketCreateAclEntryOperator
- airflow.gcp.operators.gcs.GoogleCloudStorageObjectCreateAclEntryOperator
- airflow.operators.sql_to_gcs.BaseSQLToGoogleCloudStorageOperator
- airflow.operators.adls_to_gcs.AdlsToGoogleCloudStorageOperator
- airflow.operators.gcs_to_s3.GoogleCloudStorageToS3Operator
- airflow.operators.gcs_to_gcs.GoogleCloudStorageToGoogleCloudStorageOperator
- airflow.operators.bigquery_to_gcs.BigQueryToCloudStorageOperator
- airflow.operators.local_to_gcs.FileToGoogleCloudStorageOperator
- airflow.operators.cassandra_to_gcs.CassandraToGoogleCloudStorageOperator
- airflow.operators.bigquery_to_bigquery.BigQueryToBigQueryOperator
kubernetes_annotations
configuration section has been removed.
A new key worker_annotations
has been added to existing kubernetes
section instead.
That is to remove restriction on the character set for k8s annotation keys.
All key/value pairs from kubernetes_annotations
should now go to worker_annotations
as a json. I.e. instead of e.g.
[kubernetes_annotations]
annotation_key = annotation_value
annotation_key2 = annotation_value2
it should be rewritten to
[kubernetes]
worker_annotations = { "annotation_key" : "annotation_value", "annotation_key2" : "annotation_value2" }
According to AIP-21 operators related to Google Cloud Platform has been moved from contrib to core. The following table shows changes in import paths.
Old path | New path |
---|---|
airflow.contrib.hooks.bigquery_hook.BigQueryHook | airflow.gcp.hooks.bigquery.BigQueryHook |
airflow.contrib.hooks.datastore_hook.DatastoreHook | airflow.gcp.hooks.datastore.DatastoreHook |
airflow.contrib.hooks.gcp_bigtable_hook.BigtableHook | airflow.gcp.hooks.bigtable.BigtableHook |
airflow.contrib.hooks.gcp_cloud_build_hook.CloudBuildHook | airflow.gcp.hooks.cloud_build.CloudBuildHook |
airflow.contrib.hooks.gcp_compute_hook.GceHook | airflow.gcp.hooks.compute.ComputeEngineHook |
airflow.contrib.hooks.gcp_container_hook.GKEClusterHook | airflow.gcp.hooks.kubernetes_engine.GKEClusterHook |
airflow.contrib.hooks.gcp_dataflow_hook.DataFlowHook | airflow.gcp.hooks.dataflow.DataFlowHook |
airflow.contrib.hooks.gcp_dataproc_hook.DataProcHook | airflow.gcp.hooks.dataproc.DataProcHook |
airflow.contrib.hooks.gcp_dlp_hook.CloudDLPHook | airflow.gcp.hooks.dlp.CloudDLPHook |
airflow.contrib.hooks.gcp_function_hook.GcfHook | airflow.gcp.hooks.functions.GcfHook |
airflow.contrib.hooks.gcp_kms_hook.GoogleCloudKMSHook | airflow.gcp.hooks.kms.GoogleCloudKMSHook |
airflow.contrib.hooks.gcp_mlengine_hook.MLEngineHook | airflow.gcp.hooks.mlengine.MLEngineHook |
airflow.contrib.hooks.gcp_natural_language_hook.CloudNaturalLanguageHook | airflow.gcp.hooks.natural_language.CloudNaturalLanguageHook |
airflow.contrib.hooks.gcp_pubsub_hook.PubSubHook | airflow.gcp.hooks.pubsub.PubSubHook |
airflow.contrib.hooks.gcp_spanner_hook.CloudSpannerHook | airflow.gcp.hooks.spanner.CloudSpannerHook |
airflow.contrib.hooks.gcp_speech_to_text_hook.GCPSpeechToTextHook | airflow.gcp.hooks.speech_to_text.GCPSpeechToTextHook |
airflow.contrib.hooks.gcp_sql_hook.CloudSqlDatabaseHook | airflow.gcp.hooks.cloud_sql.CloudSqlDatabaseHook |
airflow.contrib.hooks.gcp_sql_hook.CloudSqlHook | airflow.gcp.hooks.cloud_sql.CloudSqlHook |
airflow.contrib.hooks.gcp_tasks_hook.CloudTasksHook | airflow.gcp.hooks.tasks.CloudTasksHook |
airflow.contrib.hooks.gcp_text_to_speech_hook.GCPTextToSpeechHook | airflow.gcp.hooks.text_to_speech.GCPTextToSpeechHook |
airflow.contrib.hooks.gcp_transfer_hook.GCPTransferServiceHook | airflow.gcp.hooks.cloud_storage_transfer_service.GCPTransferServiceHook |
airflow.contrib.hooks.gcp_translate_hook.CloudTranslateHook | airflow.gcp.hooks.translate.CloudTranslateHook |
airflow.contrib.hooks.gcp_video_intelligence_hook.CloudVideoIntelligenceHook | airflow.gcp.hooks.video_intelligence.CloudVideoIntelligenceHook |
airflow.contrib.hooks.gcp_vision_hook.CloudVisionHook | airflow.gcp.hooks.vision.CloudVisionHook |
airflow.contrib.hooks.gcs_hook.GoogleCloudStorageHook | airflow.gcp.hooks.gcs.GoogleCloudStorageHook |
airflow.contrib.operators.adls_to_gcs.AdlsToGoogleCloudStorageOperator | airflow.operators.adls_to_gcs.AdlsToGoogleCloudStorageOperator |
airflow.contrib.operators.bigquery_check_operator.BigQueryCheckOperator | airflow.gcp.operators.bigquery.BigQueryCheckOperator |
airflow.contrib.operators.bigquery_check_operator.BigQueryIntervalCheckOperator | airflow.gcp.operators.bigquery.BigQueryIntervalCheckOperator |
airflow.contrib.operators.bigquery_check_operator.BigQueryValueCheckOperator | airflow.gcp.operators.bigquery.BigQueryValueCheckOperator |
airflow.contrib.operators.bigquery_get_data.BigQueryGetDataOperator | airflow.gcp.operators.bigquery.BigQueryGetDataOperator |
airflow.contrib.operators.bigquery_operator.BigQueryCreateEmptyDatasetOperator | airflow.gcp.operators.bigquery.BigQueryCreateEmptyDatasetOperator |
airflow.contrib.operators.bigquery_operator.BigQueryCreateEmptyTableOperator | airflow.gcp.operators.bigquery.BigQueryCreateEmptyTableOperator |
airflow.contrib.operators.bigquery_operator.BigQueryCreateExternalTableOperator | airflow.gcp.operators.bigquery.BigQueryCreateExternalTableOperator |
airflow.contrib.operators.bigquery_operator.BigQueryDeleteDatasetOperator | airflow.gcp.operators.bigquery.BigQueryDeleteDatasetOperator |
airflow.contrib.operators.bigquery_operator.BigQueryOperator | airflow.gcp.operators.bigquery.BigQueryOperator |
airflow.contrib.operators.bigquery_table_delete_operator.BigQueryTableDeleteOperator | airflow.gcp.operators.bigquery.BigQueryTableDeleteOperator |
airflow.contrib.operators.bigquery_to_bigquery.BigQueryToBigQueryOperator | airflow.operators.bigquery_to_bigquery.BigQueryToBigQueryOperator |
airflow.contrib.operators.bigquery_to_gcs.BigQueryToCloudStorageOperator | airflow.operators.bigquery_to_gcs.BigQueryToCloudStorageOperator |
airflow.contrib.operators.bigquery_to_mysql_operator.BigQueryToMySqlOperator | airflow.operators.bigquery_to_mysql.BigQueryToMySqlOperator |
airflow.contrib.operators.dataflow_operator.DataFlowJavaOperator | airflow.gcp.operators.dataflow.DataFlowJavaOperator |
airflow.contrib.operators.dataflow_operator.DataFlowPythonOperator | airflow.gcp.operators.dataflow.DataFlowPythonOperator |
airflow.contrib.operators.dataflow_operator.DataflowTemplateOperator | airflow.gcp.operators.dataflow.DataflowTemplateOperator |
airflow.contrib.operators.dataproc_operator.DataProcHadoopOperator | airflow.gcp.operators.dataproc.DataProcHadoopOperator |
airflow.contrib.operators.dataproc_operator.DataProcHiveOperator | airflow.gcp.operators.dataproc.DataProcHiveOperator |
airflow.contrib.operators.dataproc_operator.DataProcJobBaseOperator | airflow.gcp.operators.dataproc.DataProcJobBaseOperator |
airflow.contrib.operators.dataproc_operator.DataProcPigOperator | airflow.gcp.operators.dataproc.DataProcPigOperator |
airflow.contrib.operators.dataproc_operator.DataProcPySparkOperator | airflow.gcp.operators.dataproc.DataProcPySparkOperator |
airflow.contrib.operators.dataproc_operator.DataProcSparkOperator | airflow.gcp.operators.dataproc.DataProcSparkOperator |
airflow.contrib.operators.dataproc_operator.DataProcSparkSqlOperator | airflow.gcp.operators.dataproc.DataProcSparkSqlOperator |
airflow.contrib.operators.dataproc_operator.DataprocClusterCreateOperator | airflow.gcp.operators.dataproc.DataprocClusterCreateOperator |
airflow.contrib.operators.dataproc_operator.DataprocClusterDeleteOperator | airflow.gcp.operators.dataproc.DataprocClusterDeleteOperator |
airflow.contrib.operators.dataproc_operator.DataprocClusterScaleOperator | airflow.gcp.operators.dataproc.DataprocClusterScaleOperator |
airflow.contrib.operators.dataproc_operator.DataprocOperationBaseOperator | airflow.gcp.operators.dataproc.DataprocOperationBaseOperator |
airflow.contrib.operators.dataproc_operator.DataprocWorkflowTemplateInstantiateInlineOperator | airflow.gcp.operators.dataproc.DataprocWorkflowTemplateInstantiateInlineOperator |
airflow.contrib.operators.dataproc_operator.DataprocWorkflowTemplateInstantiateOperator | airflow.gcp.operators.dataproc.DataprocWorkflowTemplateInstantiateOperator |
airflow.contrib.operators.datastore_export_operator.DatastoreExportOperator | airflow.gcp.operators.datastore.DatastoreExportOperator |
airflow.contrib.operators.datastore_import_operator.DatastoreImportOperator | airflow.gcp.operators.datastore.DatastoreImportOperator |
airflow.contrib.operators.file_to_gcs.FileToGoogleCloudStorageOperator | airflow.operators.local_to_gcs.FileToGoogleCloudStorageOperator |
airflow.contrib.operators.gcp_bigtable_operator.BigtableClusterUpdateOperator | airflow.gcp.operators.bigtable.BigtableClusterUpdateOperator |
airflow.contrib.operators.gcp_bigtable_operator.BigtableInstanceCreateOperator | airflow.gcp.operators.bigtable.BigtableInstanceCreateOperator |
airflow.contrib.operators.gcp_bigtable_operator.BigtableInstanceDeleteOperator | airflow.gcp.operators.bigtable.BigtableInstanceDeleteOperator |
airflow.contrib.operators.gcp_bigtable_operator.BigtableTableCreateOperator | airflow.gcp.operators.bigtable.BigtableTableCreateOperator |
airflow.contrib.operators.gcp_bigtable_operator.BigtableTableDeleteOperator | airflow.gcp.operators.bigtable.BigtableTableDeleteOperator |
airflow.contrib.operators.gcp_bigtable_operator.BigtableTableWaitForReplicationSensor | airflow.gcp.sensors.bigtable.BigtableTableWaitForReplicationSensor |
airflow.contrib.operators.gcp_cloud_build_operator.CloudBuildCreateBuildOperator | airflow.gcp.operators.cloud_build.CloudBuildCreateBuildOperator |
airflow.contrib.operators.gcp_compute_operator.GceBaseOperator | airflow.gcp.operators.compute.GceBaseOperator |
airflow.contrib.operators.gcp_compute_operator.GceInstanceGroupManagerUpdateTemplateOperator | airflow.gcp.operators.compute.GceInstanceGroupManagerUpdateTemplateOperator |
airflow.contrib.operators.gcp_compute_operator.GceInstanceStartOperator | airflow.gcp.operators.compute.GceInstanceStartOperator |
airflow.contrib.operators.gcp_compute_operator.GceInstanceStopOperator | airflow.gcp.operators.compute.GceInstanceStopOperator |
airflow.contrib.operators.gcp_compute_operator.GceInstanceTemplateCopyOperator | airflow.gcp.operators.compute.GceInstanceTemplateCopyOperator |
airflow.contrib.operators.gcp_compute_operator.GceSetMachineTypeOperator | airflow.gcp.operators.compute.GceSetMachineTypeOperator |
airflow.contrib.operators.gcp_container_operator.GKEClusterCreateOperator | airflow.gcp.operators.kubernetes_engine.GKEClusterCreateOperator |
airflow.contrib.operators.gcp_container_operator.GKEClusterDeleteOperator | airflow.gcp.operators.kubernetes_engine.GKEClusterDeleteOperator |
airflow.contrib.operators.gcp_container_operator.GKEPodOperator | airflow.gcp.operators.kubernetes_engine.GKEPodOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCancelDLPJobOperator | airflow.gcp.operators.dlp.CloudDLPCancelDLPJobOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateDLPJobOperator | airflow.gcp.operators.dlp.CloudDLPCreateDLPJobOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateDeidentifyTemplateOperator | airflow.gcp.operators.dlp.CloudDLPCreateDeidentifyTemplateOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateInspectTemplateOperator | airflow.gcp.operators.dlp.CloudDLPCreateInspectTemplateOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateJobTriggerOperator | airflow.gcp.operators.dlp.CloudDLPCreateJobTriggerOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateStoredInfoTypeOperator | airflow.gcp.operators.dlp.CloudDLPCreateStoredInfoTypeOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeidentifyContentOperator | airflow.gcp.operators.dlp.CloudDLPDeidentifyContentOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteDeidentifyTemplateOperator | airflow.gcp.operators.dlp.CloudDLPDeleteDeidentifyTemplateOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteDlpJobOperator | airflow.gcp.operators.dlp.CloudDLPDeleteDlpJobOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteInspectTemplateOperator | airflow.gcp.operators.dlp.CloudDLPDeleteInspectTemplateOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteJobTriggerOperator | airflow.gcp.operators.dlp.CloudDLPDeleteJobTriggerOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteStoredInfoTypeOperator | airflow.gcp.operators.dlp.CloudDLPDeleteStoredInfoTypeOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetDeidentifyTemplateOperator | airflow.gcp.operators.dlp.CloudDLPGetDeidentifyTemplateOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetDlpJobOperator | airflow.gcp.operators.dlp.CloudDLPGetDlpJobOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetInspectTemplateOperator | airflow.gcp.operators.dlp.CloudDLPGetInspectTemplateOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetJobTripperOperator | airflow.gcp.operators.dlp.CloudDLPGetJobTripperOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetStoredInfoTypeOperator | airflow.gcp.operators.dlp.CloudDLPGetStoredInfoTypeOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPInspectContentOperator | airflow.gcp.operators.dlp.CloudDLPInspectContentOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListDeidentifyTemplatesOperator | airflow.gcp.operators.dlp.CloudDLPListDeidentifyTemplatesOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListDlpJobsOperator | airflow.gcp.operators.dlp.CloudDLPListDlpJobsOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListInfoTypesOperator | airflow.gcp.operators.dlp.CloudDLPListInfoTypesOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListInspectTemplatesOperator | airflow.gcp.operators.dlp.CloudDLPListInspectTemplatesOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListJobTriggersOperator | airflow.gcp.operators.dlp.CloudDLPListJobTriggersOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListStoredInfoTypesOperator | airflow.gcp.operators.dlp.CloudDLPListStoredInfoTypesOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPRedactImageOperator | airflow.gcp.operators.dlp.CloudDLPRedactImageOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPReidentifyContentOperator | airflow.gcp.operators.dlp.CloudDLPReidentifyContentOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPUpdateDeidentifyTemplateOperator | airflow.gcp.operators.dlp.CloudDLPUpdateDeidentifyTemplateOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPUpdateInspectTemplateOperator | airflow.gcp.operators.dlp.CloudDLPUpdateInspectTemplateOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPUpdateJobTriggerOperator | airflow.gcp.operators.dlp.CloudDLPUpdateJobTriggerOperator |
airflow.contrib.operators.gcp_dlp_operator.CloudDLPUpdateStoredInfoTypeOperator | airflow.gcp.operators.dlp.CloudDLPUpdateStoredInfoTypeOperator |
airflow.contrib.operators.gcp_function_operator.GcfFunctionDeleteOperator | airflow.gcp.operators.functions.GcfFunctionDeleteOperator |
airflow.contrib.operators.gcp_function_operator.GcfFunctionDeployOperator | airflow.gcp.operators.functions.GcfFunctionDeployOperator |
airflow.contrib.operators.gcp_natural_language_operator.CloudLanguageAnalyzeEntitiesOperator | airflow.gcp.operators.natural_language.CloudLanguageAnalyzeEntitiesOperator |
airflow.contrib.operators.gcp_natural_language_operator.CloudLanguageAnalyzeEntitySentimentOperator | airflow.gcp.operators.natural_language.CloudLanguageAnalyzeEntitySentimentOperator |
airflow.contrib.operators.gcp_natural_language_operator.CloudLanguageAnalyzeSentimentOperator | airflow.gcp.operators.natural_language.CloudLanguageAnalyzeSentimentOperator |
airflow.contrib.operators.gcp_natural_language_operator.CloudLanguageClassifyTextOperator | airflow.gcp.operators.natural_language.CloudLanguageClassifyTextOperator |
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDatabaseDeleteOperator | airflow.gcp.operators.spanner.CloudSpannerInstanceDatabaseDeleteOperator |
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDatabaseDeployOperator | airflow.gcp.operators.spanner.CloudSpannerInstanceDatabaseDeployOperator |
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDatabaseQueryOperator | airflow.gcp.operators.spanner.CloudSpannerInstanceDatabaseQueryOperator |
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDatabaseUpdateOperator | airflow.gcp.operators.spanner.CloudSpannerInstanceDatabaseUpdateOperator |
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDeleteOperator | airflow.gcp.operators.spanner.CloudSpannerInstanceDeleteOperator |
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDeployOperator | airflow.gcp.operators.spanner.CloudSpannerInstanceDeployOperator |
airflow.contrib.operators.gcp_speech_to_text_operator.GcpSpeechToTextRecognizeSpeechOperator | airflow.gcp.operators.speech_to_text.GcpSpeechToTextRecognizeSpeechOperator |
airflow.contrib.operators.gcp_text_to_speech_operator.GcpTextToSpeechSynthesizeOperator | airflow.gcp.operators.text_to_speech.GcpTextToSpeechSynthesizeOperator |
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceJobCreateOperator | airflow.gcp.operators.cloud_storage_transfer_service.GcpTransferServiceJobCreateOperator |
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceJobDeleteOperator | airflow.gcp.operators.cloud_storage_transfer_service.GcpTransferServiceJobDeleteOperator |
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceJobUpdateOperator | airflow.gcp.operators.cloud_storage_transfer_service.GcpTransferServiceJobUpdateOperator |
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationCancelOperator | airflow.gcp.operators.cloud_storage_transfer_service.GcpTransferServiceOperationCancelOperator |
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationGetOperator | airflow.gcp.operators.cloud_storage_transfer_service.GcpTransferServiceOperationGetOperator |
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationPauseOperator | airflow.gcp.operators.cloud_storage_transfer_service.GcpTransferServiceOperationPauseOperator |
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationResumeOperator | airflow.gcp.operators.cloud_storage_transfer_service.GcpTransferServiceOperationResumeOperator |
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationsListOperator | airflow.gcp.operators.cloud_storage_transfer_service.GcpTransferServiceOperationsListOperator |
airflow.contrib.operators.gcp_transfer_operator.GoogleCloudStorageToGoogleCloudStorageTransferOperator | airflow.gcp.operators.cloud_storage_transfer_service.GoogleCloudStorageToGoogleCloudStorageTransferOperator |
airflow.contrib.operators.gcp_translate_operator.CloudTranslateTextOperator | airflow.gcp.operators.translate.CloudTranslateTextOperator |
airflow.contrib.operators.gcp_translate_speech_operator.GcpTranslateSpeechOperator | airflow.gcp.operators.translate_speech.GcpTranslateSpeechOperator |
airflow.contrib.operators.gcp_video_intelligence_operator.CloudVideoIntelligenceDetectVideoExplicitContentOperator | airflow.gcp.operators.video_intelligence.CloudVideoIntelligenceDetectVideoExplicitContentOperator |
airflow.contrib.operators.gcp_video_intelligence_operator.CloudVideoIntelligenceDetectVideoLabelsOperator | airflow.gcp.operators.video_intelligence.CloudVideoIntelligenceDetectVideoLabelsOperator |
airflow.contrib.operators.gcp_video_intelligence_operator.CloudVideoIntelligenceDetectVideoShotsOperator | airflow.gcp.operators.video_intelligence.CloudVideoIntelligenceDetectVideoShotsOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionAddProductToProductSetOperator | airflow.gcp.operators.vision.CloudVisionAddProductToProductSetOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionAnnotateImageOperator | airflow.gcp.operators.vision.CloudVisionAnnotateImageOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionDetectDocumentTextOperator | airflow.gcp.operators.vision.CloudVisionDetectDocumentTextOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionDetectImageLabelsOperator | airflow.gcp.operators.vision.CloudVisionDetectImageLabelsOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionDetectImageSafeSearchOperator | airflow.gcp.operators.vision.CloudVisionDetectImageSafeSearchOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionDetectTextOperator | airflow.gcp.operators.vision.CloudVisionDetectTextOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductCreateOperator | airflow.gcp.operators.vision.CloudVisionProductCreateOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductDeleteOperator | airflow.gcp.operators.vision.CloudVisionProductDeleteOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductGetOperator | airflow.gcp.operators.vision.CloudVisionProductGetOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductSetCreateOperator | airflow.gcp.operators.vision.CloudVisionProductSetCreateOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductSetDeleteOperator | airflow.gcp.operators.vision.CloudVisionProductSetDeleteOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductSetGetOperator | airflow.gcp.operators.vision.CloudVisionProductSetGetOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductSetUpdateOperator | airflow.gcp.operators.vision.CloudVisionProductSetUpdateOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductUpdateOperator | airflow.gcp.operators.vision.CloudVisionProductUpdateOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionReferenceImageCreateOperator | airflow.gcp.operators.vision.CloudVisionReferenceImageCreateOperator |
airflow.contrib.operators.gcp_vision_operator.CloudVisionRemoveProductFromProductSetOperator | airflow.gcp.operators.vision.CloudVisionRemoveProductFromProductSetOperator |
airflow.contrib.operators.gcs_acl_operator.GoogleCloudStorageBucketCreateAclEntryOperator | airflow.gcp.operators.gcs.GoogleCloudStorageBucketCreateAclEntryOperator |
airflow.contrib.operators.gcs_acl_operator.GoogleCloudStorageObjectCreateAclEntryOperator | airflow.gcp.operators.gcs.GoogleCloudStorageObjectCreateAclEntryOperator |
airflow.contrib.operators.gcs_delete_operator.GoogleCloudStorageDeleteOperator | airflow.gcp.operators.gcs.GoogleCloudStorageDeleteOperator |
airflow.contrib.operators.gcs_download_operator.GoogleCloudStorageDownloadOperator | airflow.gcp.operators.gcs.GoogleCloudStorageDownloadOperator |
airflow.contrib.operators.gcs_list_operator.GoogleCloudStorageListOperator | airflow.gcp.operators.gcs.GoogleCloudStorageListOperator |
airflow.contrib.operators.gcs_operator.GoogleCloudStorageCreateBucketOperator | airflow.gcp.operators.gcs.GoogleCloudStorageCreateBucketOperator |
airflow.contrib.operators.gcs_to_bq.GoogleCloudStorageToBigQueryOperator | airflow.operators.gcs_to_bq.GoogleCloudStorageToBigQueryOperator |
airflow.contrib.operators.gcs_to_gcs.GoogleCloudStorageToGoogleCloudStorageOperator | airflow.operators.gcs_to_gcs.GoogleCloudStorageToGoogleCloudStorageOperator |
airflow.contrib.operators.gcs_to_s3.GoogleCloudStorageToS3Operator | airflow.operators.gcs_to_s3.GoogleCloudStorageToS3Operator |
airflow.contrib.operators.mlengine_operator.MLEngineBatchPredictionOperator | airflow.gcp.operators.mlengine.MLEngineBatchPredictionOperator |
airflow.contrib.operators.mlengine_operator.MLEngineModelOperator | airflow.gcp.operators.mlengine.MLEngineModelOperator |
airflow.contrib.operators.mlengine_operator.MLEngineTrainingOperator | airflow.gcp.operators.mlengine.MLEngineTrainingOperator |
airflow.contrib.operators.mlengine_operator.MLEngineVersionOperator | airflow.gcp.operators.mlengine.MLEngineVersionOperator |
airflow.contrib.operators.mssql_to_gcs.MsSqlToGoogleCloudStorageOperator | airflow.operators.mssql_to_gcs.MsSqlToGoogleCloudStorageOperator |
airflow.contrib.operators.mysql_to_gcs.MySqlToGoogleCloudStorageOperator | airflow.operators.mysql_to_gcs.MySqlToGoogleCloudStorageOperator |
airflow.contrib.operators.postgres_to_gcs_operator.PostgresToGoogleCloudStorageOperator | airflow.operators.postgres_to_gcs.PostgresToGoogleCloudStorageOperator |
airflow.contrib.operators.pubsub_operator.PubSubPublishOperator | airflow.gcp.operators.pubsub.PubSubPublishOperator |
airflow.contrib.operators.pubsub_operator.PubSubSubscriptionCreateOperator | airflow.gcp.operators.pubsub.PubSubSubscriptionCreateOperator |
airflow.contrib.operators.pubsub_operator.PubSubSubscriptionDeleteOperator | airflow.gcp.operators.pubsub.PubSubSubscriptionDeleteOperator |
airflow.contrib.operators.pubsub_operator.PubSubTopicCreateOperator | airflow.gcp.operators.pubsub.PubSubTopicCreateOperator |
airflow.contrib.operators.pubsub_operator.PubSubTopicDeleteOperator | airflow.gcp.operators.pubsub.PubSubTopicDeleteOperator |
airflow.contrib.operators.sql_to_gcs.BaseSQLToGoogleCloudStorageOperator | airflow.operators.sql_to_gcs.BaseSQLToGoogleCloudStorageOperator |
airflow.contrib.sensors.bigquery_sensor.BigQueryTableSensor | airflow.gcp.sensors.bigquery.BigQueryTableSensor |
airflow.contrib.sensors.gcp_transfer_sensor.GCPTransferServiceWaitForJobStatusSensor | airflow.gcp.sensors.cloud_storage_transfer_service.GCPTransferServiceWaitForJobStatusSensor |
airflow.contrib.sensors.gcs_sensor.GoogleCloudStorageObjectSensor | airflow.gcp.sensors.gcs.GoogleCloudStorageObjectSensor |
airflow.contrib.sensors.gcs_sensor.GoogleCloudStorageObjectUpdatedSensor | airflow.gcp.sensors.gcs.GoogleCloudStorageObjectUpdatedSensor |
airflow.contrib.sensors.gcs_sensor.GoogleCloudStoragePrefixSensor | airflow.gcp.sensors.gcs.GoogleCloudStoragePrefixSensor |
airflow.contrib.sensors.gcs_sensor.GoogleCloudStorageUploadSessionCompleteSensor | airflow.gcp.sensors.gcs.GoogleCloudStorageUploadSessionCompleteSensor |
airflow.contrib.sensors.pubsub_sensor.PubSubPullSensor | airflow.gcp.sensors.pubsub.PubSubPullSensor |
provide_context
argument on the PythonOperator was removed. The signature of the callable passed to the PythonOperator is now inferred and argument values are always automatically provided. There is no need to explicitly provide or not provide the context anymore. For example:
def myfunc(execution_date):
print(execution_date)
python_operator = PythonOperator(task_id='mytask', python_callable=myfunc, dag=dag)
Notice you don't have to set provide_context=True, variables from the task context are now automatically detected and provided.
All context variables can still be provided with a double-asterisk argument:
def myfunc(**context):
print(context) # all variables will be provided to context
python_operator = PythonOperator(task_id='mytask', python_callable=myfunc)
The task context variable names are reserved names in the callable function, hence a clash with op_args
and op_kwargs
results in an exception:
def myfunc(dag):
# raises a ValueError because "dag" is a reserved name
# valid signature example: myfunc(mydag)
python_operator = PythonOperator(
task_id='mytask',
op_args=[1],
python_callable=myfunc,
)
The change is backwards compatible, setting provide_context
will add the provide_context
variable to the kwargs
(but won't do anything).
PR: #5990
FileSensor is now takes a glob pattern, not just a filename. If the filename you are looking for has *
, ?
, or [
in it then you should replace these with [*]
, [?]
, and [[]
.
Change DAG file loading duration metric from
dag.loading-duration.<dag_id>
to dag.loading-duration.<dag_file>
. This is to
better handle the case when a DAG file has multiple DAGs.
ImapHook:
- The order of arguments has changed for
has_mail_attachment
,retrieve_mail_attachments
anddownload_mail_attachments
. - A new
mail_filter
argument has been added to each of those.
ImapAttachmentSensor:
- The order of arguments has changed for
__init__
. - A new
mail_filter
argument has been added to__init__
.
ImapAttachmentToS3Operator:
- The order of arguments has changed for
__init__
. - A new
imap_mail_filter
argument has been added to__init__
.
SubDagOperator
is changed to use Airflow scheduler instead of backfill
to schedule tasks in the subdag. User no longer need to specify the executor
in SubDagOperator
.
The following variables were removed from the task instance context:
- end_date
- latest_date
- tables
To simplify the code, the decorator has been moved from the inner-class.
Instead of @GoogleCloudBaseHook._Decorators.provide_gcp_credential_file
,
you should write @GoogleCloudBaseHook.provide_gcp_credential_file
Note: The order of arguments has changed for check_for_prefix
.
The bucket_name
is now optional. It falls back to the connection schema
attribute.
To obtain pylint compatibility the filter
argument in GcpTransferServiceOperationsListOperator
has been renamed to request_filter
.
To obtain pylint compatibility the filter
argument in GCPTransferServiceHook.list_transfer_job
and
GCPTransferServiceHook.list_transfer_operations
has been renamed to request_filter
.
MySqlToGoogleCloudStorageOperator
now exports TIMESTAMP columns as UTC
by default, rather than using the default timezone of the MySQL server.
This is the correct behavior for use with BigQuery, since BigQuery
assumes that TIMESTAMP columns without time zones are in UTC. To
preserve the previous behavior, set ensure_utc
to False.
The Airflow CLI has been organized so that related commands are grouped
together as subcommands. The airflow list_dags
command is now airflow dags list
, airflow pause
is airflow dags pause
, etc. For a complete
list of updated CLI commands, see https://airflow.apache.org/cli.html.
The Mesos Executor is removed from the code base as it was not widely used and not maintained. Mailing List Discussion on deleting it.
It is highly recommended to have 1TB+ disk size for Dataproc to have sufficient throughput: https://cloud.google.com/compute/docs/disks/performance
Hence, the default value for master_disk_size
in DataprocClusterCreateOperator has beeen changes from 500GB to 1TB.
- renamed
sign_in
function toget_conn
The HTTPHook is now secured by default: verify=True
.
This can be overwriten by using the extra_options param as {'verify': False}
.
-
The following parameters have been replaced in all the methods in GCSHook:
bucket
is changed tobucket_name
object
is changed toobject_name
-
The
maxResults
parameter inGoogleCloudStorageHook.list
has been renamed tomax_results
for consistency.
- upgraded cloudant version from
>=0.5.9,<2.0
to>=2.0
- removed the use of the
schema
attribute in the connection - removed
db
function since the database object can also be retrieved by callingcloudant_session['database_name']
For example:
from airflow.contrib.hooks.cloudant_hook import CloudantHook
with CloudantHook().get_conn() as cloudant_session:
database = cloudant_session['database_name']
See the docs for more information on how to use the new cloudant version.
Previously not all hooks and operators related to Google Cloud Platform use
google_cloud_default
as a default conn_id. There is currently one default
variant. Values like google_cloud_storage_default
, bigquery_default
,
google_cloud_datastore_default
have been deprecated. The configuration of
existing relevant connections in the database have been preserved. To use those
deprecated GCP conn_id, you need to explicitly pass their conn_id into
operators/hooks. Otherwise, google_cloud_default
will be used as GCP's conn_id
by default.
If dag_discovery_safe_mode
is enabled, only check files for DAGs if
they contain the strings "airflow" and "DAG". For backwards
compatibility, this option is enabled by default.
The deprecated import mechanism has been removed so the import of modules becomes more consistent and explicit.
For example: from airflow.operators import BashOperator
becomes from airflow.operators.bash_operator import BashOperator
Sensors are now accessible via airflow.sensors
and no longer via airflow.operators.sensors
.
For example: from airflow.operators.sensors import BaseSensorOperator
becomes from airflow.sensors.base_sensor_operator import BaseSensorOperator
Subpackages for specific services have been combined into one variant for each cloud provider. The name of the subpackage for the Google Cloud Platform has changed to follow style.
If you want to install integration for Microsoft Azure, then instead of
pip install 'apache-airflow[azure_blob_storage,azure_data_lake,azure_cosmos,azure_container_instances]'
you should execute pip install 'apache-airflow[azure]'
If you want to install integration for Amazon Web Services, then instead of
pip install 'apache-airflow[s3,emr]'
, you should execute pip install 'apache-airflow[aws]'
If you want to install integration for Google Cloud Platform, then instead of
pip install 'apache-airflow[gcp_api]'
, you should execute pip install 'apache-airflow[gcp]'
.
The old way will work until the release of Airflow 2.1.
Previously we were using two versions of UI, which were hard to maintain as we need to implement/update the same feature in both versions. With this change we've removed the older UI in favor of Flask App Builder RBAC UI. No need to set the RBAC UI explicitly in the configuration now as this is the only default UI. Please note that that custom auth backends will need re-writing to target new FAB based UI.
As part of this change, a few configuration items in [webserver]
section are removed and no longer applicable,
including authenticate
, filter_by_owner
, owner_mode
, and rbac
.
We should not use the run_duration
option anymore. This used to be for restarting the scheduler from time to time, but right now the scheduler is getting more stable and therefore using this setting is considered bad and might cause an inconsistent state.
The DAG parsing manager log now by default will be log into a file, where its location is
controlled by the new dag_processor_manager_log_location
config option in core section.
The scheduler.min_file_parsing_loop_time config option has been temporarily removed due to some bugs.
The ability to manipulate users from the command line has been changed. 'airflow create_user' and 'airflow delete_user' and 'airflow list_users' has been grouped to a single command airflow users
with optional flags --create
, --list
and --delete
.
Example Usage:
To create a new user:
airflow users --create --username jondoe --lastname doe --firstname jon --email [email protected] --role Viewer --password test
To list users:
airflow users --list
To delete a user:
airflow users --delete --username jondoe
To add a user to a role:
airflow users --add-role --username jondoe --role Public
To remove a user from a role:
airflow users --remove-role --username jondoe --role Public
The do_xcom_push
flag (a switch to push the result of an operator to xcom or not) was appearing in different incarnations in different operators. It's function has been unified under a common name (do_xcom_push
) on BaseOperator
. This way it is also easy to globally disable pushing results to xcom.
See AIRFLOW-3249 to check if your operator was affected.
The 'properties' and 'jars' properties for the Dataproc related operators (DataprocXXXOperator
) have been renamed from
dataproc_xxxx_properties
and dataproc_xxx_jars
to dataproc_properties
and dataproc_jars
respectively.
Arguments for dataproc_properties dataproc_jars
No breaking changes.
Airflow 1.10 will be the last release series to support Python 2. Airflow 2.0.0 will only support Python 3.5 and up.
If you have a specific task that still requires Python 2 then you can use the PythonVirtualenvOperator for this.
- removed argument
version
fromget_conn
function and added it to the hook's__init__
function instead and renamed it toapi_version
- renamed the
partialKeys
argument of functionallocate_ids
topartial_keys
-
the discovery-based api (
googleapiclient.discovery
) used inGoogleCloudStorageHook
is now replaced by the recommended client based api (google-cloud-storage
). To know the difference between both the libraries, read https://cloud.google.com/apis/docs/client-libraries-explained. PR: #5054 -
as a part of this replacement, the
multipart
&num_retries
parameters forGoogleCloudStorageHook.upload
method have been deprecated.The client library uses multipart upload automatically if the object/blob size is more than 8 MB - source code. The client also handles retries automatically
-
the
generation
parameter is deprecated inGoogleCloudStorageHook.delete
andGoogleCloudStorageHook.insert_object_acl
.
Updating to google-cloud-storage >= 1.16
changes the signature of the upstream client.get_bucket()
method from get_bucket(bucket_name: str)
to get_bucket(bucket_or_name: Union[str, Bucket])
. This method is not directly exposed by the airflow hook, but any code accessing the connection directly (GoogleCloudStorageHook().get_conn().get_bucket(...)
or similar) will need to be updated.
The elasticsearch_
prefix has been removed from all config items under the [elasticsearch]
section. For example elasticsearch_host
is now just host
.
non_pooled_task_slot_count
and non_pooled_backfill_task_slot_count
are removed in favor of a real pool, e.g. default_pool
.
By default tasks are running in default_pool
.
default_pool
is initialized with 128 slots and user can change the
number of slots through UI/CLI. default_pool
cannot be removed.
The new pool
config option allows users to choose different pool
implementation. Default value is "prefork", while choices include "prefork" (default),
"eventlet", "gevent" or "solo". This may help users achieve better concurrency performance
in different scenarios.
For more details about Celery pool implementation, please refer to:
- https://docs.celeryproject.org/en/latest/userguide/workers.html#concurrency
- https://docs.celeryproject.org/en/latest/userguide/concurrency/eventlet.html
The signature of the get_task_instances
method in the BaseOperator
and DAG
classes has changed. The change does not change the behavior of the method in either case.
Old signature:
def get_task_instances(self, session, start_date=None, end_date=None):
New signature:
@provide_session
def get_task_instances(self, start_date=None, end_date=None, session=None):
Old signature:
def get_task_instances(
self, session, start_date=None, end_date=None, state=None):
New signature:
@provide_session
def get_task_instances(
self, start_date=None, end_date=None, state=None, session=None):
In either case, it is necessary to rewrite calls to the get_task_instances
method that currently provide the session
positional argument. New calls to this method look like:
# if you can rely on @provide_session
dag.get_task_instances()
# if you need to provide the session
dag.get_task_instances(session=your_session)
If you are using the Redis Sensor or Hook you may have to update your code. See redis-py porting instructions to check if your code might be affected (MSET, MSETNX, ZADD, and ZINCRBY all were, but read the full doc).
It is no longer required to set one of the environment variables to avoid a GPL dependency. Airflow will now always use text-unidecode if unidecode was not installed before.
The new sync_parallelism
config option will control how many processes CeleryExecutor will use to
fetch celery task state in parallel. Default value is max(1, number of cores - 1)
BashTaskRunner has been renamed to StandardTaskRunner. It is the default task runner so you might need to update your config.
task_runner = StandardTaskRunner
If the AIRFLOW_CONFIG
environment variable was not set and the
~/airflow/airflow.cfg
file existed, airflow previously used
~/airflow/airflow.cfg
instead of $AIRFLOW_HOME/airflow.cfg
. Now airflow
will discover its config file using the $AIRFLOW_CONFIG
and $AIRFLOW_HOME
environment variables rather than checking for the presence of a file.
If dag_discovery_safe_mode
is enabled, only check files for DAGs if
they contain the strings "airflow" and "DAG". For backwards
compatibility, this option is enabled by default.
Most GCP-related operators have now optional PROJECT_ID
parameter. In case you do not specify it,
the project id configured in
GCP Connection is used.
There will be an AirflowException
thrown in case PROJECT_ID
parameter is not specified and the
connection used has no project id defined. This change should be backwards compatible as earlier version
of the operators had PROJECT_ID
mandatory.
Operators involved:
- GCP Compute Operators
- GceInstanceStartOperator
- GceInstanceStopOperator
- GceSetMachineTypeOperator
- GCP Function Operators
- GcfFunctionDeployOperator
- GCP Cloud SQL Operators
- CloudSqlInstanceCreateOperator
- CloudSqlInstancePatchOperator
- CloudSqlInstanceDeleteOperator
- CloudSqlInstanceDatabaseCreateOperator
- CloudSqlInstanceDatabasePatchOperator
- CloudSqlInstanceDatabaseDeleteOperator
Other GCP operators are unaffected.
The change in GCP operators implies that GCP Hooks for those operators require now keyword parameters rather
than positional ones in all methods where project_id
is used. The methods throw an explanatory exception
in case they are called using positional parameters.
Hooks involved:
- GceHook
- GcfHook
- CloudSqlHook
Other GCP hooks are unaffected.
It's now possible to use None
as a default value with the default_var
parameter when getting a variable, e.g.
foo = Variable.get("foo", default_var=None)
if foo is None:
handle_missing_foo()
(Note: there is already Variable.setdefault()
which me be helpful in some cases.)
This changes the behaviour if you previously explicitly provided None
as a default value. If your code expects a KeyError
to be thrown, then don't pass the default_var
argument.
There were previously two ways of specifying the Airflow "home" directory
(~/airflow
by default): the AIRFLOW_HOME
environment variable, and the
airflow_home
config setting in the [core]
section.
If they had two different values different parts of the code base would end up
with different values. The config setting has been deprecated, and you should
remove the value from the config file and set AIRFLOW_HOME
environment
variable if you need to use a non default value for this.
(Since this setting is used to calculate what config file to load, it is not possible to keep just the config option)
The signature of the create_transfer_job
method in GCPTransferServiceHook
class has changed. The change does not change the behavior of the method.
Old signature:
def create_transfer_job(self, description, schedule, transfer_spec, project_id=None):
New signature:
def create_transfer_job(self, body):
It is necessary to rewrite calls to method. The new call looks like this:
body = {
'status': 'ENABLED',
'projectId': project_id,
'description': description,
'transferSpec': transfer_spec,
'schedule': schedule,
}
gct_hook.create_transfer_job(body)
The change results from the unification of all hooks and adjust to the official recommendations for the Google Cloud Platform.
The signature of wait_for_transfer_job
method in GCPTransferServiceHook
has changed.
Old signature:
def wait_for_transfer_job(self, job):
New signature:
def wait_for_transfer_job(self, job, expected_statuses=(GcpTransferOperationStatus.SUCCESS, )):
The behavior of wait_for_transfer_job
has changed:
Old behavior:
wait_for_transfer_job
would wait for the SUCCESS status in specified jobs operations.
New behavior:
You can now specify an array of expected statuses. wait_for_transfer_job
now waits for any of them.
The default value of expected_statuses
is SUCCESS so that change is backwards compatible.
The class GoogleCloudStorageToGoogleCloudStorageTransferOperator
has been moved from
airflow.contrib.operators.gcs_to_gcs_transfer_operator
to airflow.contrib.operators.gcp_transfer_operator
the class S3ToGoogleCloudStorageTransferOperator
has been moved from
airflow.contrib.operators.s3_to_gcs_transfer_operator
to airflow.contrib.operators.gcp_transfer_operator
The change was made to keep all the operators related to GCS Transfer Services in one file.
The previous imports will continue to work until Airflow 2.0
The driver_classapth
argument to SparkSubmit Hook and Operator was
generating --driver-classpath
on the spark command line, but this isn't a
valid option to spark.
The argument has been renamed to driver_class_path
and the option it
generates has been fixed.
Extend and enhance new Airflow RBAC UI to support DAG level ACL. Each dag now has two permissions(one for write, one for read) associated('can_dag_edit', 'can_dag_read').
The admin will create new role, associate the dag permission with the target dag and assign that role to users. That user can only access / view the certain dags on the UI
that he has permissions on. If a new role wants to access all the dags, the admin could associate dag permissions on an artificial view(all_dags
) with that role.
We also provide a new cli command(sync_perm
) to allow admin to auto sync permissions.
ts_nodash
previously contained TimeZone information along with execution date. For Example: 20150101T000000+0000
. This is not user-friendly for file or folder names which was a popular use case for ts_nodash
. Hence this behavior has been changed and using ts_nodash
will no longer contain TimeZone information, restoring the pre-1.10 behavior of this macro. And a new macro ts_nodash_with_tz
has been added which can be used to get a string with execution date and timezone info without dashes.
Examples:
ts_nodash
:20150101T000000
ts_nodash_with_tz
:20150101T000000+0000
next_ds/prev_ds now map to execution_date instead of the next/previous schedule-aligned execution date for DAGs triggered in the UI.
This patch changes the User.superuser
field from a hardcoded boolean to a Boolean()
database column. User.superuser
will default to False
, which means that this privilege will have to be granted manually to any users that may require it.
For example, open a Python shell and
from airflow import models, settings
session = settings.Session()
users = session.query(models.User).all() # [admin, regular_user]
users[1].superuser # False
admin = users[0]
admin.superuser = True
session.add(admin)
session.commit()
We have updated the version of flask-login we depend upon, and as a result any
custom auth backends might need a small change: is_active
,
is_authenticated
, and is_anonymous
should now be properties. What this means is if
previously you had this in your user class
def is_active(self):
return self.active
then you need to change it like this
@property
def is_active(self):
return self.active
GoogleCloudStorageToBigQueryOperator is now support schema auto-detection is available when you load data into BigQuery. Unfortunately, changes can be required.
If BigQuery tables are created outside of airflow and the schema is not defined in the task, multiple options are available:
define a schema_fields:
gcs_to_bq.GoogleCloudStorageToBigQueryOperator(
...
schema_fields={...})
or define a schema_object:
gcs_to_bq.GoogleCloudStorageToBigQueryOperator(
...
schema_object='path/to/schema/object)
or enabled autodetect of schema:
gcs_to_bq.GoogleCloudStorageToBigQueryOperator(
...
autodetect=True)
The scheduler_heartbeat
metric has been changed from a gauge to a counter. Each loop of the scheduler will increment the counter by 1. This provides a higher degree of visibility and allows for better integration with Prometheus using the StatsD Exporter. The scheduler's activity status can be determined by graphing and alerting using a rate of change of the counter. If the scheduler goes down, the rate will drop to 0.
EMRHook.create_job_flow has been changed to pass all keys to the create_job_flow API, rather than just specific known keys for greater flexibility.
However prior to this release the "emr_default" sample connection that was created had invalid configuration, so creating EMR clusters might fail until your connection is updated. (Ec2KeyName, Ec2SubnetId, TerminationProtection and KeepJobFlowAliveWhenNoSteps were all top-level keys when they should be inside the "Instances" dict)
Connecting to an LDAP server over plain text is not supported anymore. The
certificate presented by the LDAP server must be signed by a trusted
certificate, or you must provide the cacert
option under [ldap]
in the
config file.
If you want to use LDAP auth backend without TLS then you will have to create a custom-auth backend based on https://github.com/apache/airflow/blob/1.10.0/airflow/contrib/auth/backends/ldap_auth.py
Installation and upgrading requires setting SLUGIFY_USES_TEXT_UNIDECODE=yes
in your environment or
AIRFLOW_GPL_UNIDECODE=yes
. In case of the latter a GPL runtime dependency will be installed due to a
dependency (python-nvd3 -> python-slugify -> unidecode).
The method name was changed to be compatible with the Python 3.7 async/await keywords
Add a configuration variable(default_dag_run_display_number) to control numbers of dag run for display
Add a configuration variable(default_dag_run_display_number) under webserver section to control the number of dag runs to show in UI.
The current webserver UI uses the Flask-Admin extension. The new webserver UI uses the Flask-AppBuilder (FAB) extension. FAB has built-in authentication support and Role-Based Access Control (RBAC), which provides configurable roles and permissions for individual users.
To turn on this feature, in your airflow.cfg file (under [webserver]), set the configuration variable rbac = True
, and then run airflow
command, which will generate the webserver_config.py
file in your $AIRFLOW_HOME.
FAB has built-in authentication support for DB, OAuth, OpenID, LDAP, and REMOTE_USER. The default auth type is AUTH_DB
.
For any other authentication type (OAuth, OpenID, LDAP, REMOTE_USER), see the Authentication section of FAB docs for how to configure variables in webserver_config.py file.
Once you modify your config file, run airflow db init
to generate new tables for RBAC support (these tables will have the prefix ab_
).
Once configuration settings have been updated and new tables have been generated, create an admin account with airflow create_user
command.
Run airflow webserver
to start the new UI. This will bring up a log in page, enter the recently created admin username and password.
There are five roles created for Airflow by default: Admin, User, Op, Viewer, and Public. To configure roles/permissions, go to the Security
tab and click List Roles
in the new UI.
- AWS Batch Operator renamed property queue to job_queue to prevent conflict with the internal queue from CeleryExecutor - AIRFLOW-2542
- Users created and stored in the old users table will not be migrated automatically. FAB's built-in authentication support must be reconfigured.
- Airflow dag home page is now
/home
(instead of/admin
). - All ModelViews in Flask-AppBuilder follow a different pattern from Flask-Admin. The
/admin
part of the URL path will no longer exist. For example:/admin/connection
becomes/connection/list
,/admin/connection/new
becomes/connection/add
,/admin/connection/edit
becomes/connection/edit
, etc. - Due to security concerns, the new webserver will no longer support the features in the
Data Profiling
menu of old UI, includingAd Hoc Query
,Charts
, andKnown Events
. - HiveServer2Hook.get_results() always returns a list of tuples, even when a single column is queried, as per Python API 2.
- UTC is now the default timezone: Either reconfigure your workflows scheduling in UTC or set
default_timezone
as explained in https://airflow.apache.org/timezone.html#default-time-zone
We now rename airflow.contrib.sensors.hdfs_sensors to airflow.contrib.sensors.hdfs_sensor for consistency purpose.
We now rely on more strict ANSI SQL settings for MySQL in order to have sane defaults. Make sure
to have specified explicit_defaults_for_timestamp=1
in your my.cnf under [mysqld]
To make the config of Airflow compatible with Celery, some properties have been renamed:
celeryd_concurrency -> worker_concurrency
celery_result_backend -> result_backend
celery_ssl_active -> ssl_active
celery_ssl_cert -> ssl_cert
celery_ssl_key -> ssl_key
Resulting in the same config parameters as Celery 4, with more transparency.
Dataflow job labeling is now supported in Dataflow{Java,Python}Operator with a default "airflow-version" label, please upgrade your google-cloud-dataflow or apache-beam version to 2.2.0 or greater.
The GoogleCloudStorageDownloadOperator
can either write to a supplied filename
or return the content of a file via xcom through store_to_xcom_key
- both options are mutually exclusive.
The bql
parameter passed to BigQueryOperator
and BigQueryBaseCursor.run_query
has been deprecated and renamed to sql
for consistency purposes. Using bql
will still work (and raise a DeprecationWarning
), but is no longer
supported and will be removed entirely in Airflow 2.0
With Airflow 1.9 or lower, Unload operation always included header row. In order to include header row,
we need to turn off parallel unload. It is preferred to perform unload operation using all nodes so that it is
faster for larger tables. So, parameter called include_header
is added and default is set to False.
Header row will be added only if this parameter is set True and also in that case parallel will be automatically turned off (PARALLEL OFF
)
With Airflow 1.9 or lower, there were two connection strings for the Google Cloud operators, both google_cloud_storage_default
and google_cloud_default
. This can be confusing and therefore the google_cloud_storage_default
connection id has been replaced with google_cloud_default
to make the connection id consistent across Airflow.
With Airflow 1.9 or lower, FILENAME_TEMPLATE
, PROCESSOR_FILENAME_TEMPLATE
, LOG_ID_TEMPLATE
, END_OF_LOG_MARK
were configured in airflow_local_settings.py
. These have been moved into the configuration file, and hence if you were using a custom configuration file the following defaults need to be added.
[core]
fab_logging_level = WARN
log_filename_template = {{{{ ti.dag_id }}}}/{{{{ ti.task_id }}}}/{{{{ ts }}}}/{{{{ try_number }}}}.log
log_processor_filename_template = {{{{ filename }}}}.log
[elasticsearch]
elasticsearch_log_id_template = {{dag_id}}-{{task_id}}-{{execution_date}}-{{try_number}}
elasticsearch_end_of_log_mark = end_of_log
The previous setting of log_task_reader
is not needed in many cases now when using the default logging config with remote storages. (Previously it needed to be set to s3.task
or similar. This is not needed with the default config anymore)
With the change to Airflow core to be timezone aware the default log path for task instances will now include timezone information. This will by default mean all previous task logs won't be found. You can get the old behaviour back by setting the following config options:
[core]
log_filename_template = {{ ti.dag_id }}/{{ ti.task_id }}/{{ execution_date.strftime("%%Y-%%m-%%dT%%H:%%M:%%S") }}/{{ try_number }}.log
SSH Hook now uses the Paramiko library to create an ssh client connection, instead of the sub-process based ssh command execution previously (<1.9.0), so this is backward incompatible.
- update SSHHook constructor
- use SSHOperator class in place of SSHExecuteOperator which is removed now. Refer to test_ssh_operator.py for usage info.
- SFTPOperator is added to perform secure file transfer from serverA to serverB. Refer to test_sftp_operator.py for usage info.
- No updates are required if you are using ftpHook, it will continue to work as is.
The airflow.hooks.S3_hook.S3Hook has been switched to use boto3 instead of the older boto (a.k.a. boto2). This results in a few backwards incompatible changes to the following classes: S3Hook:
- the constructors no longer accepts
s3_conn_id
. It is now calledaws_conn_id
. - the default connection is now "aws_default" instead of "s3_default"
- the return type of objects returned by
get_bucket
is now boto3.s3.Bucket - the return type of
get_key
, andget_wildcard_key
is now an boto3.S3.Object.
If you are using any of these in your DAGs and specify a connection ID you will need to update the parameter name for the connection to "aws_conn_id": S3ToHiveTransfer, S3PrefixSensor, S3KeySensor, RedshiftToS3Transfer.
The logging structure of Airflow has been rewritten to make configuration easier and the logging system more transparent.
A logger is the entry point into the logging system. Each logger is a named bucket to which messages can be written for processing. A logger is configured to have a log level. This log level describes the severity of the messages that the logger will handle. Python defines the following log levels: DEBUG, INFO, WARNING, ERROR or CRITICAL.
Each message that is written to the logger is a Log Record. Each log record contains a log level indicating the severity of that specific message. A log record can also contain useful metadata that describes the event that is being logged. This can include details such as a stack trace or an error code.
When a message is given to the logger, the log level of the message is compared to the log level of the logger. If the log level of the message meets or exceeds the log level of the logger itself, the message will undergo further processing. If it doesn’t, the message will be ignored.
Once a logger has determined that a message needs to be processed, it is passed to a Handler. This configuration is now more flexible and can be easily be maintained in a single file.
Airflow's logging mechanism has been refactored to use Python’s built-in logging
module to perform logging of the application. By extending classes with the existing LoggingMixin
, all the logging will go through a central logger. Also the BaseHook
and BaseOperator
already extend this class, so it is easily available to do logging.
The main benefit is easier configuration of the logging by setting a single centralized python file. Disclaimer; there is still some inline configuration, but this will be removed eventually. The new logging class is defined by setting the dotted classpath in your ~/airflow/airflow.cfg
file:
# Logging class
# Specify the class that will specify the logging configuration
# This class has to be on the python classpath
logging_config_class = my.path.default_local_settings.LOGGING_CONFIG
The logging configuration file needs to be on the PYTHONPATH
, for example $AIRFLOW_HOME/config
. This directory is loaded by default. Any directory may be added to the PYTHONPATH
, this might be handy when the config is in another directory or a volume is mounted in case of Docker.
The config can be taken from airflow/config_templates/airflow_local_settings.py
as a starting point. Copy the contents to ${AIRFLOW_HOME}/config/airflow_local_settings.py
, and alter the config as is preferred.
# -*- coding: utf-8 -*-
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
import os
from airflow import configuration as conf
# TODO: Logging format and level should be configured
# in this file instead of from airflow.cfg. Currently
# there are other log format and level configurations in
# settings.py and cli.py. Please see AIRFLOW-1455.
LOG_LEVEL = conf.get('core', 'LOGGING_LEVEL').upper()
LOG_FORMAT = conf.get('core', 'log_format')
BASE_LOG_FOLDER = conf.get('core', 'BASE_LOG_FOLDER')
PROCESSOR_LOG_FOLDER = conf.get('scheduler', 'child_process_log_directory')
FILENAME_TEMPLATE = '{{ ti.dag_id }}/{{ ti.task_id }}/{{ ts }}/{{ try_number }}.log'
PROCESSOR_FILENAME_TEMPLATE = '{{ filename }}.log'
DEFAULT_LOGGING_CONFIG = {
'version': 1,
'disable_existing_loggers': False,
'formatters': {
'airflow.task': {
'format': LOG_FORMAT,
},
'airflow.processor': {
'format': LOG_FORMAT,
},
},
'handlers': {
'console': {
'class': 'logging.StreamHandler',
'formatter': 'airflow.task',
'stream': 'ext://sys.stdout'
},
'file.task': {
'class': 'airflow.utils.log.file_task_handler.FileTaskHandler',
'formatter': 'airflow.task',
'base_log_folder': os.path.expanduser(BASE_LOG_FOLDER),
'filename_template': FILENAME_TEMPLATE,
},
'file.processor': {
'class': 'airflow.utils.log.file_processor_handler.FileProcessorHandler',
'formatter': 'airflow.processor',
'base_log_folder': os.path.expanduser(PROCESSOR_LOG_FOLDER),
'filename_template': PROCESSOR_FILENAME_TEMPLATE,
}
# When using s3 or gcs, provide a customized LOGGING_CONFIG
# in airflow_local_settings within your PYTHONPATH, see UPDATING.md
# for details
# 's3.task': {
# 'class': 'airflow.utils.log.s3_task_handler.S3TaskHandler',
# 'formatter': 'airflow.task',
# 'base_log_folder': os.path.expanduser(BASE_LOG_FOLDER),
# 's3_log_folder': S3_LOG_FOLDER,
# 'filename_template': FILENAME_TEMPLATE,
# },
# 'gcs.task': {
# 'class': 'airflow.utils.log.gcs_task_handler.GCSTaskHandler',
# 'formatter': 'airflow.task',
# 'base_log_folder': os.path.expanduser(BASE_LOG_FOLDER),
# 'gcs_log_folder': GCS_LOG_FOLDER,
# 'filename_template': FILENAME_TEMPLATE,
# },
},
'loggers': {
'': {
'handlers': ['console'],
'level': LOG_LEVEL
},
'airflow': {
'handlers': ['console'],
'level': LOG_LEVEL,
'propagate': False,
},
'airflow.processor': {
'handlers': ['file.processor'],
'level': LOG_LEVEL,
'propagate': True,
},
'airflow.task': {
'handlers': ['file.task'],
'level': LOG_LEVEL,
'propagate': False,
},
'airflow.task_runner': {
'handlers': ['file.task'],
'level': LOG_LEVEL,
'propagate': True,
},
}
}
To customize the logging (for example, use logging rotate), define one or more of the logging handles that Python has to offer. For more details about the Python logging, please refer to the official logging documentation.
Furthermore, this change also simplifies logging within the DAG itself:
root@ae1bc863e815:/airflow# python
Python 3.6.2 (default, Sep 13 2017, 14:26:54)
[GCC 4.9.2] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> from airflow.settings import *
>>>
>>> from datetime import datetime
>>> from airflow import DAG
>>> from airflow.operators.dummy_operator import DummyOperator
>>>
>>> dag = DAG('simple_dag', start_date=datetime(2017, 9, 1))
>>>
>>> task = DummyOperator(task_id='task_1', dag=dag)
>>>
>>> task.log.error('I want to say something..')
[2017-09-25 20:17:04,927] {<stdin>:1} ERROR - I want to say something..
The file_task_handler
logger has been made more flexible. The default format can be changed, {dag_id}/{task_id}/{execution_date}/{try_number}.log
by supplying Jinja templating in the FILENAME_TEMPLATE
configuration variable. See the file_task_handler
for more information.
If you are logging to Google cloud storage, please see the Google cloud platform documentation for logging instructions.
If you are using S3, the instructions should be largely the same as the Google cloud platform instructions above. You will need a custom logging config. The REMOTE_BASE_LOG_FOLDER
configuration key in your airflow config has been removed, therefore you will need to take the following steps:
- Copy the logging configuration from
airflow/config_templates/airflow_logging_settings.py
. - Place it in a directory inside the Python import path
PYTHONPATH
. If you are using Python 2.7, ensuring that any__init__.py
files exist so that it is importable. - Update the config by setting the path of
REMOTE_BASE_LOG_FOLDER
explicitly in the config. TheREMOTE_BASE_LOG_FOLDER
key is not used anymore. - Set the
logging_config_class
to the filename and dict. For example, if you placecustom_logging_config.py
on the base of yourPYTHONPATH
, you will need to setlogging_config_class = custom_logging_config.LOGGING_CONFIG
in your config as Airflow 1.8.
A new DaskExecutor allows Airflow tasks to be run in Dask Distributed clusters.
These features are marked for deprecation. They may still work (and raise a DeprecationWarning
), but are no longer
supported and will be removed entirely in Airflow 2.0
-
If you're using the
google_cloud_conn_id
ordataproc_cluster
argument names explicitly incontrib.operators.Dataproc{*}Operator
(s), be sure to rename them togcp_conn_id
orcluster_name
, respectively. We've renamed these arguments for consistency. (AIRFLOW-1323) -
post_execute()
hooks now take two arguments,context
andresult
(AIRFLOW-886)Previously, post_execute() only took one argument,
context
. -
contrib.hooks.gcp_dataflow_hook.DataFlowHook
starts to use--runner=DataflowRunner
instead ofDataflowPipelineRunner
, which is removed from the packagegoogle-cloud-dataflow-0.6.0
. -
The pickle type for XCom messages has been replaced by json to prevent RCE attacks. Note that JSON serialization is stricter than pickling, so if you want to e.g. pass raw bytes through XCom you must encode them using an encoding like base64. By default pickling is still enabled until Airflow 2.0. To disable it set enable_xcom_pickling = False in your Airflow config.
The Airflow package name was changed from airflow
to apache-airflow
during this release. You must uninstall
a previously installed version of Airflow before installing 1.8.1.
The database schema needs to be upgraded. Make sure to shutdown Airflow and make a backup of your database. To
upgrade the schema issue airflow upgradedb
.
Systemd unit files have been updated. If you use systemd please make sure to update these.
Please note that the webserver does not detach properly, this will be fixed in a future version.
Airflow 1.7.1 has issues with being able to over subscribe to a pool, ie. more slots could be used than were available. This is fixed in Airflow 1.8.0, but due to past issue jobs may fail to start although their dependencies are met after an upgrade. To workaround either temporarily increase the amount of slots above the amount of queued tasks or use a new pool.
Using a dynamic start_date (e.g. start_date = datetime.now()
) is not considered a best practice. The 1.8.0 scheduler
is less forgiving in this area. If you encounter DAGs not being scheduled you can try using a fixed start_date and
renaming your DAG. The last step is required to make sure you start with a clean slate, otherwise the old schedule can
interfere.
Please read through the new scheduler options, defaults have changed since 1.7.1.
In order to increase the robustness of the scheduler, DAGS are now processed in their own process. Therefore each
DAG has its own log file for the scheduler. These log files are placed in child_process_log_directory
which defaults to
<AIRFLOW_HOME>/scheduler/latest
. You will need to make sure these log files are removed.
DAG logs or processor logs ignore and command line settings for log file locations.
Previously the command line option num_runs
was used to let the scheduler terminate after a certain amount of
loops. This is now time bound and defaults to -1
, which means run continuously. See also num_runs.
Previously num_runs
was used to let the scheduler terminate after a certain amount of loops. Now num_runs specifies
the number of times to try to schedule each DAG file within run_duration
time. Defaults to -1
, which means try
indefinitely. This is only available on the command line.
After how much time should an updated DAG be picked up from the filesystem.
CURRENTLY DISABLED DUE TO A BUG How many seconds to wait between file-parsing loops to prevent the logs from being spammed.
The frequency with which the scheduler should relist the contents of the DAG directory. If while developing +dags, they are not being picked up, have a look at this number and decrease it when necessary.
By default the scheduler will fill any missing interval DAG Runs between the last execution date and the current date.
This setting changes that behavior to only execute the latest interval. This can also be specified per DAG as
catchup = False / True
. Command line backfills will still work.
Due to changes in the way Airflow processes DAGs the Web UI does not show an error when processing a faulty DAG. To
find processing errors go the child_process_log_directory
which defaults to <AIRFLOW_HOME>/scheduler/latest
.
Previously, new DAGs would be scheduled immediately. To retain the old behavior, add this to airflow.cfg:
[core]
dags_are_paused_at_creation = False
If you specify a hive conf to the run_cli command of the HiveHook, Airflow add some convenience variables to the config. In case you run a secure Hadoop setup it might be required to whitelist these variables by adding the following to your configuration:
<property>
<name>hive.security.authorization.sqlstd.confwhitelist.append</name>
<value>airflow\.ctx\..*</value>
</property>
All Google Cloud Operators and Hooks are aligned and use the same client library. Now you have a single connection type for all kinds of Google Cloud Operators.
If you experience problems connecting with your operator make sure you set the connection type "Google Cloud Platform".
Also the old P12 key file type is not supported anymore and only the new JSON key files are supported as a service account.
These features are marked for deprecation. They may still work (and raise a DeprecationWarning
), but are no longer
supported and will be removed entirely in Airflow 2.0
-
Hooks and operators must be imported from their respective submodules
airflow.operators.PigOperator
is no longer supported;from airflow.operators.pig_operator import PigOperator
is. (AIRFLOW-31, AIRFLOW-200) -
Operators no longer accept arbitrary arguments
Previously,
Operator.__init__()
accepted any arguments (either positional*args
or keyword**kwargs
) without complaint. Now, invalid arguments will be rejected. (apache#1285) -
The config value secure_mode will default to True which will disable some insecure endpoints/features
There is a report that the default of "-1" for num_runs creates an issue where errors are reported while parsing tasks.
It was not confirmed, but a workaround was found by changing the default back to None
.
To do this edit cli.py
, find the following:
'num_runs': Arg(
("-n", "--num_runs"),
default=-1, type=int,
help="Set the number of runs to execute before exiting"),
and change default=-1
to default=None
. If you have this issue please report it on the mailing list.
To continue using the default smtp email backend, change the email_backend line in your config file from:
[email]
email_backend = airflow.utils.send_email_smtp
to:
[email]
email_backend = airflow.utils.email.send_email_smtp
To continue using S3 logging, update your config file so:
s3_log_folder = s3://my-airflow-log-bucket/logs
becomes:
remote_base_log_folder = s3://my-airflow-log-bucket/logs
remote_log_conn_id = <your desired s3 connection>