Amazon S3 to Azure Blob Storage Transfer Operator¶
The Blob service stores text and binary data as objects in the cloud. The Blob service offers the following three resources: the storage account, containers, and blobs. Within your storage account, containers provide a way to organize sets of blobs. For more information about the service visit Azure Blob Storage API documentation. This page shows how to upload data from local filesystem to Azure Blob Storage.
Use the S3ToWasbOperator
transfer to copy the data from Amazon Simple Storage Service (S3) to Azure Blob Storage.
Prerequisite Tasks¶
To use these operators, you must do a few things:
Create necessary resources using AZURE PORTAL or AZURE CLI.
Install API libraries via pip.
pip install 'apache-airflow[azure]'Detailed information is available Installation of Airflow®
Operators¶
Transfer Data from Amazon S3 to Blob Storage¶
To copy data from an Amazon AWS S3 Bucket to an Azure Blob Storage container, the following operator can be used:
S3ToWasbOperator
Example usage:
tests/system/microsoft/azure/example_s3_to_wasb.py
s3_to_wasb = S3ToAzureBlobStorageOperator(
task_id="s3_to_wasb",
s3_bucket=s3_bucket_name,
container_name=wasb_container_name,
s3_key=S3_KEY,
blob_prefix=BLOB_PREFIX, # Using a prefix for this
trigger_rule=TriggerRule.ALL_DONE,
replace=True,
)
Reference¶
For further information, please refer to the following links: