Loading...

How to migrate all Azure Storage Queue data between two different Storage Accounts with Python

How to migrate all Azure Storage Queue data between two different Storage Accounts with Python

Background

This article describes how to migrate all Azure Storage Queues data between two different storage accounts.

 

For this, we will use Azure Storage SDK for Python to copy all queues (and the respective data) from one Azure Storage Queue to another Azure Storage Queue. This approach will keep the data in the source queues, and will create new queues with the respective data in the destination Azure Storage Queue.

 

This script was developed and tested using the following versions but it is expected to work with previous versions:

  • Python 3.11.7
  • azure-identity (version: 1.15.0)
  • azure-storage-queue (version: 12.9.0)


Approach

 

In this section, you can find a sample code to copy all queues data between two Storage Accounts using the Azure Storage SDK for Python.

 

This Python sample code is based on Azure Storage SDK for Python. Please review our documentation here Quickstart: Azure Queue Storage client library for Python.

 

Prerequisites

 

Download or use any Python IDE of your choice.

  • On Python side, we will use the following packages:
    • azure-identity (more information here azure-identity · PyPI). To install, please run:
      pip install azure-identity​
    • azure-storage-queue (more information here azure-storage-queue · PyPI). To install, please run:
      pip install azure-storage-queue​​

 

Please see below the sample code to copy all the queues data between two Azure Storage Accounts using the storage connection string.

 

import os from azure.identity import DefaultAzureCredential from azure.storage.queue import QueueServiceClient, QueueClient, QueueMessage try: # Connect to the source and target storage accounts source_connection_string = "XXXX" target_connection_string = "XXXX" # Create a QueueServiceClient for both source and target account)s source_client = QueueServiceClient.from_connection_string(source_connection_string) target_client = QueueServiceClient.from_connection_string(target_connection_string) # List all queues from the source account for queue in source_client.list_queues(): print(queue) # Create the same queue in the target account target_queue_client = target_client.create_queue(queue.name) # Read messages from the source queue for message in source_client.get_queue_client(queue.name).receive_messages(): # Add the message to the target queue target_queue_client.send_message(message.content) print("Data migration completed successfully!") except Exception as ex: print("Exception:") print(ex)

 

 

After executing this sample code, it is expected that you will find all the queues from the source Storage Account in the destination Storage Account, as well as the data/messages from those queues.

 

Disclaimer:

  • These steps are provided for the purpose of illustration only. 
  • These steps and any related information are provided "as is" without warranty of any kind, either expressed or implied, including but not limited to the implied warranties of merchantability and/or fitness for a particular purpose.
  • We grant You a nonexclusive, royalty-free right to use and modify the Steps and to reproduce and distribute the steps, provided that. You agree:
    • to not use Our name, logo, or trademarks to market Your software product in which the steps are embedded;
    • to include a valid copyright notice on Your software product in which the steps are embedded; and
    • to indemnify, hold harmless, and defend Us and Our suppliers from and against any claims or lawsuits, including attorneys’ fees, that arise or result from the use or distribution of steps.

 

 

Published on:

Learn more
Azure PaaS Blog articles
Azure PaaS Blog articles

Azure PaaS Blog articles

Share post:

Related posts

Announcing Data API builder General Availability for Azure Cosmos DB

Great news for developers out there! The Data API Builder for Azure Cosmos DB is now officially available, eliminating the need for writing co...

23 hours ago

Organizing rule collections and rule collection groups in Azure Firewall Policy

Firewall Policy is the recommended method to manage Azure Firewall security and operational configurations. When using Firewall Policy, any ru...

1 day ago

Get Started with Azure AI Services | Open AI and Deployment Models

Table of Contents   Overview - Azure AI Services Kind of Azure AI Services Responsible AI Services Limited Access Features Cognitive Acc...

1 day ago

Using Sempy to Authenticate to Fabric/Power BI APIs using Service Principal and Azure Key Vault

In this blog post, the author demonstrates how to use Azure Key Vault and Azure identity to authenticate securely when working with Fabric Not...

1 day ago

Unlock Your Python Potential with Azure

Microsoft's appreciation for Python's versatility and developer-friendly features has led to the creation of tools and resources aimed at assi...

1 day ago

Azure Lab Services - Lab Plan Outage

Azure Lab Services is currently experiencing an outage that affects Lab Plans, but not Lab Accounts. This outage intermittently impacts all op...

2 days ago

Azure Cosmos DB Conf 2024: Accelerating Innovation in AI and Data

The fourth annual Azure Cosmos DB Conf held on April 16, 2024, was a highly anticipated event for those at the forefront of cloud data managem...

2 days ago

New ‘ExecutionMetrics’ event in Azure Log Analytics for Power BI Semantic Models

The Power BI integration with Azure Log Analytics just got better with the introduction of a new event - 'ExecutionMetrics'. Customers can now...

2 days ago
Stay up to date with latest Microsoft Dynamics 365 and Power Platform news!
* Yes, I agree to the privacy policy