Optimizing Azure Table Storage: Automated Data Cleanup using a PowerShell script with Azure Automate
Scenario
This blog’s aim is to manage Table Storage data efficiently. Imagine you have a large Azure Table Storage that accumulates logs from various applications or any unused older data. Over time, this data grows significantly, making it necessary to periodically clean up old entries to maintain performance and manage costs. You decide to automate this process using Azure Automation. However, lifecycle management policies are limited to the Blob service only.
By scheduling a PowerShell script, you can efficiently delete outdated data from your Azure Table Storage without manual intervention. This approach ensures that your storage remains optimized, and your applications continue to run smoothly.
Below is the PowerShell script which delete Table Entities based on Timestamp: -
|
Connect-AzAccount -Identity $SubscriptionID = "xxxxxxxxxxxxxxxxxx" $StorageAccount = "xxxxxxxxxxxxxxxxxxx" foreach ($table in $alltablename) {
#DISCLAIMER |
Here are the steps to schedule a PowerShell script in Azure Automation :-
-
Create an Azure Automation Account by following the link:
-
Add Modules to Azure Automation Account:
- Navigate to the created automation account page.
- Go to the "Modules" tab under the "Shared Resources" section and choose the "Add a module" option.
- You can either manually import modules from your local machine or import inbuilt modules from the gallery.
- In this article, we will proceed with the gallery option.
- Search for the Storage Modules.
- Add the module with recommended Runtime version.
-
Create a PowerShell Runbook:
- In the Azure Portal, navigate to your Automation Account.
- Under "Process Automation", select "Runbooks".
- Click on "Create a runbook".
- Enter a name for the runbook, select "PowerShell" as the Runbook type, and click "Create".
- Once Runbook is created, in the "Edit PowerShell Runbook" page.
- Enter your PowerShell script and click "Publish".
-
Schedule the Runbook:
- Go to the respective Runbook and choose the "Link to schedule" option.
- Select the "Link a schedule to your Runbook" option and choose the appropriate schedule.
- If you go ahead wit Schedule option, you can create a new schedule by specifying the name, description, start date, time, time zone, and repeating information.
- Go to the respective Runbook and choose the "Link to schedule" option.
-
Monitor the Runbook:
- You can monitor the runbook's execution by going to the Jobs section under Process Automation in your Automation account.
- Here, you can see the status of the runbook jobs, view job details, and troubleshoot any issues.
These steps should help you schedule your PowerShell script in Azure Automation. If you have any more questions or need further assistance, feel free to ask!
References :-
Published on:
Learn moreRelated posts
What’s New with Microsoft Foundry (formerly Azure AI Foundry) from Ignite 2025
Microsoft Ignite 2025 just wrapped up, and one of the biggest themes this year was the evolution of Azure AI Foundry, now simply called Micros...
Announcing: Dynamic Data Masking for Azure Cosmos DB (Preview)
Today marks a big step forward with the public preview of Dynamic Data Masking (DDM) for Azure Cosmos DB. This feature helps organizations pro...
Use Azure SRE Agent with Azure Cosmos DB: Smarter Diagnostics for Your Applications
We’re excited to announce the Azure Cosmos DB SRE Agent built on Azure SRE Agent; a new capability designed to simplify troubleshooting and im...
General Availability: Priority-Based Execution in Azure Cosmos DB
Have you ever faced a situation where two different workloads share the same container, and one ends up slowing down the other? This is a comm...
Announcing Preview of Online Copy Jobs in Azure Cosmos DB: Migrate Data with Minimal Downtime!
We are excited to announce the preview of Online Copy Jobs, a powerful new feature designed to make data migration between containers seamless...