Optimize object storage costs automatically with smart tier—now generally available

We are excited to announce the general availability (GA) of smart tier for Azure Blob and Data Lake Storage. Smart tier is a fully managed, automated tiering capability for Azure Blob Storage and Data Lake Storage that helps optimize storage costs without ongoing operational effort. By continuously optimizing data placement, smart tier ensures your storage costs are aligned with actual usage.

Get in-depth details about smart tier

As data estates expand and access patterns evolve, managing lifecycle rules at scale becomes complex. Customers need automated, continuous tiering to keep costs aligned with usage.

Smart tier continuously evaluates your data access patterns and automatically moves objects across the hot, cool, and cold tiers to keep your costs aligned with usage without manual configuration.

Since launching the public preview of smart tier at Ignite in November 2025, customers and partners have adopted it across a range of data estates and over 50% of smart-tier–managed capacity has automatically shifted to cooler tiers based on actual access patterns:

We see a significant and measurable benefit from adopting smart tier in Azure Storage for our Azure Data Explorer (ADX) clusters. By intelligently placing data in the most cost‑effective tier based on actual usage patterns, smart tier allows us to optimize storage spend without sacrificing performance. Hot data remains instantly accessible for query workloads, while cooler, less frequently accessed data is automatically shifted to lower‑cost tiers. Smart tier effectively removed the guesswork from storage optimization, enabling us to focus on delivering insights rather than managing data placement.
Brad Watts, Principal PM for Azure Data Explorer

The Azure Blob and Data Lake Storage partner ecosystem is also integrating smart tier into their solutions:

Smart Tier represents a major step forward in simplifying how enterprises optimize storage in the cloud. The ability to automate tiering while maintaining resilience and predictable economics is highly complementary to Qumulo’s data services on Azure. Together with Microsoft, we’re enabling customers to modernize file workloads on Azure while reducing operational complexity and improving long‑term cost efficiency.
Brandon Whitelaw, SVP and Head of Product at Qumulo

Smart tier is generally available today in nearly all zonal public cloud regions, supporting both Azure Blob and Data Lake Storage.

How smart tier makes tiering decisions

Smart tier continuously evaluates the last access time of each individual object on the storage account where smart tier is enabled.

Frequently accessed data stays in the hot tier to support performance and transaction efficiency; inactive data transitions to the cool tier after 30 days and to the cold tier after an additional 60 days. When data is accessed again, it is immediately promoted back to hot and the tiering cycle restarts. This means your datasets remain in the most cost-effective tier automatically, removing the need to predict access patterns.

Read and write operations against an object, i.e. Get Blob or Put Blob operations are restarting the tiering cycle. Metadata operations, i.e. Get Blob Properties, are not impacting transitions. These static tiering rules are part of the underlying service and ensure automatic optimizations without the need for manual maintenance.

Setting up smart tier

Enabling smart tier is straightforward and designed to minimize change management while delivering immediate cost-optimization benefits:

During storage account creation, just select smart tier as the default access tier through the storage account configuration for any storage account with zonal redundancy. This is supported both via API and the Azure portal.

Enable existing accounts with zonal redundancies by switching the blob access tier from default to smart through the same tooling.

Let Azure optimize automatically: Objects inheriting the default tier are continuously managed without manual interventions needed.

Please note: Smart tier doesn’t support legacy account types such as Standard general-purpose v1 (GPv1) and is not applicable on page or append blobs.

For objects managed by smart tier, you pay standard hot, cool, and cold capacity rates, without additional charges for tier transitions, early deletion, or data retrieval. Moving existing objects into smart tier does not incur tier-change fees; a monitoring fee covers the orchestration.

Over time, automated down-tiering of inactive data combined with smart tier’s simplified billing can translate into meaningful savings at scale.

Best practices for maximizing smart tier value

After enabling smart tier on the account level, you can explicitly pin objects that you don’t want to be managed by smart tier to other tiers. No monitoring fee will apply to those objects.

Don’t exclude small objects. Objects less than 128 KiB stay in hot, don’t tier down, and don’t incur the monitoring fee. If an object later grows to equal to or greater than 128 KiB, smart tier policies apply automatically.

Common pitfall: Avoid trying to influence tiering behavior using lifecycle rules or other tier optimization mechanisms for smart tier–managed objects.

Based on patterns observed across multiple large smart tier preview deployments, customers commonly see the following outcomes after enabling smart tier:

Smart tier adoption for a large analytics workload

During public preview, a large data analytics customer enabled smart tier across hundreds of tebibytes of telemetry and log data with mixed and evolving access patterns.

Before enabling smart tier, the team relied on custom lifecycle rules that required frequent retuning as access patterns evolved and often led to unexpected cost spikes after re-access.

After enabling smart tier:

More than half of this customer’s managed data footprint automatically transitioned to cooler tiers based on actual usage patterns.

The team eliminated lifecycle policy management entirely, freeing engineering time.

Storage costs became more predictable and resilient to re-access spikes, since rehydration occurred automatically without retrieval or early deletion charges.

While savings vary by workload, this pattern reflects how smart tier helps align object storage costs with real usage.

Who should use smart tier?

Smart tier is well suited for organizations that:

Manage large or fast-growing object data estates.

Have mixed, evolving, or unpredictable access patterns.

Want to optimize costs without maintaining lifecycle rules.

Need data to remain online and immediately accessible, even when infrequently used.

Want safeguards against billing spikes caused by unplanned rehydration of cooler-tier datasets.

This includes analytics pipelines, data lakes, logs, telemetry, and application data where usage naturally changes over time.

Why enable smart tier now?

Reduce operational overhead: No lifecycle rules to design, test, or maintain.

Align costs with real usage: Data continuously moves to the most appropriate tier based on access patterns.

Preserve performance: Frequently accessed data remains hot; re‑access is automatic.

Simplify billing: No tier transition, early deletion, or retrieval charges within smart tier; a monthly monitoring fee occurs for each object in scope.

Scale with confidence: Built for large, evolving data estates.

What’s next for smart tier?

Smart tier is designed as a foundational capability that will continue to evolve. Upcoming improvements focus on:

Broader regional availability, including additional public cloud regions as GA rollout progresses.

Client tooling support: Watch out for upcoming releases of our Storage SDKs and tooling supporting this new capability.

Get started with smart tier

Enable smart tier during storage account creation or update an existing zonal storage account by setting smart tier as the default access tier. Once enabled, Azure continuously optimizes data placement—no ongoing configuration required.

Optimize data placement with smart tier

The post Optimize object storage costs automatically with smart tier—now generally available appeared first on Microsoft Azure Blog.
Quelle: Azure

Cloud Cost Optimization: Principles that still matter

In this article

What is cloud cost optimization and why does it still matter?How AI workloads change traditional cost optimizationCloud cost optimization best practices for AI and modern workloadsCloud cost management versus cost optimizationMeasuring value alongside cloud cost optimizationNext steps for cloud cost optimization on Azure

This blog post is the second in a multi-part series called Cloud Cost Optimization. Throughout this series, we’ll share practical strategies, best practices, and actionable guidance to help you plan, design, and manage AI investments for sustainable value and efficiency.

Cloud cost optimization continues to be a top priority for organizations of every size. As cloud environments grow and workloads scale, leaders are under constant pressure to control spend, reduce waste, and ensure that resources are being used efficiently. What was once a secondary operational concern has become a strategic capability tied directly to business performance, resilience, and long‑term growth.

At the same time, the rapid growth of AI workloads is adding a new layer of complexity to managing cloud costs. AI‑powered workloads and evolving usage patterns are transforming how organizations approach cloud optimization and investment planning. However, these changes do not replace the need for strong cost optimization practices. Instead, they make cloud cost optimization and AI cost management more critical than ever.

Maximize the return on your AI investment with Azure

This article provides a practical, evergreen overview of cloud cost optimization, how AI changes the cost landscape, and the principles organizations can apply to optimize cloud and AI workloads over time.

What is cloud cost optimization and why does it still matter?

Cloud cost optimization refers to the ongoing practice of analyzing cloud usage and making informed decisions to reduce unnecessary spend while maintaining performance, reliability, and scalability. It is not about cutting costs indiscriminately, but about ensuring that cloud resources are aligned to real workload demand and business value.

Unlike traditional IT environments, cloud platforms operate on consumption‑based pricing models. This means costs are directly tied to how resources are used, not just what is deployed. As a result, cost optimization is not a one‑time exercise. It requires continuous attention as environments evolve, workloads change, and new services are introduced.

Organizations that invest in cloud cost optimization benefit from:

Improved visibility into where cloud spend is going.

Reduced waste from underutilized or idle resources.

Better alignment between cloud usage and business needs.

Greater confidence when scaling workloads.

As cloud environments grow more complex (spanning multiple services, regions, and architectures), the importance of structured cloud cost management and optimization only increases. For organizations operating in the cloud, this makes cost optimization a foundational capability rather than an operational afterthought.

How AI workloads change traditional cost optimization

AI workloads introduce new cost dynamics that can challenge traditional cloud cost optimization approaches. While many principles still apply, the pace and variability of AI usage amplify the need for strong cost governance.

AI consumption patterns are often less predictable. Training models, running inference, and experimenting with different architectures can cause rapid fluctuations in compute and storage usage. Costs may spike during experimentation phases and stabilize later in production or shift again as models evolve.

AI development typically involves a higher degree of iteration. Teams may test multiple models, datasets, or configurations before settling on a production approach. Without strong visibility and controls, these experiments can quietly drive significant cloud costs and complicate efforts to optimize cloud costs effectively.

AI workloads often rely on specialized infrastructure and services that increase cost sensitivity. As a result, maintaining visibility and control requires intentional AI cost optimization and disciplined cloud cost management practices.

This makes cloud cost optimization even more critical in AI‑powered environments, not optional.

Cloud cost optimization best practices for AI and modern workloads

While technologies change, many cloud cost optimization best practices remain consistent across traditional and AI workloads. The key is applying them continuously and adapting them to modern usage patterns.

Visibility and usage awareness

Effective cost optimization starts with understanding how resources are being consumed. Organizations need clear insight into usage patterns across environments, workloads, and services to identify inefficiencies and optimization opportunities. Visibility is the foundation of both cloud cost management and AI cost management.

Governance guardrails

Guardrails help prevent unnecessary spend before it occurs. These can include usage boundaries, policy‑driven controls, and standardized approaches that encourage efficient resource consumption without slowing innovation. Strong governance supports sustainable cost optimization as environments scale.

Rightsizing and lifecycle thinking

Workloads change over time. Resources that were appropriate during development may be inefficient in production, or vice versa. Rightsizing and lifecycle awareness help ensure resources match actual needs at every stage, which is essential to optimizing cloud costs over the long term.

Continuous review and iteration

Cloud cost optimization is not static. Regular review cycles allow teams to adapt to changing usage patterns, new workloads, and evolving priorities, especially as AI solutions move from experimentation to scale.

These cloud cost optimization best practices apply whether organizations are optimizing traditional applications, data platforms, or AI workloads running at scale.

Cloud cost management versus cost optimization

Cloud cost management and cost optimization are closely related, but not the same.

Cloud cost management focuses on tracking, reporting, and understanding cloud spend. It answers questions like:

Where is money being spent?

How is usage trending over time?

Which workloads or services are driving costs?

Cloud cost optimization, on the other hand, is about action and decision‑making. It builds on cost management insights to determine:

Where inefficiencies exist.

What changes can reduce waste.

How to improve efficiency without compromising outcomes.

Organizations need both. Cloud cost management provides visibility, while cost optimization turns that visibility into informed decisions that improve efficiency, scalability, and resiliency (especially in AI‑heavy environments).

Measuring value alongside cloud cost optimization

Reducing cloud costs alone is rarely the goal. The real objective is ensuring that cloud and AI investments deliver sustainable value over time.

Effective cost optimization balances efficiency with outcomes. This means considering how resources contribute to workload performance, reliability, and long‑term viability (not just minimizing spend). For AI workloads, this balance is particularly important, as experimentation and innovation are essential but must still be managed responsibly.

By measuring efficiency and aligning cloud cost optimization and AI cost optimization efforts with workload value, organizations can avoid short‑term savings that undermine long‑term success. This value‑driven approach to managing cloud costs ensures optimization supports growth rather than constraining it.

Explore how Azure can help maximize your AI return on investment

Next steps for cloud cost optimization on Azure

Azure provides a broad set of resources designed to help organizations manage and optimize cloud and AI costs over time.

To explore guidance, best practices, and curated resources that support cost optimization across cloud and AI workloads, visit the solutions pages:

Maximize ROI from AI.

FinOps on Azure.

For deeper perspectives on related topics, you may also find these resources helpful:

Defining roles and responsibilities for cloud cost optimization.

Optimize your Azure costs to help meet your financial objectives.

Cost optimization is a continuous journey, one that becomes even more important as AI adoption accelerates. By applying durable principles and maintaining ongoing visibility and control, organizations can scale cloud and AI investments responsibly while maximizing long‑term value.

To go deeper, explore the Cloud Cost Optimization series for best practices and guidance on optimizing cloud and AI investments for long-term business impact.

Did you miss these posts in the Cloud Cost Optimization series?

Cloud Cost Optimization: How to maximize ROI from AI, manage costs, and unlock real business value

The post Cloud Cost Optimization: Principles that still matter appeared first on Microsoft Azure Blog.
Quelle: Azure

Amazon CloudWatch RUM now available in AWS European Sovereign Cloud

Amazon CloudWatch RUM (Real User Monitoring) is a feature of Amazon CloudWatch that enables developers and operations teams to collect, view, and analyze client-side performance data from real end-user sessions in web and mobile applications. With its expansion to the AWS European Sovereign Cloud, customers operating under strict European data residency and sovereignty requirements can now monitor their web application performance without data leaving the sovereign boundary. This capability is designed for enterprises, public sector organizations, and regulated industries in Europe that require full control over where their data is stored and processed.
CloudWatch RUM helps teams proactively identify and resolve performance bottlenecks across both web and mobile applications by surfacing real-time metrics such as page load times, JavaScript errors, HTTP failures, and mobile-specific signals like crash rates and network latency — enabling faster root cause analysis and improved end-user experience. For example, a European public sector organization can use CloudWatch RUM within the AWS European Sovereign Cloud to monitor citizen-facing web portals and mobile apps while maintaining full data sovereignty compliance.
CloudWatch RUM in the AWS European Sovereign Cloud is available today in the EU Sovereign (eusc-de-east-1) region — to get started, visit the Amazon CloudWatch RUM documentation.
Quelle: aws.amazon.com

SageMaker JumpStart now offers optimized deployments for foundation models

SageMaker JumpStart now offers optimized deployments, enabling customers to deploy foundation models with pre-configured settings tailored to specific use cases and performance constraints. SageMaker JumpStart optimized deployments simplify model deployment by offering task-aware configurations that optimize for cost, throughput, or latency based on your workload requirements – whether content generation, summarization, or Q&A. This launch includes support for 30+ popular models from Meta, Microsoft, Mistral AI, Qwen, Google, and TII, with visibility into key performance metrics like P50 latency, time-to-first token (TTFT), and throughput before deployment.
With SageMaker JumpStart optimized deployments, customers can select from use case-specific configurations (such as generative writing or chat-style interactions) and choose optimization targets including cost-optimized, throughput-optimized, latency-optimized, or balanced performance. Models deploy to SageMaker AI Managed Inference endpoints or SageMaker HyperPod clusters with pre-set configurations that eliminate guesswork while maintaining full visibility into deployment details. Available models include Meta Llama 3.1 and 3.2 variants, Microsoft Phi-3, Mistral AI models including the new Mistral-Small-24B-Instruct-2501, Qwen 2 and 3 series including multimodal Qwen2-VL, Google Gemma, and TII Falcon3. All deployments leverage SageMaker’s VPC deployment capabilities, ensuring data control and production-ready infrastructure with enterprise-grade security. The feature is available in all AWS regions where SageMaker JumpStart is curretly supported.
To get started with optimized deployments, navigate to Models in SageMaker Studio, select your desired foundation model in the JumpStart Models tab, choose “Deploy,” and select your use case and performance optimization target. For details, visit the SageMaker JumpStart documentation. AWS is actively expanding support to include additional models.
Quelle: aws.amazon.com

Amazon Managed Grafana now supports creating Grafana 12.4 workspaces

Amazon Managed Grafana now supports creating new workspaces with Grafana version 12.4.  This release includes features that were launched as a part of open source Grafana versions 11.0 to 12.4, including Drilldown apps, scenes powered dashboards, variables in transformations, visualization enhancements, and new features with the Amazon CloudWatch plugin.
Queryless Drilldown apps enable customers to perform point-and-click exploration of Prometheus metrics, Loki logs, Tempo traces, and Pyroscope profiles. The Scenes-powered rendering engine boosts dashboard performance. Amazon CloudWatch Logs adds support for PPL and SQL queries, cross-account Metrics Insights, and log anomaly detection. The rebuilt table visualization improves performance with CSS cell styling and interactive Actions buttons, while trendline transformations and navigation bookmarks enhance data exploration. Grafana 12.4 is supported in all AWS regions where Amazon Managed Grafana is generally available.
You can create a new Amazon Managed Grafana workspace from the AWS Console, SDK, or CLI. To explore the complete list of new features, please refer to the user documentation. Follow the instructions here to create workspaces with version 12.4. To learn more about Amazon Managed Grafana features and its pricing, visit the product page and pricing page.
Quelle: aws.amazon.com

AWS Deadline Cloud announces AI-powered troubleshooting assistant for render jobs

Today, AWS Deadline Cloud announces an AI-powered troubleshooting assistant that helps you quickly diagnose and resolve render job failures. AWS Deadline Cloud is a fully managed service that simplifies render management for computer-generated 2D/3D graphics and visual effects for films, TV shows, commercials, games, and industrial design. Render job failures from missing assets, software errors, configuration mismatches, and resource constraints can stall production pipelines and waste compute resources. Previously, diagnosing these issues required specialized technical staff to manually parse logs and identify root causes — a process that is time-consuming, difficult to scale, and often unavailable to smaller studios. The new Deadline Cloud assistant investigates failed jobs you identify, analyzes logs and metrics, detects common issues, and provides troubleshooting recommendations based on industry best practices and a pre-trained knowledge base covering Deadline Cloud, common render farm issues, and popular digital content creation applications including Autodesk Maya, 3ds Max, VRED, Blender, SideFX Houdini, Maxon Cinema 4D, Foundry Nuke, and Adobe After Effects. The assistant runs within your AWS account using Amazon Bedrock, keeping all data and analysis within your control. The Deadline Cloud assistant is available today in all AWS Regions where AWS Deadline Cloud is supported. Watch a demo on YouTube to see it in action, or visit the AWS Deadline Cloud documentation to learn more.
Quelle: aws.amazon.com