Microsoft Cost Management updates—July 2024

Whether you’re a new student, a thriving startup, or the largest enterprise, you have financial constraints, and you need to know what you’re spending, where it’s being spent, and how to plan for the future. Nobody wants a surprise when it comes to the bill, and this is where Microsoft Cost Management comes in.

We’re always looking for ways to learn more about your challenges and how Microsoft Cost Management can help you better understand where you’re accruing costs in the cloud, identify and prevent bad spending patterns, and optimize costs to empower you to do more with less. Here are a few of the latest improvements and updates based on your feedback:

Exports enhancements: Parquet format support, file compression, and Fabric ingestion

Pricing updates on Azure.com

Your feedback matters: Take our quick survey! 

New ways to save money with Microsoft Cloud

Documentation updates

Let’s dig into the details.

Exports enhancements: Parquet format support, file compression, and Fabric ingestion 

In our last blog, I spoke about the support for FOCUS 1.0 (FinOps Cost Usage and Specification) datasets in Exports. We continue to make enhancements to the Exports functionality bringing support for the parquet format and file compression which can potentially help you achieve 40 to 70% file size reduction. These new cost saving features are initially available for the following datasets: Cost and usage details (Actual, Amortized, FOCUS) and Price Sheet. They aim to streamline your cost management processes, improve data handling efficiency, and reduce storage and network costs, all while providing comprehensive insights into your Azure spending.

Parquet is an open-source, columnar storage file format designed for efficient data processing and analytics. It offers several benefits over traditional formats like Comma-Separated Values (CSV), some of which are included below:

Efficient storage and reduced network cost: Parquet’s columnar format allows for better compression and encoding schemes, resulting in smaller file sizes. Compressed datasets occupy less space, translating to lower storage expenses and file transfer network cost.

Improved data transfer speed: Smaller file sizes mean faster data transfer rates, enhancing the efficiency of data operations.

Faster query performance: By storing data by column, parquet enables faster data retrieval and query performance, especially for large datasets.

Optimized analytics: Parquet format is optimized for big data tools and can be easily integrated with various analytics platforms.

To further reduce the size of your datasets, you can now compress your CSV files using GNU ZIP (GZIP) and parquet files using Snappy.

Here is the screenshot showing the new configuration options:

Please refer to this article to get started.

Microsoft Fabric ingestion 

Microsoft Fabric, as we know, is a great tool for data reporting and analytics where you can reference datasets from multiple sources without copying the data. We have now added new documentation to make it easy for you to ingest your exported costs datasets into new or existing Fabric workspaces. Just follow the steps included in this article. 

Pricing updates on Azure.com

We’ve been working hard to make some changes to our Azure pricing experiences, and we’re excited to share them with you. These changes will help make it easier for you to estimate the costs of your solutions.

We’re thrilled to announce the launch of new pricing pages for Azure AI Health (now generally available) and the innovative Phi-3 service (now in preview), ensuring you have the latest information at your fingertips.

Our Azure AI suite has seen significant enhancements, with updated calculators for Azure AI Vision and Azure AI Language, ensuring you have access to the most current offers and SKUs. The Azure AI Speech service now proudly offers generally available pricing for the cutting-edge Text to Speech add-on feature “Avatar”, and Azure AI Document Intelligence has added pricing for new training and custom generative stock-keeping units (SKUs).

To maintain the accuracy and relevance of our offers, we’ve deprecated the Azure HPC Cache and SQL Server Stretch pricing pages and calculators. This step ensures that you’re only presented with the most up-to-date and valid options.

The pricing calculator has been updated with the latest offers and SKUs for Azure Container Storage, Azure AI Vision, Azure Monitor, and PostgreSQL, reflecting our commitment to providing you with the most accurate cost estimates.

We’ve introduced new prices and SKUs across various services, including pricing for the new Intel Dv6/Ev6 series (preview) and ND Mi300X v5 series for Virtual Machines, auxiliary logs offer for Azure Monitor, and audio streaming and closed caption SKUs for Azure Communication Services. The Azure Databricks service now features pricing for Automated Serverless Compute, and the Azure Container Storage service pricing page now reflects generally available pricing.

Our dedication to enhancing your pricing experience is reflected in the continuous improvements made to several pages, including Azure Synapse Analytics, Azure SQL Database, Azure Migrate, Azure Cosmos DB (autoscale-provisioned), Microsoft Purview, Microsoft Fabric, Linux Virtual Machines, Azure VMware Solution, Azure Web PubSub, Azure Content Delivery Network, and Azure SignalR Service.

We’re constantly working to improve our pricing tools and make them more accessible and user-friendly. We hope you find these changes helpful in estimating the costs for your Azure Solutions. If you have any feedback or suggestions for future improvements, please let us know!

Your Feedback Matters: Take our quick survey!

If you use Azure in your day-to-day work from deploying resources to managing costs and billing, we would love to hear from you. (All experience levels welcome!) Please take a few moments to complete this short, 5 to 10-minute survey to help us understand your roles, responsibilities, and the challenges you face in managing the cloud. Your feedback will help us improve our services to better meet your personal needs. 

New ways to save money in the Microsoft Cloud

Here are new and updated offerings which can potentially help with your cost optimization needs:

Generally Available: Azure Virtual Network Manager mesh and direct connectivity

Generally Available: Announcing kube-egress-gateway for Kubernetes

Generally Available: Run your Databricks Jobs with Serverless compute for workflows

Generally Available: Azure Elastic SAN Feature Updates

Generally Available: Azure Virtual Network Manager mesh and direct connectivity

Public Preview: Summary rules in Azure Monitor Log Analytics, for optimal consumption experiences and cost

Public Preview: Continuous Performance Diagnostics for Windows VMs to enhance VM Troubleshooting

Public Preview: Azure cross-subscription Load Balancer

Public Preview: Advanced Network Observability for your Azure Kubernetes Service clusters through Azure Monitor

New Azure Advisor recommendations for Azure Database for PostgreSQL—Flexible Server

Want a more guided experience? Start with Control Azure spending and manage bills with Microsoft Cost Management.

Documentation updates 

Here are a few costs related documentation updates you might be interested in:

Update: Centrally managed Azure Hybrid Benefit FAQ

Update: Pay for your Azure subscription by wire transfer

Update: Tutorial: Create and manage budgets

Update: Understand cost details fields

Update: Quickstart: Start using Cost analysis

Update: Tutorial: Improved exports experience—Preview

Update: Transfer Azure Enterprise enrollment accounts and subscriptions

Update:  Migrate from Consumption Usage Details API

Update: Change contact information for an Azure billing account

New: Avoid unused subscriptions

Want to keep an eye on all documentation updates? Check out the Cost Management and Billing documentation change history in the azure-docs repository on GitHub. If you see something missing, select Edit at the top of the document and submit a quick pull request. You can also submit a GitHub issue. We welcome and appreciate all contributions!

What’s next?

These are just a few of the big updates from last month. Don’t forget to check out the previous Microsoft Cost Management updates. We’re always listening and making constant improvements based on your feedback, so please keep the feedback coming.

Follow the Microsoft Cost Management YouTube channel to stay in the loop with new videos as they’re released and let us know what you’d like to see next.
The post Microsoft Cost Management updates—July 2024 appeared first on Azure Blog.
Quelle: Azure

New Azure Data Box capabilities to accelerate your offline data migration

Azure Data Box offline data transfer solution allows you send petabytes of data into Azure Storage in a quick, inexpensive, and reliable manner. The secure data transfer is accelerated by hardware transfer devices that enable offline data ingestion to Azure.

We’re excited to announce several new service capabilities including:

General availability of self-encrypted drives Azure Data Box Disk SKU that enables fast transfers on Linux systems.

Support for data ingestion to multiple blob access tiers in a single order.

Preview of cross-region data transfers for seamless data ingest from source country or region to select Azure destinations in a different country or region.

Support in Azure Storage Mover for online catch-up data copy of any changes active workloads may have generated post offline migrations with Azure Data Box.

Additionally, we’re happy to share the Azure Data Box cloud service is HIPAA/BAA, PCI 3DS and PCI DSS certified. More details on each of these new capabilities can be found below.

Azure Data Box
Move stored or in-flight data to Azure quickly and cost-effectively

Learn more

Azure Data Box Disk: self-encrypted drives

Azure Data Box Disk is now generally available in a hardware-encrypted option in the European Union, United States, and Japan. These self-encrypting drives (SEDs) use the dedicated/native/specialized hardware for data encryption, without any software dependency from the host machine. These SEDs use the specialized native hardware present on the disk for data encryption, without any software dependencies on the host machine. With this offering, we now support comparable data transfer rates on Linux as that of our BitLocker-encrypted Data Box Disk drives on Windows.

Azure Data Box Disk SED is popular with some of our Automotive customers as it connects directly to the in-car Linux-based data loggers through a SATA interface, thereby eliminating the need for a secondary data copy from another in-car storage and saving time. Here is how Xylon, manufacturer of automotive data loggers uses Azure Data Box Disk: self encrypted drives to migrate Advanced driver-assistance systems (ADAS) sensor data to Azure: 

Through the cooperation with the Microsoft Azure team, we have enabled direct data logging to the hardware-encrypted Data Box Disks plugged into our logiRECORDER Automotive HIL Video Logger. It enables our common customers to transfer precious data from the test fleet to the cloud in the simplest and fastest possible way, without wasting time on unnecessary data copying and reformatting along the way.” 
—Jura Ivanovic, Product Director, Automotive HIL Video Logger, Xylon 

Learn more about Data Box Disk: self encrypted drives and get started migrating your on-premises data to Azure. 

Multi-access tier ingestion support

You can now transfer data to different blob access tiers including Cold Tier in a single Azure Data Box order. Previously, Azure Data Box only supported transferring data to the default access tiers of Azure Storage Accounts. For example, if you wanted to move data to the Cool tier in an Azure Storage Account that has the default set to hot, you would have had to first move the data to hot tier via Azure Data Box and then leverage life cycle management to move the data to the Cool tier after it’s uploaded to Azure. 

We have now introduced new “access tier” folders in the folder hierarchy on the device. All data that you copy to the “Cool” folder will have it’s access tier set as cool, irrespective of the default access tier of the destination Storage account, and similarly for data copied to other folders representing the various access tiers. Learn more about multi-access tier ingestion support. 

Cross-region data transfer to select Azure regions 

We’re excited to share that Azure Data Box cross-region data transfer capabilities, now in preview, supports seamless ingest of on-premises data from a source country or region to select Azure destinations in a different country or region. For example, with this capability you can now copy on-premises data from Singapore or India to the West United States Azure destination region. Note that the Azure Data Box device isn’t shipped across commerce boundaries. Instead, it’s transported from and to an Azure data Center within the originating country or region where the on-premises data resides. Data transfer to the destination Azure region takes place across the Azure network without incurring additional fees. 

Learn more about this capability and the supported country or region combinations for Azure Data Box, Azure Data Box Disk, and Azure Data Box Heavy respectively. 

Support for online catch-up copy with Azure Storage Mover Integration 

If your data source has any active workloads, it will likely make changes while your Azure Data Box is in transit to Azure. Consequently, you’ll also need to bring those changes to your cloud storage, before a workload can be cut over to it. We’re happy to announce that you can now combine the Azure Storage Mover and Data Box services to form an effective file and folder migration solution to minimize downtime for your workloads. Storage Mover jobs can detect differences between your on-site and cloud storage to effectively transfer any updates and new files not previously captured by your Data Box transfer. For example, if only a file’s metadata (such as permissions) has changed, Azure Storage Mover will upload only the new metadata instead of the entire file content. 

Learn more about how catch-up copies with Azure Storage Mover’s merge and mirror copy mode can help transfer only the delta data to Azure.

Certifications

The Azure Data Box cloud service has achieved HIPAA/BAA, PCI 3DS & PCI DSS certifications. These certifications have been key requests from many of our customers across the healthcare and financial sectors respectively, and we’re happy to have achieved the compliance status to enable our customers’ data transfer needs.

Additional product updates

Support for up to 4 TB Azure files across the product family. 

Support for data transfer to “Poland Central” and “Italy North” Azure regions. 

Transfers to Premium Azure Files and Blob Archive tiers now supported with Data Box Disk. 

The data copy service, which significantly improves the ingestion and upload time for small files, is now generally available.

Our goal is to continually enhance the simplicity of your offline data transfers, and your input is invaluable. Should you have any suggestions or feedback regarding Azure Data Box, feel free to reach out via email at DataBox@microsoft.com. We look forward to you reviewing your feedback and comments.
The post New Azure Data Box capabilities to accelerate your offline data migration appeared first on Azure Blog.
Quelle: Azure

Announcing a new OpenAI feature for developers on Azure 

We are thrilled to announce the launch of OpenAI’s latest model on Azure. This new model, officially named GPT-4o-2024-08-06, brings innovative features designed to elevate developer experiences on Azure. Specifically, the new model focuses on enhancing productivity through Structured Outputs, like JSON Schemas, for the new GPT-4o and GPT-4o mini models. 

Azure OpenAI Service
Build your own copilot and generative AI applications.

Learn more

A focus on Structured Outputs 

GPT-4o was first announced in May 2024, as OpenAI’s new multimodal model, followed by GPT-4o mini in July 2024. Today’s version is designed with a specific use case in mind: simplifying the process of generating well-defined, structured outputs from AI models. This feature is particularly valuable for developers who need to validate and format AI outputs into structures like JSON Schemas. Developers often face challenges validating and formatting AI outputs into well-defined structures like JSON Schemas.  

Structured Outputs addresses this by allowing developers to specify the desired output format directly from the AI model. This feature enables developers to define a JSON Schema for text outputs, simplifying the process of generating data payloads that can seamlessly integrate with other systems or enhance user experiences. 

Use cases for JSON 

JSON Schema is essential for defining the structure and constraints of JSON documents, ensuring they follow specific formats with mandatory properties and value types. It enhances data understandability through semantic annotation and serves as a domain-specific language for optimized application requirements. Development teams use JSON Schema to maintain consistency across platforms, drive model-driven UI constraints, and automatically generate user interfaces. It aids in data serialization, security testing, and partial validation in technical scenarios. JSON Schema also supports automated testing, Schema inference, and machine-readable web profiles, improving data interoperability. It standardizes validation interfaces and reporting, handles external validation, and ensures data consistency within and across documents. It can also help with customer support and how to communicate in a timely manner. 

Two flavors of Structured Outputs 

Structured Outputs is available in two forms: 

User-defined JSON Schema: This option allows developers to specify the exact JSON Schema they want the AI to follow, supported by both GPT-4o-2024-08-06 and GPT-4o-mini-2024-07-18.

More Accurate Tool Output (“Strict Mode”): This limited version lets developers define specific function signatures for tool use, supported by all models that support function calling, including GPT-3.5 Turbo, GPT-4, GPT-4 Turbo, and GPT-4o models from June 2023 onwards. 

Technical guidance on using Structured Outputs 

To help you get started with Structured Outputs, we recommend the following approach. 

Getting started with Structured Outputs 

Define Your JSON Schema: Determine the structure you want your AI outputs to follow. This can include required fields, data types, and other constraints. 

Configure the AI model: Use the Structured Outputs feature to specify your JSON Schema within the API call. This ensures that the AI output adheres to your defined structure. 

Integration and testing: Integrate the output into your application or system, and test thoroughly to ensure compliance with your JSON Schema. 

Example use case: Customer support automation 

Imagine you’re developing a customer support chatbot that needs to generate responses in a specific format for logging and analytics. By using Structured Outputs, you can define a JSON Schema that includes fields like responseText, intent, confidenceScore, and timestamp. This ensures that every response generated by the chatbot is formatted correctly, making it easier to log, analyze, and act upon. 

Example API call 

Here’s an example API call to illustrate how to use Structured Outputs:

{
"model": "gpt-4o-2024-08-06",
"prompt": "Generate a customer support response",
"structured_output": {
"schema": {
"type": "object",
"properties": {
"responseText": { "type": "string" },
"intent": { "type": "string" },
"confidenceScore": { "type": "number" },
"timestamp": { "type": "string", "format": "date-time" }
},
"required": ["responseText", "intent", "confidenceScore", "timestamp"]
}
}
}

Pricing 

We will make pricing for this feature available soon. Please bookmark the Azure OpenAI Service pricing page. 

Learn more about the future of AI

We’ve been rolling out several new models recently, and we understand it can be a lot to keep up with. This flurry of activity is all about empowering developer innovation. Each new model brings unique capabilities and enhancements, helping you build even more powerful and versatile applications. 

The launch of this new model feature for GPT-4o and GPT-4o mini marks a significant milestone in our ongoing efforts to push the boundaries of AI capabilities. We’re excited to see how developers will leverage these new features to create innovative and impactful applications. 

Azure ai studio

Craft AI solutions your way

Stay tuned for more updates and get ready to experience the future of AI with these new developer features for GPT-4o and mini. Start experimenting in the Azure OpenAI Playground. 

Explore Azure OpenAI Service

The post Announcing a new OpenAI feature for developers on Azure  appeared first on Azure Blog.
Quelle: Azure

Build AI-enabled applications with Azure AI and NVIDIA

Learn how Azure AI, combined with NVIDIA AI, can help you create innovative and intelligent solutions using your preferred tools and workflows.

An explosion of interest in generative AI across many industries has sparked, a direct result of the collaboration of Microsoft and NVIDIA and the breakthrough technology behind OpenAI’s ChatGPT. As a result, artificial intelligence (AI) is transforming the way we interact with digital products and services, from chatbots and voice assistants to smart cameras and recommendation systems. It’s now almost a daily demand to leverage the power of AI to create applications that can understand, interact, and learn from their users and environments.

However, building these AI applications presents significant challenges in terms of time, resources, access to AI infrastructure, and costs—which can be prohibitive for many developers and organizations.

To alleviate these challenges, developers can benefit from the combined benefits of Azure AI—a set of cloud-based AI and machine learning services that can help you build, train, and deploy AI-enabled applications with ease—and the NVIDIA AI platform to maximize application performance throughout every stage of development and deployment. To make access and entry even easier to build the best AI-enabled applications, Microsoft and NVIDIA have launched a collaborative resource for developers and organizations to experience the better together benefits.

Azure AI
Lead your market with multimodal and generative AI

Innovate with AI

In this blog, we’ll discuss how combining the power of both Azure AI and the NVIDIA AI Platform can help you create your most impactful AI-enabled applications, providing you with flexibility, productivity, efficiency, and innovation.

Better together: Microsoft and NVIDIA

Recognizing the barriers developers face, NVIDIA and Microsoft have worked closely to democratize access to the same core technology that powers ChatGPT to accelerate adoption. The partnership focuses on optimizing every layer of the generative AI stack—from highly performant and scalable AI infrastructure to developer-friendly tools and services—to reduce complexity and cost, making advanced AI capabilities more accessible and feasible for a broader range of applications and industries.

Used by more than 60,000 organizations, Azure AI integrates with popular developer environments Visual Studio Code and GitHub, allowing you to use your preferred tools and workflows to develop, test, and deploy your AI solutions. Whether you want to use pre-built models and APIs, or build and train your own custom models, Azure AI can support various AI scenarios including building your own copilot with enterprise chat, speech analytics, document processing automation, and more.

Azure’s leading cloud AI supercomputing infrastructure, leveraging both state of the art NVIDIA GPUs and NVIDIA InfiniBand networking, provides the best performance, scalability, and built-in security needed to build, train, and deploy the most demanding AI workloads with confidence, at any scale. This combination accelerates time to solution, lowers deployment costs by supporting more users with fewer compute resources, and enhances user experience through optimized performance and faster data throughput.

Benefits for developers in GitHub and Visual Studio Code

Whether a developer that uses Visual Studio Code or GitHub, Azure AI integrates with your existing development environment, allowing you to use the same tools and workflows that you’re already familiar with.

Some of the benefits of using these AI tools and services for developers in GitHub and Visual Studio Code include:

Flexibility and choice: Choose the AI solution that best suits you, whether using pre-built models and APIs or building and training your own custom models. Choose the framework and language that you prefer, such as LangChain, Semantic Kernel, TensorFlow, PyTorch, Scikit-learn, Python, or R. You can even use the Azure OpenAI Service to access the latest GPT models from OpenAI. Additionally, folks can use the new Prompty format to work with prompts in their preferred environments (like Visual Studio Code and GitHub) all while using the trusted platform.

Productivity and efficiency: Simplify and accelerate the AI development process by using Visual Studio Code extensions and GitHub Actions. For example, use prompt flow to manage various versions of your flow assets like prompts, code, configurations, and environments via code repo, with tracked changes and rollback to previous versions, promoting a collaborative LLMOps ethos. For machine learning workloads, use GitHub Actions for Azure Machine Learning to automate model training, testing, and deployment.

Performance and scalability: Harness NVIDIA-optimized software from NVIDIA AI Enterprise, available in the Azure marketplace, to streamline workflows and embrace powerful AI capabilities. With support for remote development using Visual Studio Code extensions, you can write, debug, and optimize GPU-accelerated applications—including AI models—while using NVIDIA GPU-powered Azure Virtual Machines.

Innovation and creativity: Build applications that understand, interact, and learn from their users and environments, and that deliver personalized and engaging experiences. Use Azure AI to build a comprehensive generative AI stack and enrich your applications with retrieval-augmented generation, natural language processing, machine learning, and more.

Start building your most innovative applications

The strategic partnership between Microsoft and NVIDIA has significantly enhanced the Azure AI ecosystem. The integration of the NVIDIA AI Enterprise software platform combined with Azure’s AI toolsets and libraries ensures a robust and efficient environment for advancing your AI projects. Accelerate your time to deployment with the optimized NVIDIA Nemotron models, NVIDIA NIM inference microservices, Langchain and Hugging Face integrations, and APIs, inside of your Azure AI environment.

By building AI-enabled applications within the Microsoft ecosystem, developers can benefit from the productivity and efficiency gains that come from using a single, integrated set of tools and services. This can help reduce development time, support costs, and enhance collaboration and communication among team members. You can also benefit from the innovation and creativity that Azure AI enables, allowing you to create applications that understand, interact, and learn from users and environments, and deliver personalized and engaging experiences.

Learn more about how you can streamline development and build AI-enabled applications faster and easier with the combined power of Microsoft and NVIDIA.
The post Build AI-enabled applications with Azure AI and NVIDIA appeared first on Azure Blog.
Quelle: Azure

Accelerating AI app development with Azure AI and GitHub

Microsoft is empowering developers to become AI developers, bringing Azure AI industry leading models to the global GitHub community of more than 100 million

More than 60,000 organizations use Microsoft Azure AI today to explore the power of custom AI applications. However, the market is quickly moving from experimentation to scale, and we see more developers around the world becoming AI developers. With this shift, the needs of developers and their requirements to access and build with AI models and tools are evolving.

GitHub Models
The latest models with unique capabilities, performance metrics, and cost efficiencies for developers.

Learn more

To support this shift to scale, we are excited to partner with GitHub to empower their more than 100 million developers to build AI applications directly from Github.com with seamless integrations with Codespaces and Microsoft Visual Studio Code. Our collaboration starts today as we bring Azure AI’s leading model selection to developers through GitHub Models, along with simple APIs to empower responsible, production-ready AI applications.

For more insights into how GitHub Models can help you increase experimentation and accelerate your development cycles, all in GitHub, please read the blog from GitHub CEO Thomas Dohmke.

Simplifying AI development 

As AI model innovation accelerates, Azure remains committed to delivering the leading model selection and greatest model diversity to meet the unique cost, latency, design, and safety needs of AI developers. Today, we offer the largest and most complete model library in the market, including the latest models from OpenAI, Meta, Mistral and Cohere, introduced in July of this year, and updates to our own Phi-3 family of small language models. With GitHub Models, developers can now explore and utilize the latest models along with AI innovations and next-generation frontier models. This offering gives every developer the flexibility to choose the best combination of unique capabilities, performance metrics, and cost efficiencies.

While continuous model innovation brings more choice, it also brings complexity when selecting the right model for the right scenario. Today, developers have a range of options for cloud vs. edge, general-purpose vs. task specific, and more. On top of that, organizations often need multiple models to enable better quality, lower cost of goods sold, and to address complex use cases for each industry. GitHub Models opens the door for developers to experiment with multiple models, simplifying model experimentation and selection across the best of the Azure AI catalog, quickly comparing models, parameters, and prompts.

By making Azure AI an open, modular platform, we aim to help our customers rapidly go from idea to code to cloud. With Azure AI on GitHub, developers can do just that by utilizing Codespaces to set up a prototype or use the Prompty extension to generate code with GitHub Models directly in Microsoft Visual Studio Code.

In the coming months, we will expand our integration even further, bringing Azure AI’s language, vision, and multi-modal services to GitHub, along with additional Azure AI toolchain elements, further streamlining the AI application development process.

Integrating safety by default 

Developers building with AI want to be confident their AI applications are trustworthy, safe, and secure. GitHub Models gives developers a strong foundation from the start with built-in safety and security controls from Azure AI.

Azure AI works with model providers and other partners such as HiddenLayer to reduce emerging threats, from cybersecurity vulnerabilities, to malware, and other signs of tampering. And we have taken this further in GitHub Models by integrating Azure AI Content Safety for top foundation models including Azure OpenAI Service, Llama, and Mistral. Azure AI Content Safety enables built-in, real time protection for risks such as the generation of harmful content, copyright materials, hallucination, and new AI specific attacks such as jailbreaks and prompt injection attacks.

If developers want to go deeper, they can customize these controls in Azure AI, using evaluations to test and monitor their applications for ongoing quality and safety.

AI simplicity with a single API

Increased model selection gives developers the broadest range of options for the individual applications they are building. But each model naturally brings with it increased complexity. To counteract this, we’re making it incredibly easy for every developer to experiment with a range of models through the Azure AI model inference API. Using this single API, GitHub developers can now access a common set of capabilities to compare performance across a diverse set of foundational models in a uniform and consistent way, easily switching between models to compare performance without changing the underlying code.

The Azure AI Inference SDK provides client libraries in Python and JavaScript with support for C# and .NET coming soon. This SDK makes it easy to integrate AI into your applications by simplifying common tasks related to authentication, security and retries in your programming language of choice. You can get started today with Python and JavaScript samples.

Streamlining GitHub Enterprise access through Microsoft Azure 

Beyond these new integrations, we are also making it easier than ever for organizations to access GitHub Enterprise through Azure, combining GitHub’s cloud-native platform with Azure’s robust enterprise-grade security and scalability.

Organizations with an existing Azure subscription can purchase GitHub products via self-service, directly through Microsoft Sales or via Microsoft Cloud Solution Providers and can adjust the number of GitHub seats as needed to ensure efficient usage. Additionally, eligible organizations may take advantage of the Microsoft Azure Consumption Commitment (MACC) and Azure Commitment Discount (ACD). 

Companies can now spin-up a GitHub instance directly from the Azure Portal and connect their Microsoft Entra ID with GitHub to facilitate user management and access control. With an Azure subscription, you have all the necessary tools for creating an intelligent AI application, including access to GitHub’s complete range of services like repositories, Actions, Advanced Security, and Copilot. This makes it incredibly simple and efficient to give developers everything they need to build and deploy AI applications at scale.

We invite you to experience the power of this integrated end-to-end development experience. New customers can explore these capabilities with a free 30-day trial of GitHub Enterprise. 

We can’t wait to see what you will build with GitHub and Azure. 

Learn more about GitHub Models Launch. 

Explore and experiment with Azure AI models in GitHub. 

Get deeper technical details on GitHub Models. 

Use GitHub Codespaces to setup a prototype from your repo. 

Learn more about the free 30-day trial of GitHub Enterprise. 

Activate your GitHub free trial today

The post Accelerating AI app development with Azure AI and GitHub appeared first on Azure Blog.
Quelle: Azure

Embrace the future of container native storage with Azure Container Storage

We are thrilled to announce the general availability of Microsoft Azure Container Storage, the industry’s first platform-managed container native storage service in the public cloud. With Kubernetes driving cloud evolution, we are witnessing a transformative shift as enterprises move from virtual machines (VMs) to containers, optimizing for scalability, flexibility, and cost efficiency. We introduce Azure Container Storage to meet these demands, providing best in class price performance for hosting stateful containers on cloud-based storage and delivering lowest latency on locally attached storage.

Azure Container Storage joins our suite of container services, tightly integrating with Kubernetes and simplifying stateful workload management across Azure’s set of comprehensive storage offerings. Previously, customers needed to retrofit stateful workloads to VM-centric storage options with scalability constraints or deploy self-managed open-source container storage solutions. Since Azure Container Storage is built purposefully for Azure Kubernetes Service (AKS), it simplifies the process, allowing developers to focus on innovating and running applications without worrying about managing storage. With the ability to perform all storage operations directly through Kubernetes APIs—such as creating persistent volumes and scaling up capacity on demand, it eliminates the need to interact with control plane APIs of the underlying infrastructure.

Azure Container Storage

Try it today

Azure Container Storage also streamlines storage management across multiple backing storage options. With its general availability, Azure Container Storage supports Ephemeral Disks (local NVMe and temp SSD) and Azure Disks, just the start in our journey to transform the container storage landscape. Ephemeral Disks support marks a pivotal moment for container users, providing the most comprehensive volume management support for containers on local storage in the cloud. Beyond basic persistent volume (PV) provisioning, Azure Container Storage offers built-in capabilities such as snapshots and autoscaling, capabilities that cannot be found outside of Azure.  

During preview, customers have already begun taking advantage of Azure Container Storage to evolve their business-critical, next-generation solutions. Whether it be optimizing Redpanda cluster performance on Ephemeral Disks or scaling past existing persistent volume limits for Postgres workloads on Azure Disks, Azure Container Storage supports a wide range of workloads. For building stateful applications operating containers, this is just the beginning. Shortly after general availability, we will expand our offerings to include Elastic SAN and later, options like Azure Blobs and Azure Files for shared storage use cases.  

A cloud native solution for all use cases

Azure Container Storage ensures essential resiliency and security for every workload through built-in resiliency design and security enforcements. 

Built-in resiliency: Easily run highly available stateful applications on Azure Container Storage and protect against zonal failures on all levels of the resource hierarchy. You can choose between zone-redundant storage (ZRS) options or multi-zone storage pools on local-redundant storage (LRS) to deliver a highly available solution across zones. For local storage, we optimally place a pod’s persistent volumes on ephemeral disks that exist on the same node as the AKS pod, reducing the number of failure points that could impact your application’s runtime. Moreover, we offer the best balance for availability, cost, and performance—providing the most cost-efficient block storage offering on the cloud with multi-zonal high availability support and sub millisecond read latency. 

Security by default: Security is our top priority. We offer server-side encryption (SSE) with platform-managed keys by default and enforce network security per respective backing storage options. Customers can further enhance security through extensive options, such as SSE with customer-managed keys, per their security standards.

Modernizing existing applications

For any enterprise looking to modernize its applications, Azure Container Storage consolidates management across familiar block storage offerings, simplifies the movement of workloads, and provides continuity in backup and disaster recovery.  

We streamline and consolidate the management experience across our comprehensive portfolio of familiar Azure block storage offerings. Rather than needing to certify and manage multiple container orchestration solutions for each storage resource you deploy, Azure Container Storage efficiently coordinates volume provisioning within a storage pool, a concept we introduce to group storage resources into a unified resource for your AKS cluster. This storage pool can be backed by your preferred storage option, empowering you to choose the most cost-efficient resource tailored to your specific workload performance requirements. For example, Ephemeral Disk, newly introduced as a supported block storage offering for containers, is well-suited for latency-sensitive workloads that benefit from local NVMe or temp SSD storage. KPN, a Dutch telecommunications company, shared their positive experience using Azure Container Storage with local NVMe to host a mail solution on AKS: 

“With Azure Container Storage, we have been able to achieve improved performance in our KPN consumer mail workload by leveraging ephemeral disks and taking advantage of the pooling of resources that Azure Container Storage enables. Instead of the manual configuration of storage, we can focus on running our workloads, and Azure Container Storage will take care of auto-discovering and formatting the NVMe disks, making it simple to use and in line with the Kubernetes way.” 
—Peter Teeninga, Cloud Architect, Personal Cloud by KPN

To make your journey to the cloud as painless as possible we partnered with CloudCasa, a key player in Kubernetes data mobility, to simplify mission-critical migration to the cloud. To continue supporting your cloud estate, we partnered with Kasten, the leading service for data protection for Kubernetes, offering robust backup and disaster recovery capabilities. For more details on our data migration and backup experience provided through our partners, please refer to the later section. 

Building cloud native applications

For application developers building solutions in the cloud, Azure Container Storage offers seamless integration with Kubernetes, providing a container-native experience designed for scalability from the ground up. This ensures that your applications can grow easily and cost-efficiently over time. By supporting industry-standard protocols, such as NVMe-of and iSCSI, we simplify interoperability, providing additional performance options. For instance, you can take advantage of the lower persistent volume attach and detach latencies these protocols offer and achieve rapid scale-out and fast failover. Azure Container Storage allows customers to attach more storage resources to a single VM, increasing the limit to 75 volumes for any VM sizes. The added flexibility increases customers’ ability to optimize Azure resources to meet their cost and performance goals. Sesam, a Norwegian data synchronization and management company, has effectively leveraged this capability to reduce costs by scaling up their persistent volumes more efficiently: 

“Azure Container Storage (ACS) has enabled us to achieve lower total cost of ownership in our workload. We have a large number of pods that need their own persistent volumes, and through the use of Azure Disks and storage pool resource pooling, we are able to fulfill this in a more cost-effective manner, without hitting the limits on the number of disks that can be attached to a node. In practice this makes us able to allocate capacity more easily and more efficiently.”
—Geir Ove Grønmo, Product Manager, Sesam.io 

Data migration and backup support through our trusted partners

Highly efficient and operational storage management is the baseline experience Azure Container Storage strives for. Azure Container Storage tightly integrates with two key third-party solutions—CloudCasa and Kasten—to offer you an integrated migrate, backup, and disaster recovery experience for workloads hosted on stateful containers.  

With the ability to automatically recreate an entire cluster, CloudCasa centralizes the management of cluster recovery and migration, making it easy to move your existing Kubernetes workloads to and within AKS. To modernize your existing workloads on Azure, simply do a full backup of the existing storage resources then set up a restore, indicating Azure Container Storage as the new storage resource for your cluster. 

“With Azure Container Storage, Microsoft has removed much of the management burden from Kubernetes storage, allowing development and DevOps teams to focus on their data and applications. This approach enables organizations to more easily operate stateful production applications at scale. We are pleased to have worked with the Azure Container Storage team to certify CloudCasa for backup and recovery of stateful applications running on it, and to provide a jointly tested solution for easy migration to it.”
—Bob Adair, Head of Product Management, CloudCasa By Catalogic

Kasten automates the end-to-end workflow of backup and disaster recovery, protecting your Kubernetes clusters and application operations. When you deploy your storage pool in Azure Container Storage, you can enable Kasten during snapshot setup. Using dynamic policies, Kasten helps you manage backups at scale in a crash-consistent manner. 

“With Azure Container Storage and Kasten by Veeam, organizations can maximize performance, flexibility and resiliency, while protecting their cloud native workloads from ransomware attacks. Kasten by Veeam collaborated with Microsoft Product and Engineering teams to validate provisioning, volume snapshot and restore capabilities on Azure Container Storage to ensure joint Microsoft and Kasten by Veeam customers can backup, protect, and migrate their stateful workloads to Azure Kubernetes Service (AKS). Through our strategic partnership, we simplify organizations’ cloud journeys without sacrificing performance, scalability or resiliency.” 
—Matt Slotten, Principal Solution Architect, Cloud Native Partnerships Kasten by Veeam  

What is new with general availability?

Our announcement builds on the updates we’ve shared throughout our preview, highlighting several differentiated capabilities. We’ve enhanced the resiliency of stateful containers with multi-zone storage pools and volume replication for local NVMe storage pools to protect against availability loss during single node failures. We’ve also added snapshot support across all storage options for backup and disaster recovery. Additionally, we’ve expanded the Ephemeral Disk portfolio from local NVMe to include temp SSD support, enhancing cost efficiency for use cases that can leverage directly attached local storage. With this announcement, we are excited to introduce three new capabilities that will further increase the resiliency and performance of running stateful workloads: 

Enhance the resiliency of your persistent volumes hosted on local NVMe storage (L-series ephemeral disks) with replication support.

Improved persistent volume recovery after a restart of an Azure Kubernetes Service (AKS) cluster.

Customize the performance of your local NVMe storage with new performance tier options. 

Learn more about Azure Container Storage

Get started with installing Azure Container Storage to your AKS cluster! For a comprehensive guide, watch our step-by-step walkthrough video. You can also explore workload samples from our newly launched community repository to create your first stateful application. To learn more, refer to our AKS engineering blog. We encourage everyone to contribute and share your insights as you explore our newest storage offering.  

If you have any questions, please reach out to AskContainerStorage@microsoft.com. Embrace the future of stateful containers with Azure and unlock new possibilities! 

Explore Azure Container Storage capabilities

The post Embrace the future of container native storage with Azure Container Storage appeared first on Azure Blog.
Quelle: Azure

OpenAI’s fastest model, GPT-4o mini is now available on Azure AI

We are also announcing safety features by default for GPT-4o mini, expanded data residency and service availability, plus performance upgrades to Microsoft Azure OpenAI Service.

GPT-4o mini allows customers to deliver stunning applications at a lower cost with blazing speed. GPT-4o mini is significantly smarter than GPT-3.5 Turbo—scoring 82% on Measuring Massive Multitask Language Understanding (MMLU) compared to 70%—and is more than 60% cheaper.1 The model delivers an expanded 128K context window and integrates the improved multilingual capabilities of GPT-4o, bringing greater quality to languages from around the world.

GPT-4o mini, announced by OpenAI today, is available simultaneously on Azure AI, supporting text processing capabilities with excellent speed and with image, audio, and video coming later. Try it at no cost in the Azure OpenAI Studio Playground.

Azure AI
Where innovators are creating the future

Try for free

We’re most excited about the new customer experiences that can be enhanced with GPT-4o mini, particularly streaming scenarios such as assistants, code interpreter, and retrieval which will benefit from this model’s capabilities. For instance, we observed the incredible speed while testing GPT-4o mini on GitHub Copilot, an AI pair programmer that assists you by delivering code completion suggestions in the tiny pauses between keystrokes, rapidly updating recommendations with each new character typed.

We are also announcing updates to Azure OpenAI Service, including extending safety by default for GPT-4o mini, expanded data residency, and worldwide pay-as-you-go availability, plus performance upgrades. 

Azure AI brings safety by default to GPT-4o mini

Safety continues to be paramount to the productive use and trust that we and our customers expect.

We’re pleased to confirm that our Azure AI Content Safety features—including prompt shields and protected material detection— are now ‘on by default’ for you to use with GPT-4o mini on Azure OpenAI Service.

We have invested in improving the throughput and speed of the Azure AI Content Safety capabilities—including the introduction of an asynchronous filter—so you can maximize the advancements in model speed while not compromising safety. Azure AI Content Safety is already supporting developers across industries to safeguard their generative AI applications, including game development (Unity), tax filing (H&R Block), and education (South Australia Department for Education).

In addition, our Customer Copyright Commitment will apply to GPT-4o mini, giving peace of mind that Microsoft will defend customers against third-party intellectual property claims for output content.

Azure AI now offers data residency for all 27 regions

From day one, Azure OpenAI Service has been covered by Azure’s data residency commitments.

Azure AI gives customers both flexibility and control over where their data is stored and where their data is processed, offering a complete data residency solution that helps customers meet their unique compliance requirements. We also provide choice over the hosting structure that meets business, application, and compliance requirements. Regional pay-as-you-go and Provisioned Throughput Units (PTUs) offer control over both data processing and data storage.

We’re excited to share that Azure OpenAI Service is now available in 27 regions including Spain, which launched earlier this month as our ninth region in Europe.

Azure AI announces global pay-as-you-go with the highest throughput limits for GPT-4o mini

GPT-4o mini is now available using our global pay-as-you-go deployment at 15 cents per million input tokens and 60 cents per million output tokens, which is significantly cheaper than previous frontier models.

We are pleased to announce that the global pay-as-you-go deployment option is generally available this month, allowing customers to pay for the resources they consume, making it flexible for variable workloads, while traffic is routed globally to provide higher throughput, and still offering control over where data resides at rest.

Additionally, we recognize that one of the challenges customers face with new models is not being able to upgrade between model versions in the same region as their existing deployments. Now, with global pay-as-you-go deployments, customers will be able to upgrade from existing models to the latest models.

Global pay-as-you-go offers customers the highest possible scale, offering 15M tokens per minute (TPM) throughput for GPT-4o mini and 30M TPM throughput for GPT-4o. Azure OpenAI Service offers GPT-4o mini with 99.99% availability and the same industry leading speed as our partner OpenAI.

Azure AI offers leading performance and flexibility for GPT-4o mini

Azure AI is continuing to invest in driving efficiencies for AI workloads across Azure OpenAI Service.

GPT-4o mini comes to Azure AI with availability on our Batch service this month. Batch delivers high throughput jobs with a 24-hour turnaround at a 50% discount rate by using off-peak capacity. This is only possible because Microsoft runs on Azure AI, which allows us to make off-peak capacity available to customers.

We are also releasing fine-tuning for GPT-4o mini this month which allows customers to further customize the model for your specific use case and scenario to deliver exceptional value and quality at unprecedented speeds. Following our update last month to switch to token based billing for training, we’ve reduced the hosting charges by up to 43%. Paired with our low price for inferencing, this makes Azure OpenAI Service fine-tuned deployments the most cost-effective offering for customers with production workloads.

With more than 53,000 customers turning to Azure AI to deliver breakthrough experiences at impressive scale, we’re excited to see the innovation from companies like Vodafone (customer agent solution), the University of Sydney (AI assistants), and GigXR (AI virtual patients). More than 50% of the Fortune 500 are building their applications with Azure OpenAI Service.

We can’t wait to see what our customers do with GPT-4o mini on Azure AI!

1GPT-4o mini: advancing cost-efficient intelligence | OpenAI
The post OpenAI’s fastest model, GPT-4o mini is now available on Azure AI appeared first on Azure Blog.
Quelle: Azure

Microsoft Cost Management updates—June 2024

Whether you’re a new student, a thriving startup, or the largest enterprise, you have financial constraints, and you need to know what you’re spending, where it’s being spent, and how to plan for the future. Nobody wants a surprise when it comes to the bill, and this is where Microsoft Cost Management comes in. 

We’re always looking for ways to learn more about your challenges and how Microsoft Cost Management can help you better understand where you’re accruing costs in the cloud, identify and prevent bad spending patterns, and optimize costs to empower you to do more with less. Here are a few of the latest improvements and updates. 

FOCUS 1.0 support in Exports

Cost card in Azure portal

Kubernetes cost views (New entry point)

Pricing updates on Azure.com

New ways to save money with Microsoft Cloud

Documentation updates

Before we dig into details, kudos to the FinOps foundation for successfully hosting FinOps X 2024 in San Diego, California last month. Microsoft participated as a platinum sponsor for a second consecutive year. Our team members enjoyed connecting with customers and getting insights into their FinOps practice. We also shared our vision of simplifying FinOps through AI, demonstrated in this short video—Bring your FinOps practice into the era of AI.

For all our updates from FinOps X 2024, refer to the blog post by my colleague, Michael Flanakin, who also serves in the FinOps Technical Advisory Council. 

FOCUS 1.0 support in exports 

As you may already know, the FinOps foundation announced the general availability of the FinOps Cost and Usage Specification (FOCUS) Version 1 in June 2024. We are thrilled to announce that you can get the newly released version through exports experience in the Microsoft Azure portal or the REST API. You can review the updated schema and the differences from the previous version in this Microsoft Learn article. We will continue to support the ability for you to export the preview version of the FOCUS dataset.

For all the datasets supported through exports and to learn more about the functionality, refer to our documentation.

Cost card in Azure portal 

You have always had the ability to estimate costs for Azure services using the pricing calculator so that you can better plan your expenses. Now, we are excited to announce the estimation capability within the Azure portal itself. Engineers now can quickly get a breakdown of their estimated virtual machine (VM) costs before deploying them and adjust as needed. This new experience is currently available only for VMs running on pay-as-you-go subscriptions and will be expanded in the future. Empowering engineers with cost data without disrupting their workflow enables them to make the right decisions for managing their spending and drives accountability.

Kubernetes cost views (new entry point) 

I had spoken about the Azure Kubernetes Service cost views in our November 2023 blog post. We know how important it is for you to get visibility into the granular costs of running your clusters. To make it even easier to access these cost views, we have added an entry point to the cluster page itself. Engineers and admins who are already on the cluster page potentially making configuration changes or just monitoring their cluster, can now quickly reference the costs as well.

Pricing updates on Azure.com

We’ve been working hard to make some changes to our Azure pricing experiences, and we’re excited to share them with you. These changes will help make it easier for you to estimate the costs of your solutions. 

We’ve expanded our global reach with pricing support for new Azure regions, including Spain Central and Mexico Central. 

We’ve introduced pricing for several new services—enhancing our Azure portfolio—including Trusted Signing, Azure Advanced Container Networking Services, Azure AI Studio, Microsoft Entra External ID, and Azure API Center (now available on the Azure API Management pricing calculator.)

The Azure pricing calculator now supports a new example to help you get started with estimating costs for your Azure Arc enabled servers scenarios.  

Azure AI has seen significant updates with pricing support for Basic Video Indexing Analysis for Azure AI Video Indexer, new GPT-4o models and improved Fine Tuning models for Azure OpenAI Service, the deprecation of S2 to S4 volume discount tiers for Azure AI Translator, and the introduction of standard fast transcription and video dubbing, both in preview, for Azure AI Speech.  

We’re thrilled to announce new features in both preview and general availability stages with Azure flex consumption (preview) for Azure Functions, Advanced messaging (generally available) for Azure Communication Services, and Azure API Center (generally available) for Azure API Management, and AKS Automatic (preview) for Azure Kubernetes.  

We’ve made comprehensive updates to our pricing models to reflect the latest offerings and ensure you have the most accurate information, including changes to

Azure Bastion: Added pricing for premium and developer stock-keeping units (SKUs).

Virtual Machines: Removal of CentOS for Linux, added 5 year reserved instances (RI) pricing for the Hx and HBv4 series, as well as pricing for the new NDsr H100 v5 and E20 v4 series.

Databricks: Added pricing for all-purpose serverless compute jobs.

Azure Communication Gateway: Added pricing for the new “Lab” SKU.

Azure Virtual Desktop for Azure Stack HCI: Pricing added to the Azure Virtual Desktop calculator.

Azure Data Factory: Added RI pricing for Dataflow.

Azure Container Apps: Added pricing for dynamic session feature.

Azure Backup: Added pricing for the new comprehensive Blob Storage data protection feature.

 Azure SQL Database: Added 3 year RI pricing for hyperscale series, zone redundancy pricing for hyperscale elastic pools, and disaster recovery pricing options for single database.

Azure PostgreSQL: Added pricing for Premium SSD v2.

Defender for Cloud: Added pricing for the “Pre-Purchase Plan”.

Azure Stack Hub: Added pricing for site recovery.

Azure Monitor: Added pricing for pricing for workspace replication as well as data restore in the pricing calculator.

We’re constantly working to improve our pricing tools and make them more accessible and user-friendly. We hope you find these changes helpful in estimating the costs for your Azure solutions. If you have any feedback or suggestions for future improvements, please let us know! 

New ways to save money in the Microsoft Cloud 

VM Hibernation is now generally available 

Documentation updates 

Here are a few documentation updates you might be interested in: 

Update: Understand Cost Management data  

Update: Azure Hybrid Benefit documentation 

Update: Automation for partners  

Update: View and download your Microsoft Azure invoice  

Update: Tutorial: Create and manage exported data  

Update: Automatically renew reservations  

Update: Changes to the Azure reservation exchange policy  

Update: Migrate from EA Marketplace Store Charge API

Update: Azure product transfer hub  

Update: Get started with your Microsoft Partner Agreement billing account  

Update: Manage billing across multiple tenants using associated billing tenants

Want to keep an eye on all documentation updates? Check out the Cost Management and Billing documentation change history in the azure-docs repository on GitHub. If you see something missing, select Edit at the top of the document and submit a quick pull request. You can also submit a GitHub issue. We welcome and appreciate all contributions! 

What’s next? 

These are just a few of the big updates from last month. Don’t forget to check out the previous Microsoft Cost Management updates. We’re always listening and making constant improvements based on your feedback, so please keep the feedback coming.  

Follow @MSCostMgmt on X and subscribe to the Microsoft Cost Management YouTube channel for updates, tips, and tricks.
The post Microsoft Cost Management updates—June 2024 appeared first on Azure Blog.
Quelle: Azure

Harnessing the full power of AI in the cloud: The economic impact of migrating to Azure for AI readiness

As the digital landscape rapidly evolves, AI stands at the forefront, driving significant innovation across industries. However, to fully harness the power of AI, businesses must be AI-ready; this means having defined use-cases for their AI apps, being equipped with modernized databases that seamlessly integrate with AI models, and most importantly, having the right infrastructure in place to power and realize their AI ambitions. When we talk to our customers, many have expressed that traditional on-premises systems often fall short in providing the necessary scalability, stability, and flexibility required for modern AI applications.

A recent Forrester study1, commissioned by Microsoft, surveyed over 300 IT leaders and interviewed representatives from organizations globally to learn about their experience migrating to Azure and if that enhanced their AI impact. The results showed that migrating from on-premises infrastructure to Azure can support AI-readiness in organizations, with lower costs to stand up and consume AI services plus improved flexibility and ability to innovate with AI. Here’s what you should know before you start leveraging AI in the cloud.

Challenges faced by customers with on-premises infrastructure

Many organizations who attempted to implement AI on-premises encountered significant challenges with their existing infrastructure. The top challenges with on-premises infrastructure cited were:

Aging and costly infrastructure: Maintaining or replacing aging on-premises systems is both expensive and complex, diverting resources from strategic initiatives.

Infrastructure instability: Unreliable infrastructure impacts business operations and profitability, creating an urgent need for a more stable solution.

Lack of scalability: Traditional systems often lack the scalability required for AI and machine learning (ML) workloads, necessitating substantial investments for infrequent peak capacity needs.

High capital costs: The substantial upfront costs of on-premises infrastructure limit flexibility and can be a barrier to adopting new technologies.

Forrester’s study highlights that migrating to Azure effectively addresses these issues, enabling organizations to focus on innovation and business growth rather than infrastructure maintenance.

Azure AI
Where innovators are creating for the future

Try for free today

Key Benefits

Improved AI-readiness: When asked whether being on Azure helped with AI-readiness, 75% of survey respondents with Azure infrastructure reported that migrating to the cloud was essential or significantly reduced barriers to AI and ML adoption. Interviewees noted that the AI services are readily available in Azure, and colocation of data and infrastructure that is billed only on consumption helps teams test and deploy faster with less upfront costs. This was summarized well by an interviewee who was the head of cloud and DevOps for a banking company:

We didn’t have to go and build an AI capability. It’s up there, and most of our data is in the cloud as well. And from a hardware-specific standpoint, we don’t have to go procure special hardware to run AI models. Azure provides that hardware today.”
—Head of cloud and DevOps for global banking company

Cost Efficiency: Migrating to Azure significantly reduces the initial costs of deploying AI and the cost to maintain AI, compared to on-premises infrastructure. The study estimates that organizations experience financial benefits of USD $500 thousand plus over three years and 15% lower costs to maintain AI/ML in Azure compared to on-premises infrastructure.

Flexibility and scalability to build and maintain AI: As mentioned above, lack of scalability was a common challenge for survey respondents with on-premises infrastructure as well. Respondents with on-premises infrastructure cited lack of scalability with existing systems as a challenge when deploying AI and ML at 1.5 times the rate of those with Azure cloud infrastructure.

Interviewees shared that migrating to Azure gave them easy access to new AI services and the scalability they needed to test and build them out without worrying about infrastructure. 90% of survey respondents with Azure cloud infrastructure agreed or strongly agreed they have the flexibility to build new AI and ML applications. This is compared to 43% of respondents with on-premises infrastructure. A CTO for a healthcare organization said:

After migrating to Azure all the infrastructure problems have disappeared, and that’s generally been the problem when you’re looking at new technologies historically.”
—CTO for a healthcare organization

They explained that now, “The scalability [of Azure] is unsurpassed, so it adds to that scale and reactiveness we can provide to the organization.” They also said: “When we were running on-prem, AI was not as easily accessible as it is from a cloud perspective. It’s a lot more available, accessible, and easy to start consuming as well. It allowed the business to start thinking outside of the box because the capabilities were there.”

Holistic organizational improvement: Beyond the cost and performance benefits, the study found that migration to Azure accelerated innovation with AI by having an impact on the people at all levels of an organization:

Bottoms-up: skilling and reinvestment in employees. Forrester has found that investing in employees to build understanding, skills, and ethics is critical to successfully using AI. Both interviewees and survey respondents expressed difficulty finding skilled resources to support AI and ML initiatives at their organizations.

Migrating to the cloud freed up resources and changed the types of work needed, allowing organizations to upskill employees and reinvest resources in new initiatives like AI. A VP of AI for a financial services organization shared: “As we have gone along this journey, we have not reduced the number of engineers as we have gotten more efficient, but we’re doing more. You could say we’ve invested in AI, but everything we have invested—my entire team—none of these people were new additions. These are people we could redeploy because we’re doing everything else more efficiently.”

Top-down: created a larger culture of innovation at organizations. As new technologies—like AI—disrupt entire industries, companies need to excel at all levels of innovation to succeed, including embracing platforms and ecosystems that help drive innovation. For interviewees, migrating to the cloud meant that new resources and capabilities were readily available, making it easier for organizations to take advantage of new technologies and opportunities with reduced risk.

Survey data indicates that 77% of respondents with Azure cloud infrastructure find it easier to innovate with AI and ML, compared to only 34% of those with on-premises infrastructure. An executive head of cloud and DevOps for a banking organization said: “Migrating to Azure changes the mindset from an organization perspective when it comes to innovation, because services are easily available in the cloud. You don’t have to go out to the market and look for them. If you look at AI, originally only our data space worked on it, whereas today, it’s being used across the organization because we were already in the cloud and it’s readily available.”

Learn more about migrating to Azure for AI-readiness

Forrester’s study underscores the significant economic and strategic advantages of migrating to Azure for be AI-ready. Lower costs, increased innovation, better resource allocation, and improved scalability make migration to Azure a clear choice for organizations looking to thrive in the AI-driven future.

Ready to get started with your migration journey? Here are some resources to learn more:

Read the full Forrester TEI study on migration to Azure for AI-readiness.

The solutions that can support your organization’s migration and modernization goals.

Our hero offerings that provide funding, unique offers, expert support, and best practices for all use-cases, from migration to innovation with AI.

Learn more in our e-book and video on how to migrate to innovate.

Refrences

Forrester Consulting The Total Economic Impact™ Of Migrating to Microsoft Azure For AI-Readiness, commissioned by Microsoft, June 2024

The post Harnessing the full power of AI in the cloud: The economic impact of migrating to Azure for AI readiness appeared first on Azure Blog.
Quelle: Azure

Announcing Phi-3 fine-tuning, new generative AI models, and other Azure AI updates to empower organizations to customize and scale AI applications

AI is transforming every industry and creating new opportunities for innovation and growth. But, developing and deploying AI applications at scale requires a robust and flexible platform that can handle the complex and diverse needs of modern enterprises and allow them to create solutions grounded in their organizational data. That’s why we are excited to announce several updates to help developers quickly create customized AI solutions with greater choice and flexibility leveraging the Azure AI toolchain:

Serverless fine-tuning for Phi-3-mini and Phi-3-medium models enables developers to quickly and easily customize the models for cloud and edge scenarios without having to arrange for compute.

Updates to Phi-3-mini including significant improvement in core quality, instruction-following, and structured output, enabling developers to build with a more performant model without additional cost.

Same day shipping earlier this month of the latest models from OpenAI (GPT-4o mini), Meta (Llama 3.1 405B), Mistral (Large 2) to Azure AI to provide customers greater choice and flexibility.

Unlocking value through model innovation and customization  

In April, we introduced the Phi-3 family of small, open models developed by Microsoft. Phi-3 models are our most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up. As developers look to tailor AI solutions to meet specific business needs and improve quality of responses, fine-tuning a small model is a great alternative without sacrificing performance. Starting today, developers can fine-tune Phi-3-mini and Phi-3-medium with their data to build AI experiences that are more relevant to their users, safely, and economically.

Given their small compute footprint, cloud and edge compatibility, Phi-3 models are well suited for fine-tuning to improve base model performance across a variety of scenarios including learning a new skill or a task (e.g. tutoring) or enhancing consistency and quality of the response (e.g. tone or style of responses in chat/Q&A). We’re already seeing adaptations of Phi-3 for new use cases.

Phi-3 models
A family of powerful, small language models (SLMs) with groundbreaking performance at low cost and low latency

Try today

Microsoft and Khan Academy are working together to help improve solutions for teachers and students across the globe. As part of the collaboration, Khan Academy uses Azure OpenAI Service to power Khanmigo for Teachers, a pilot AI-powered teaching assistant for educators across 44 countries and is experimenting with Phi-3 to improve math tutoring. Khan Academy recently published a research paper highlighting how different AI models perform when evaluating mathematical accuracy in tutoring scenarios, including benchmarks from a fine-tuned version of Phi-3. Initial data shows that when a student makes a mathematical error, Phi-3 outperformed most other leading generative AI models at correcting and identifying student mistakes.

And we’ve fine-tuned Phi-3 for the device too. In June, we introduced Phi Silica to empower developers with a powerful, trustworthy model for building apps with safe, secure AI experiences. Phi Silica builds on the Phi family of models and is designed specifically for the NPUs in Copilot+ PCs. Microsoft Windows is the first platform to have a state-of-the-art small language model (SLM) custom built for the Neural Processing Unit (NPU) and shipping inbox.

You can try fine-tuning for Phi-3 models today in Azure AI.

I am also excited to share that our Models-as-a-Service (serverless endpoint) capability in Azure AI is now generally available. Additionally, Phi-3-small is now available via a serverless endpoint so developers can quickly and easily get started with AI development without having to manage underlying infrastructure. Phi-3-vision, the multi-modal model in the Phi-3 family, was announced at Microsoft Build and is available through Azure AI model catalog. It will soon be available via a serverless endpoint as well. Phi-3-small (7B parameter) is available in two context lengths 128K and 8K whereas Phi-3-vision (4.2B parameter) has also been optimized for chart and diagram understanding and can be used to generate insights and answer questions.

We are seeing great response from the community on Phi-3. We released an update for Phi-3-mini last month that brings significant improvement in core quality and instruction following. The model was re-trained leading to substantial improvement in instruction following and support for structured output. We also improved multi-turn conversation quality, introduced support for <|system|> prompts, and significantly improved reasoning capability.

The table below highlights improvements across instruction following, structured output, and reasoning.

Benchmarks Phi-3-mini-4k Phi-3-mini-128k Apr ’24 release Jun ’24 update Apr ’24 release Jun ’24 update Instruction Extra Hard 5.7 6.0 5.7 5.9 Instruction Hard 4.9 5.1 5 5.2 JSON Structure Output 11.5 52.3 1.9 60.1 XML Structure Output 14.4 49.8 47.8 52.9 GPQA 23.7 30.6 25.9 29.7 MMLU 68.8 70.9 68.1 69.7 Average 21.7 35.8 25.7 37.6 

We continue to make improvements to Phi-3 safety too. A recent research paper highlighted Microsoft’s iterative “break-fix” approach to improving the safety of the Phi-3 models which involved multiple rounds of testing and refinement, red teaming, and vulnerability identification. This method significantly reduced harmful content by 75% and enhanced the models’ performance on responsible AI benchmarks. 

Expanding model choice, now with over 1600 models available in Azure AI

With Azure AI, we’re committed to bringing the most comprehensive selection of open and frontier models and state-of-the-art tooling to help meet customers’ unique cost, latency, and design needs. Last year we launched the Azure AI model catalog where we now have the broadest selection of models with over 1,600 models from providers including AI21, Cohere, Databricks, Hugging Face, Meta, Mistral, Microsoft Research, OpenAI, Snowflake, Stability AI and others. This month we added—OpenAI’s GPT-4o mini through Azure OpenAI Service, Meta Llama 3.1 405B, and Mistral Large 2.

Continuing the momentum today we are excited to share that Cohere Rerank is now available on Azure. Accessing Cohere’s enterprise-ready language models on Azure AI’s robust infrastructure enables businesses to seamlessly, reliably, and safely incorporate cutting-edge semantic search technology into their applications. This integration allows users to leverage the flexibility and scalability of Azure, combined with Cohere’s highly performant and efficient language models, to deliver superior search results in production.

TD Bank Group, one of the largest banks in North America, recently signed an agreement with Cohere to explore its full suite of large language models (LLMs), including Cohere Rerank.

At TD, we’ve seen the transformative potential of AI to deliver more personalized and intuitive experiences for our customers, colleagues and communities, we’re excited to be working alongside Cohere to explore how its language models perform on Microsoft Azure to help support our innovation journey at the Bank.”
Kirsti Racine, VP, AI Technology Lead, TD.

Atomicwork, a digital workplace experience platform and longtime Azure customer, has significantly enhanced its IT service management platform with Cohere Rerank. By integrating the model into their AI digital assistant, Atom AI, Atomicwork has improved search accuracy and relevance, providing faster, more precise answers to complex IT support queries. This integration has streamlined IT operations and boosted productivity across the enterprise. 

The driving force behind Atomicwork’s digital workplace experience solution is Cohere’s Rerank model and Azure AI Studio, which empowers Atom AI, our digital assistant, with the precision and performance required to deliver real-world results. This strategic collaboration underscores our commitment to providing businesses with advanced, secure, and reliable enterprise AI capabilities.”
Vijay Rayapati, CEO of Atomicwork

Command R+, Cohere’s flagship generative model which is also available on Azure AI, is purpose-built to work well with Cohere Rerank within a Retrieval Augmented Generation (RAG) system. Together they are capable of serving some of the most demanding enterprise workloads in production. 

Earlier this week, we announced that Meta Llama 3.1 405B along with the latest fine-tuned Llama 3.1 models, including 8B and 70B, are now available via a serverless endpoint in Azure AI. Llama 3.1 405B can be used for advanced synthetic data generation and distillation, with 405B-Instruct serving as a teacher model and 8B-Instruct/70B-Instruct models acting as student models. Learn more about this announcement here.

Mistral Large 2 is now available on Azure, making Azure the first leading cloud provider to offer this next-gen model. Mistral Large 2 outperforms previous versions in coding, reasoning, and agentic behavior, standing on par with other leading models. Additionally, Mistral Nemo, developed in collaboration with NVIDIA, brings a powerful 12B model that pushes the boundaries of language understanding and generation. Learn More.

And last week, we brought GPT-4o mini to Azure AI alongside other updates to Azure OpenAI Service, enabling customers to expand their range of AI applications at a lower cost and latency with improved safety and data deployment options. We will announce more capabilities for GPT-4o mini in coming weeks. We are also happy to introduce a new feature to deploy chatbots built with Azure OpenAI Service into Microsoft Teams.  

Enabling AI innovation safely and responsibly  

Building AI solutions responsibly is at the core of AI development at Microsoft. We have a robust set of capabilities to help organizations measure, mitigate, and manage AI risks across the AI development lifecycle for traditional machine learning and generative AI applications. Azure AI evaluations enable developers to iteratively assess the quality and safety of models and applications using built-in and custom metrics to inform mitigations. Additional Azure AI Content Safety features—including prompt shields and protected material detection—are now “on by default” in Azure OpenAI Service. These capabilities can be leveraged as content filters with any foundation model included in our model catalog, including Phi-3, Llama, and Mistral. Developers can also integrate these capabilities into their application easily through a single API. Once in production, developers can monitor their application for quality and safety, adversarial prompt attacks, and data integrity, making timely interventions with the help of real-time alerts.

Azure AI uses HiddenLayer Model Scanner to scan third-party and open models for emerging threats, such as cybersecurity vulnerabilities, malware, and other signs of tampering, before onboarding them to the Azure AI model catalog. The resulting verifications from Model Scanner, provided within each model card, can give developer teams greater confidence as they select, fine-tune, and deploy open models for their application. 

We continue to invest across the Azure AI stack to bring state of the art innovation to our customers so you can build, deploy, and scale your AI solutions safely and confidently. We cannot wait to see what you build next.

Stay up to date with more Azure AI news

Watch this video to learn more about Azure AI model catalog.

Listen to the podcast on Phi-3 with lead Microsoft researcher Sebastien Bubeck.

The post Announcing Phi-3 fine-tuning, new generative AI models, and other Azure AI updates to empower organizations to customize and scale AI applications appeared first on Azure Blog.
Quelle: Azure