Accelerate your AI applications with Azure NC A100 v4 virtual machines

Real-world AI has revolutionized and changed how people live during the past decade, including media and entertainment, healthcare and life science, retail, automotive, finance service, manufacturing, and oil and gas. Speaking to a smart home device, browsing social media with recommended content, or taking a ride with a self-driving vehicle is no longer in the future. With the ease of your smartphone, you can now deposit checks without going to the bank? All of these advances have been made possible through new AI breakthroughs in software and hardware.

At Microsoft, we host our deep learning inferencing, cognitive science, and our applied AI services on the NC series instances. The learnings and advancements made in these areas with regard to our infrastructure are helping drive the design decisions for the next generation of NC system. Because of our approach, our Azure customers are able to benefit from our internal learnings.

We are pleased to announce that the next generation of NC A100 v4 series is now available for preview. These virtual machines (VMs) come equipped with NVIDIA A100 80GB Tensor Core PCIe GPUs and 3rd Gen AMD EPYC™ Milan processors. These new offerings improve the performance and cost-effectiveness of a variety of GPU performance-bound real-world AI training and inferencing workloads. These workloads cover object detection, video processing, image classification, speech recognition, recommender, autonomous driving reinforcement learning, oil and gas reservoir simulation, finance document parsing, web inferencing, and more.

The NC A100 v4-series offers three classes of VM ranging from one to four NVIDIA A100 80GB PCIe Tensor Core GPUs. It is more cost-effective than ever before, while still giving customers the options and flexibility they need for their workloads.

Size

vCPU

Memory (GB)

GPUs (NVIDIA A100 80 GB Tensor Core)

Azure Network (Gbps)

Standard_NC24ads_A100_v4

24

220

1

20

Standard_NC48ads_A100_v4

48

440

2

40

Standard_NC96ads_A100_v4

96

880

4

80

Compared to the previous NC generation (NCv3) with NVIDIA Volta architecture-based GPUs, customers will experience between 1.5 and 2.5 times the performance boost due to:

Two times GPU to host bandwidth.
Four times vCPU cores per GPU VM.
Two times RAM per GPU VM.
Seven independent GPU instances on a single NVIDIA A100 GPU through Multi-Instance GPU (MIG) on Linux OS.

Below is a sample of what we experienced while running ResNet50 AI model training across a variety of batch sizes using the VM size NC96ads_A100_v4 compared to the existing NCv3 4 V100 GPUs VM size NC24s_v3. Tests were conducted across a range of batch sizes, from one to 256.

Figure 1: ResNet50 results were generated using NC24s_v3 and NC96ads_A100_v4 virtual machine sizes.

For additional information on how to run this on Azure and additional results please check out our performance technical community blog.

With our latest addition the NC series, you can reduce the time it takes to train your model training in around half the time and still within budget. You can seamlessly apply the trained cognitive science models to applications through batch inferencing, run multimillion atomics biochemistry simulations for next-generation medicine, host your web and media services in the cloud for tens of thousands of end-users, and so much more.

Learn more

The NC A100 v4 series are currently available in the South Central US, East US, and Southeast Asia Azure regions. They will be available in additional regions in the coming months.
For more information on the Azure NC A100 v4-series, please see:

Sign up for the preview of the NVIDIA A100 Tensor Core PCIe GPU in the Azure NC A100 v4-series. 
Performance of NC A100 v4-series.
Find out more about high-performance computing (HPC) in Azure.
Microsoft documentation for NC A100 v4-series VM.
Azure HPC optimized OS images.
Azure GPU virtual machines.

Quelle: Azure

Optimize your cloud investment with Azure Reservations

Continuous cost optimization can take place at all stages of an Azure workload’s lifecycle, but your Azure subscription provides a very effective benefit to further optimize your investment when you are ready to deploy that workload.

For cloud workloads with consistent resource usage, you can buy reserved instances at a significant discount and reduce your workload costs by up to 72 percent compared to pay-as-you-go prices. Azure Reservations can be obtained by committing to one-year or three-year plans for virtual machines, Azure Blob storage or Azure Data Lake Storage Gen2, SQL Database compute capacity, Azure Cosmos DB throughput, and other Azure resources.

When you can predict and commit to needed capacity, it gives us visibility into your resource requirements in advance, allowing us to be more efficient in our operations. We can then pass the savings on to you. This benefit applies to both Windows and Linux virtual machines (VMs).

In addition, you now can combine the cost savings of reserved instances with the added Azure Hybrid Benefit when running on-premises and Azure workloads to save up to 80 percent over pay-as-you-go pricing.

How to get your reservation

A reservation discount only applies to resources associated with Enterprise Agreement, Microsoft Customer Agreement, Cloud Solution Provider (CSP), or subscriptions with pay-as-you-go rates. These are billing discounts (paid upfront or monthly) and do not affect the runtime state of your resources. And do not worry, you will not pay any extra fees when you choose to pay monthly.

To determine which reservation to purchase, analyze your usage data in the Azure portal, or use reservation recommendations available in Azure Advisor (VMs only), the Cost Manage Power BI app, or the Reservation Recommendations REST API.

Reservation purchase recommendations are calculated by analyzing your hourly usage data over the last seven, 30, and 60 days.

Simple and flexible

You can purchase Azure Reserved VM Instances in three easy steps—just specify your Azure region, virtual machine type, and term (one year or three years)—that's it.

Here is how it works: Discounts are generally applied to the resource usage matching the attributes you select when you buy the reservation. Attributes include the scope where the matching VMs, SQL databases, Azure Cosmos DB, or other resources run. Attributes include the SKU, regions (where applicable), and scope. Reservation scope selects where the reservation savings apply. You can scope a reservation to a subscription or resource group. When you scope the reservation to a resource group, reservation discounts apply only to the resource group—not the entire subscription.

You can manage reservations for Azure resources including updating the scope to apply reservations to a different subscription, changing who can manage the reservation, splitting a reservation into smaller parts, or changing instance size. Enhanced data for reservation costs and usage is available for Enterprise Agreement (EA) and Microsoft Customer Agreement (MCA) usage in Azure Cost Management and Billing. Those same customers can view amortized cost data for reservations and use that data to chargeback the monetary value for a subscription, resource group, or resource.

Capacity on demand

The ability for you to access compute capacity with service-level agreements, and ahead of actual VM deployments, is important to ensure the availability of mission-critical applications running on Azure. On-demand capacity reservations, now in preview, enable you to reserve compute capacity for one or more virtual machine size(s) in an Azure region or availability zone for any length of time. You can create and cancel an on-demand capacity reservation at any time, no commitment is required.

You also can exchange a reservation for another reservation of the same type or refund a reservation, up to $50,000 USD in a 12-month rolling window if you no longer need it, or cancel a reserved instance at any time and return the remaining months to Microsoft.

Learn more

Purchase reservations from the Azure portal, APIs, PowerShell, or CLI. Cloud solution providers can use the Azure portal or Partner Center to purchase Azure Reservations.

To dive deeper, check out the learning module, “Save money with Azure Reserved Instances.”
Quelle: Azure

Unlock cloud savings on the fly with autoscale on Azure

Unused cloud resources can put an unnecessary drain on your computing budget, and unlike legacy on-premises architectures, there is no need to over-provision compute resources for times of heavy usage.

Autoscaling is one of the value levers that can help unlock cost savings for your Azure workloads by automatically scaling up and down the resources in use to better align capacity to demand. This practice can greatly reduce wasted spend for those dynamic workloads with inherently “peaky” demand.

In some cases, workloads with occasionally high peak demand have extremely low average utilization, making them ill-suited for other cost optimization practices, such as rightsizing and reservations.

For periods when an app puts a heavier demand on cloud resources, autoscaling adds resources to handle the load and satisfy service-level agreements for performance and availability. And for those times when the load demand decreases (nights, weekends, holidays), autoscaling can remove idle resources to reduce costs. Autoscaling automatically scales between the minimum and maximum number of instances and will run, add, or remove VMs automatically based on a set of rules.

Autoscaling is near real-time cost optimization. Think of it this way: Rather than build an addition to your house with extra bedrooms that will go unused most of the year, you have an agreement with a nearby hotel. Your guests can check-in, at any time and at the last minute, and the hotel will automatically charge you for the days when they visit.

Not only does it utilize cloud elasticity by paying for capacity only when you need it, you can also reduce the need for an operator to continually monitor the performance of a system and make decisions about adding or removing resources.

What services can you autoscale?

Azure provides built-in autoscaling using Azure Monitor autoscale for most compute options, including:

Azure Virtual Machines Scale Sets—see How to use automatic scaling and virtual machine scale sets.
Service Fabric—see Scale a Service Fabric cluster in or out using autoscale rules.
Azure App Service—see Scale instance count manually or automatically.
Azure Cloud Services has built-in autoscaling at the role level. See How to configure autoscaling for a cloud service in the portal.

Azure Functions differs from the previous compute options because you don't need to configure any autoscale rules. The hosting plan you choose dictates how your function app is scaled:

With a consumption plan, your functions app will scale automatically, and you will only pay for compute resources when your functions are running.
With a premium plan, your app will automatically scale based on demand using pre-warmed workers that run applications with no delay after being idle.
With a dedicated plan, you will run your functions within an App Service plan at regular App Service plan rates.

Azure Monitor autoscale provides a common set of autoscaling functionality for virtual machine scale sets, Azure App Service, and Azure Cloud Service. Scaling can be performed on a schedule, or based on a runtime metric, such as CPU or memory usage.

Use the built-in autoscaling features of the platform if they meet your requirements. If not, carefully consider whether you really need more complex scaling features. Examples of additional requirements may include more granularity of control, different ways to detect trigger events for scaling, scaling across subscriptions, and scaling other types of resources.

Note that application design can impact how that app handles scale as a load increases. To review design considerations for scalable applications, including choosing the right data storage and VM size, and more, check out Design scalable Azure applications—Microsoft Azure Well-Architected Framework.

Also know that, in general, it is better to scale up than to scale down. Scaling down usually involves deprovisioning or downtime. So, choose smaller instances when a workload is highly variable and scale out to get the required level of performance.
You can set up autoscale in the Azure portal, PowerShell, Azure CLI, or Azure Monitor REST API.

Get started with autoscaling

With autoscaling, you can dynamically scale your apps to meet changing demand or anticipate loads with different schedules and set rules that trigger scaling actions. Regardless of how you set it up, the goal is to maximize the performance of your application and save money by not wasting server resources.
Quelle: Azure

AWS Amplify Studio kündigt neue Funktionen zum Speichern und Verwalten von Dateien an

AWS Amplify Studio bietet jetzt eine visuelle Benutzeroberfläche zum Einrichten und Verwalten von Dateispeicher-Ressourcen, die das Speichern und Bereitstellen von benutzergenerierten Inhalten (z. B. Fotos und Videos) aus Web- oder mobile Apps vereinfacht. Mit Amplify Studio können Sie ganz einfach einen Amazon-S3-Bucket erstellen, Dateizugriffsebenen konfigurieren, Speicher-Client-Bibliotheken in Ihre Web- oder mobilen Awendung integrieren und Dateien im Drag-and-Drop-Dateiexplorer von Studio verwalten.
Quelle: aws.amazon.com

Amazon Athena unterstützt jetzt auch die Abfrage von Amazon-Ion-Daten

Amazon Athena unterstützt jetzt Daten, die im Amazon-Ion-Format gespeichert sind, einem reich typisierten, selbstbeschreibenden Format, das von Amazon entwickelt und als Open Source bereitgestellt wurde. Amazon Ion bietet austauschbare Binär- und Textformate, die die Einfachheit der Verwendung von Text mit der Effizienz der Binärkodierung kombinieren. Das Ion-Format wird derzeit von internen Amazon-Teams, von AWS-Services wie der Amazon Quantum Ledger Database (Amazon QLDB) und in der Open-Source-SQL-Abfragesprache PartiQL verwendet.
Quelle: aws.amazon.com

AWS-Lambda-Funktion-URLs: integrierte HTTPS-Endpunkte für Ihre Lambda-Funktionen

AWS Lambda kündigt Lambda-Funktions-URLs an, eine neue Funktion, die es einfacher macht, Funktionen über einen HTTPS-Endpunkt als integrierte Fähigkeit des AWS-Lambda-Services aufzurufen. Sie können Funktions-URLs zu neuen und bestehenden Funktionen mit einem einzigen Klick von der Konsole aus hinzufügen oder sie mit wenigen Zeilen mit AWS CloudFormation oder dem AWS Serverless Application Model verwenden. Funktions-URLs sind ideal für den Einstieg in die Entwicklung von Web-Services auf Lambda oder für allgemeine Aufgaben wie die Erstellung von Web-Hooks.
Quelle: aws.amazon.com

Ankündigung von Unified Settings (Vereinheitlichte Einstellungen) in der AWS-Managementkonsole

Wir freuen uns, Unified Settings in der AWS-Managementkonsole einführen zu können. Mit Unified Settings bleiben die Einstellungen über verschieden Geräte, Browser und Services hinweg erhalten. Bei der Einführung wird Unified Settings die folgenden Einstellungen unterstützten: Standardsprache, Standardregion und Anzeige der Lieblings-Services. Die Einstellung Standardsprache zeigt Ihre bevorzugte Sprache in der Mangementkonsole an und die Standardregion legt die AWS-Region fest, die geladen wird, wenn Sie sich anmelden oder eine Service-Konsole laden. Die Anzeige der Lieblings-Services hingegen zeigt die Services in der Favoritenleiste entweder mit dem Service-Symbol und dem vollständigen Namen oder nur mit dem Service-Symbol an. Unified Settings ist in allen öffentlichen AWS-Regionen verfügbar.
Quelle: aws.amazon.com