Amazon EC2 P6-B300 instances with NVIDIA Blackwell Ultra GPUs are now available

Today, AWS announces the general availability of Amazon Elastic Compute Cloud (Amazon EC2) P6-B300 instances, accelerated by NVIDIA Blackwell Ultra B300 GPUs. Amazon EC2 P6-B300 instances provide 8x NVIDIA Blackwell Ultra GPUs with 2.1 TB high bandwidth GPU memory, 6.4 Tbps EFA networking, 300 Gbps dedicated ENA throughput, and 4 TB of system memory. 
P6-B300 instances deliver 2x networking bandwidth, 1.5x GPU memory size, and 1.5x GPU TFLOPS (at FP4, without sparsity) compared to P6-B200 instances, making them well suited to train and deploy large trillion-parameter foundation models (FMs) and large language models (LLMs) with sophisticated techniques. The higher networking and larger memory deliver faster training times and more token throughput for AI workloads. 
P6-B300 instances are now available in the p6-b300.48xlarge size through Amazon EC2 Capacity Blocks for ML and Savings Plans in the following AWS Region: US West (Oregon). For on-demand reservation of P6-B300 instances, please reach out to your account manager.
To learn more about P6-B300 instances, visit Amazon EC2 P6 instances.
Quelle: aws.amazon.com

Launch a Chat UI Agent with Docker and the Vercel AI SDK

Running a Chat UI Agent doesn’t have to involve a complicated setup. By combining Docker with the Vercel AI SDK, it’s possible to build and launch a conversational interface in a clean, reproducible way. Docker ensures that the environment is consistent across machines, while the Vercel AI SDK provides the tools for handling streaming responses and multi-turn interactions. Using Docker Compose, the entire stack can be brought online with a single command, making it easier to experiment locally or move toward production.

The Vercel AI SDK gives you a simple yet powerful framework for building conversational UIs, handling streaming responses, and managing multi-turn interactions. Pair it with Docker, and you’ve got a portable, production-ready Chat UI Agent that runs the same way on your laptop, staging, or production.

We’ll start with the Next.js AI Chatbot template from Vercel, then containerize it using a battle-tested Dockerfile from demo repo. This way, you don’t just get a demo — you get a production-ready deployment.

One command, and your Chat UI is live.

Why this setup works

Next.js 15: Modern App Router, API routes, and streaming.

Vercel AI SDK: Simple React hooks and streaming utilities for chat UIs.

Docker (standalone build): Optimized for production — lean image size, fast startup, and reliable deployments.

This stack covers both developer experience and production readiness.

Step 1: Clone the template

Start with the official Vercel chatbot template:

npx create-next-app@latest chat-ui-agent -e https://vercel.com/templates/ai/nextjs-ai-chatbot

This scaffolds a full-featured chatbot using the Vercel AI SDK.

Step 2: Configure API keys

Create a .env.local file in the root:

OPENAI_API_KEY=your_openai_key_here

Swap in your provider key if you’re using Anthropic or another backend.

Step 3: Add the production Dockerfile

Instead of writing your own Dockerfile, grab the optimized version from Kristiyan Velkov’s repo:

Next.js Standalone Dockerfile

Save it as Dockerfile in your project root.

This file:

Uses multi-stage builds.

Creates a standalone Next.js build.

Keeps the image lightweight and fast for production.

Step 4: Docker Compose Setup

Here’s a simple docker-compose.yml:

services:
chat-ui-agent:
build:
context: .
dockerfile: Dockerfile
ports:
– "3000:3000"
environment:
OPENAI_API_KEY: ${OPENAI_API_KEY}

This ensures your API key is passed securely into the container.

Step 5: Build and Run

Spin up your chatbot:

docker-compose up –build

Open http://localhost:3000, and your Chat UI Agent is ready to roll.

Why the standalone Dockerfile matters

Using the standalone Next.js Dockerfile instead of a basic one gives you real advantages:

Production-grade: Optimized builds, smaller image sizes, faster deploys.

Best practices baked in: No need to reinvent Docker configs.

Portable: Same setup runs on local dev, staging, or production servers.

This is the kind of Dockerfile you’d actually ship to production, not just test locally.

Final Thoughts

With the Next.js AI Chatbot template, the Vercel AI SDK, and a production-ready Dockerfile, spinning up a Chat UI Agent is not just quick — it’s deployment-ready from day one.

If you want to move fast without cutting corners, this setup strikes the perfect balance: modern frameworks, clean developer experience, and a solid production pipeline.

Quelle: https://blog.docker.com/feed/

Docker + Unsloth: Build Custom Models, Faster

Building and Running Custom Models Is Still Hard

Running AI models locally is still hard. Even as open-source LLMs grow more capable, actually getting them to run on your machine, with the right dependencies, remains slow, fragile, and inconsistent.

There’s two sides to this challenge:

Model creation and optimization: making fine-tuning and quantization efficient.

Model execution and portability: making models reproducible, isolated, and universal.

Solving both lets developers actually use the models they build.

Docker + Unsloth: Making Iterating on Custom Models Faster

A lot of developers want to move away from consume the API to own the model. They want to fine-tune models for their own use cases but doing it remains hard.

We’re excited to be working together with Unsloth to make building, iterating, and running custom LLMs locally faster, simpler, and more accessible for every developer.

Unsloth lowers the barrier to building (and exporting) fine-tuned custom models. Docker lowers the barrier to running them anywhere.

You can now run any model, including Unsloth Dynamic GGUFs, on Mac, Windows or Linux with Docker Model Runner. Together, friction’s removed between experimentation and execution: dependency and reproducibility gaps.

With Docker Model Runner (DMR), starting a model is as simple as docker model run. For example, running OpenAI’s open-weight model locally becomes incredibly easy:

docker model run ai/gpt-oss:20B

How It Works

Fine-tune with Unsloth. Train and optimize your model efficiently.

Export to GGUF. Quantize to a lightweight, portable format for fast local inference.

Run with Docker. Launch instantly with docker model run. No manual setup.

Unsloth’s Dynamic GGUFs help you create compact fine-tuned models. Docker Model Runner lets you spin them up instantly and run them as easily as containers, without worrying about dependency issues.

What’s Next

Building and running AI should feel as natural as developing and shipping code. Just like containers standardized application deployment, we’re now doing the same for AI.

Unsloth + Docker marks one more step in that journey. Learn more in the docs. 

Quelle: https://blog.docker.com/feed/

Amazon Redshift announces support for the SUPER data type in Databases with Case-Insensitive Collation

Amazon Redshift announces support for the SUPER data type in databases with case insensitive collation, enabling analytics on semi-structured and nested data in these databases. Using the SUPER data type with PartiQL in Amazon Redshift, you can perform advanced analytics that combine structured SQL data (such as string, numeric, and timestamp) with the semi-structured SUPER data (such as JSON) with flexibility and ease-of-use. This enhancement allows you to leverage the SUPER data type for your structured and semi-structured data processing needs in databases with case-insensitive collation. Using the COLLATE function, you can now explicitly specify case sensitivity preferences for SUPER columns, providing greater flexibility in handling data with varying case patterns. This is particularly valuable when working with JSON documents, APIs, or application data where case consistency isn’t guaranteed. Whether you’re processing user-defined identifiers or integrating data from multiple sources, you can now perform complex queries across both case-sensitive and case-insensitive data without additional normalization overhead. Amazon Redshift support for the SUPER data type in databases with case insensitive collation is available in all AWS Regions, including the AWS GovCloud (US) Regions, where Amazon Redshift is available. See AWS Region Table for more details. To learn more about the SUPER data type in databases with case insensitive collation, please visit our documentation.
Quelle: aws.amazon.com

Amazon EC2 I7i instances now available in additional AWS regions

Amazon Web Services (AWS) announces the availability of high performance Storage Optimized Amazon EC2 I7i instances in AWS Asia Pacific (Melbourne, Mumbai, Osaka), Middle East (UAE) regions. Powered by 5th generation Intel Xeon Scalable processors with an all-core turbo frequency of 3.2 GHz, these instances deliver up to 23% better compute performance and more than 10% better price performance over previous generation I4i instances. Powered by 3rd generation AWS Nitro SSDs, I7i instances offer up to 45TB of NVMe storage with up to 50% better real-time storage performance, up to 50% lower storage I/O latency, and up to 60% lower storage I/O latency variability compared to I4i instances. I7i instances are ideal for I/O intensive and latency-sensitive workloads that demand very high random IOPS performance with real-time latency to access small to medium size datasets (multi-TBs). I7i instances support torn write prevention feature with up to 16KB block sizes, enabling customers to eliminate database performance bottlenecks. I7i instances are available in eleven sizes – nine virtual sizes up to 48xlarge and two bare metal sizes – delivering up to 100Gbps of network bandwidth and 60Gbps of Amazon Elastic Block Store (EBS) bandwidth. To learn more, visit the I7i instances page.
Quelle: aws.amazon.com