Launch a Chat UI Agent with Docker and the Vercel AI SDK

Running a Chat UI Agent doesn’t have to involve a complicated setup. By combining Docker with the Vercel AI SDK, it’s possible to build and launch a conversational interface in a clean, reproducible way. Docker ensures that the environment is consistent across machines, while the Vercel AI SDK provides the tools for handling streaming responses and multi-turn interactions. Using Docker Compose, the entire stack can be brought online with a single command, making it easier to experiment locally or move toward production.

The Vercel AI SDK gives you a simple yet powerful framework for building conversational UIs, handling streaming responses, and managing multi-turn interactions. Pair it with Docker, and you’ve got a portable, production-ready Chat UI Agent that runs the same way on your laptop, staging, or production.

We’ll start with the Next.js AI Chatbot template from Vercel, then containerize it using a battle-tested Dockerfile from demo repo. This way, you don’t just get a demo — you get a production-ready deployment.

One command, and your Chat UI is live.

Why this setup works

Next.js 15: Modern App Router, API routes, and streaming.

Vercel AI SDK: Simple React hooks and streaming utilities for chat UIs.

Docker (standalone build): Optimized for production — lean image size, fast startup, and reliable deployments.

This stack covers both developer experience and production readiness.

Step 1: Clone the template

Start with the official Vercel chatbot template:

npx create-next-app@latest chat-ui-agent -e https://vercel.com/templates/ai/nextjs-ai-chatbot

This scaffolds a full-featured chatbot using the Vercel AI SDK.

Step 2: Configure API keys

Create a .env.local file in the root:

OPENAI_API_KEY=your_openai_key_here

Swap in your provider key if you’re using Anthropic or another backend.

Step 3: Add the production Dockerfile

Instead of writing your own Dockerfile, grab the optimized version from Kristiyan Velkov’s repo:

Next.js Standalone Dockerfile

Save it as Dockerfile in your project root.

This file:

Uses multi-stage builds.

Creates a standalone Next.js build.

Keeps the image lightweight and fast for production.

Step 4: Docker Compose Setup

Here’s a simple docker-compose.yml:

services:
chat-ui-agent:
build:
context: .
dockerfile: Dockerfile
ports:
– "3000:3000"
environment:
OPENAI_API_KEY: ${OPENAI_API_KEY}

This ensures your API key is passed securely into the container.

Step 5: Build and Run

Spin up your chatbot:

docker-compose up –build

Open http://localhost:3000, and your Chat UI Agent is ready to roll.

Why the standalone Dockerfile matters

Using the standalone Next.js Dockerfile instead of a basic one gives you real advantages:

Production-grade: Optimized builds, smaller image sizes, faster deploys.

Best practices baked in: No need to reinvent Docker configs.

Portable: Same setup runs on local dev, staging, or production servers.

This is the kind of Dockerfile you’d actually ship to production, not just test locally.

Final Thoughts

With the Next.js AI Chatbot template, the Vercel AI SDK, and a production-ready Dockerfile, spinning up a Chat UI Agent is not just quick — it’s deployment-ready from day one.

If you want to move fast without cutting corners, this setup strikes the perfect balance: modern frameworks, clean developer experience, and a solid production pipeline.

Quelle: https://blog.docker.com/feed/

Docker + Unsloth: Build Custom Models, Faster

Building and Running Custom Models Is Still Hard

Running AI models locally is still hard. Even as open-source LLMs grow more capable, actually getting them to run on your machine, with the right dependencies, remains slow, fragile, and inconsistent.

There’s two sides to this challenge:

Model creation and optimization: making fine-tuning and quantization efficient.

Model execution and portability: making models reproducible, isolated, and universal.

Solving both lets developers actually use the models they build.

Docker + Unsloth: Making Iterating on Custom Models Faster

A lot of developers want to move away from consume the API to own the model. They want to fine-tune models for their own use cases but doing it remains hard.

We’re excited to be working together with Unsloth to make building, iterating, and running custom LLMs locally faster, simpler, and more accessible for every developer.

Unsloth lowers the barrier to building (and exporting) fine-tuned custom models. Docker lowers the barrier to running them anywhere.

You can now run any model, including Unsloth Dynamic GGUFs, on Mac, Windows or Linux with Docker Model Runner. Together, friction’s removed between experimentation and execution: dependency and reproducibility gaps.

With Docker Model Runner (DMR), starting a model is as simple as docker model run. For example, running OpenAI’s open-weight model locally becomes incredibly easy:

docker model run ai/gpt-oss:20B

How It Works

Fine-tune with Unsloth. Train and optimize your model efficiently.

Export to GGUF. Quantize to a lightweight, portable format for fast local inference.

Run with Docker. Launch instantly with docker model run. No manual setup.

Unsloth’s Dynamic GGUFs help you create compact fine-tuned models. Docker Model Runner lets you spin them up instantly and run them as easily as containers, without worrying about dependency issues.

What’s Next

Building and running AI should feel as natural as developing and shipping code. Just like containers standardized application deployment, we’re now doing the same for AI.

Unsloth + Docker marks one more step in that journey. Learn more in the docs. 

Quelle: https://blog.docker.com/feed/

Amazon Redshift announces support for the SUPER data type in Databases with Case-Insensitive Collation

Amazon Redshift announces support for the SUPER data type in databases with case insensitive collation, enabling analytics on semi-structured and nested data in these databases. Using the SUPER data type with PartiQL in Amazon Redshift, you can perform advanced analytics that combine structured SQL data (such as string, numeric, and timestamp) with the semi-structured SUPER data (such as JSON) with flexibility and ease-of-use. This enhancement allows you to leverage the SUPER data type for your structured and semi-structured data processing needs in databases with case-insensitive collation. Using the COLLATE function, you can now explicitly specify case sensitivity preferences for SUPER columns, providing greater flexibility in handling data with varying case patterns. This is particularly valuable when working with JSON documents, APIs, or application data where case consistency isn’t guaranteed. Whether you’re processing user-defined identifiers or integrating data from multiple sources, you can now perform complex queries across both case-sensitive and case-insensitive data without additional normalization overhead. Amazon Redshift support for the SUPER data type in databases with case insensitive collation is available in all AWS Regions, including the AWS GovCloud (US) Regions, where Amazon Redshift is available. See AWS Region Table for more details. To learn more about the SUPER data type in databases with case insensitive collation, please visit our documentation.
Quelle: aws.amazon.com

Amazon EC2 I7i instances now available in additional AWS regions

Amazon Web Services (AWS) announces the availability of high performance Storage Optimized Amazon EC2 I7i instances in AWS Asia Pacific (Melbourne, Mumbai, Osaka), Middle East (UAE) regions. Powered by 5th generation Intel Xeon Scalable processors with an all-core turbo frequency of 3.2 GHz, these instances deliver up to 23% better compute performance and more than 10% better price performance over previous generation I4i instances. Powered by 3rd generation AWS Nitro SSDs, I7i instances offer up to 45TB of NVMe storage with up to 50% better real-time storage performance, up to 50% lower storage I/O latency, and up to 60% lower storage I/O latency variability compared to I4i instances. I7i instances are ideal for I/O intensive and latency-sensitive workloads that demand very high random IOPS performance with real-time latency to access small to medium size datasets (multi-TBs). I7i instances support torn write prevention feature with up to 16KB block sizes, enabling customers to eliminate database performance bottlenecks. I7i instances are available in eleven sizes – nine virtual sizes up to 48xlarge and two bare metal sizes – delivering up to 100Gbps of network bandwidth and 60Gbps of Amazon Elastic Block Store (EBS) bandwidth. To learn more, visit the I7i instances page.
Quelle: aws.amazon.com

Amazon Polly expands Generative TTS engine with additional languages and region support

Today, we are excited to announce the general availability of five highly expressive Amazon Polly Generative voices in Austrian German (Hannah), Irish English (Niamh), Brazilian Portuguese (Camila), Belgian Dutch (Lisa), and Korean (Seoyeon). This release follows our October launch of Netherlands Dutch (Laura) Generative voice, bringing our total Generative engine offering to thirty-one voices across twenty locales. Additionally, we have expanded the Generative engine to three new regions in Asia Pacific: Asia Pacific (Seoul), Asia Pacific (Singapore), and Asia Pacific (Tokyo). Amazon Polly is a fully-managed service that turns text into lifelike speech, allowing developers and builders to enable their applications for conversational AI or for speech content creation. All new and existing Generative voices are now available in the US East (North Virginia), Europe (Frankfurt), US West (Oregon), Asia Pacific (Seoul), Asia Pacific (Singapore), and Asia Pacific (Tokyo) regions. To hear how Polly voices sound, go to Amazon Polly Features. To learn more about how to use Generative engine, go to AWS Blog. For more details on the Polly offerings and use, please read the Amazon Polly documentation and visit our pricing page.
Quelle: aws.amazon.com