Langsame Datenübertragung: iPhone 17 und iPhone Air kommen mit USB 2.0
USB 2.0 ermöglicht eine Datenübertragung von nur 480 Mbit/s – Apple stattet sein neues dünnes Smartphone iPhone Air damit aus. (iPhone 17, Apple)
Quelle: Golem
USB 2.0 ermöglicht eine Datenübertragung von nur 480 Mbit/s – Apple stattet sein neues dünnes Smartphone iPhone Air damit aus. (iPhone 17, Apple)
Quelle: Golem
Einer der intensivsten Romane von Stephen King ist ein gleichermaßen intensiver Film geworden. Ab dem 11. September beginnt der lange Marsch im Kino. Eine Rezension von Peter Osteried (Science-Fiction, Fernsehen)
Quelle: Golem
Eine einfache Phishing-Mail hat unzählige Malware-Infektionen zur Folge. Auch Cloud-Systeme sind betroffen. Dennoch geht der Angreifer fast leer aus. (NPM, Server)
Quelle: Golem
How developers are embedding runtime security to safely build with AI agents
Introduction: When AI Workflows Become Attack Surfaces
The AI tools we use today are powerful, but also unpredictable and exploitable.
You prompt an LLM and it generates a Dockerfile. It looks correct. A shell script? Reasonable. You run it in dev. Then something breaks: a volume is deleted. A credential leaks into a log. An outbound request hits a production API. Nothing in your CI pipeline flagged it, because the risk only became real at runtime.
This is the new reality of AI-native development: fast-moving code, uncertain behavior, and an expanding attack surface.
Hallucinations in LLM output are only part of the story. As developers build increasingly autonomous agentic tools, they’re also exposed to prompt injection, jailbreaks, and deliberate misuse of model outputs by adversaries. A malicious user, through a cleverly crafted input, can hijack an AI agent and cause it to modify files, exfiltrate secrets, or run unauthorized commands.
In one recent case, a developer ran an LLM-generated script that silently deleted a production database, an issue that went undetected until customer data was already lost. In another, an internal AI assistant was prompted to upload sensitive internal documents to an external file-sharing site, triggered entirely through user input.
These failures weren’t caught in static analysis, code review, or CI. They surfaced only when the code ran.
In this post, we’ll explore how developers are addressing both accidental failures and intentional threats by shifting runtime security into the development loop, embedding observability, policy enforcement, and threat detection directly into their workflows using Docker.
The Hidden Risks of AI-Generated Code
LLMs and AI agents are great at generating text, but they don’t always know what they’re doing. Whether you’re using GitHub Copilot, LangChain, or building with OpenAI APIs, your generated outputs might include:
Shell scripts that escalate privileges or misconfigure file systems
Dockerfiles that expose unnecessary ports or install outdated packages
Infra-as-code templates that connect to production services by default
Hardcoded credentials or tokens hidden deep in the output
Command sequences that behave differently depending on the context
The problem is compounded when teams start running autonomous agents, AI tools designed to take actions, not just suggest code. These agents can:
Execute file writes and deletions
Make outbound API calls
Spin up or destroy containers
Alter configuration state mid-execution
Execute dangerous database queries
These risks only surface at runtime, after your build has passed and your pipeline has shipped. And that’s a problem developers are increasingly solving inside the dev loop.
Why Runtime Security Belongs in the Developer Workflow
Traditional security tooling focuses on build-time checks, SAST, SCA, linters, compliance scanners. These are essential, but they don’t protect you from what AI-generated agents do at execution time.
Developers need runtime security that fits their workflow, not a blocker added later.
What runtime security enables:
Live detection of dangerous system calls or file access
Policy enforcement when an agent attempts unauthorized actions
Observability into AI-generated code behavior in real environments
Isolation of high-risk executions in containerized sandboxes
Why it matters:
Faster feedback loops: See issues before your CI/CD fails
Reduced incident risk: Catch privilege escalation, data exposure, or network calls early
Higher confidence: Ship LLM-generated code without guesswork
Secure experimentation: Enable safe iteration without slowing down teams
Developer ROI: Catching a misconfigured agent in dev avoids hours of triage and mitigates production risk and reputation risk; saving time, cost, and compliance exposure.
Building Safer AI Workflows with Docker
Docker provides the building blocks to develop, test, and secure modern agentic applications:
Docker Desktop gives you an isolated, local runtime for testing unsafe code
Docker Hardened Images. Secure, minimal, production-ready images
Docker Scout scans container images for vulnerabilities and misconfigurations
Runtime policy enforcement (with upcoming MCP Defender integration) provides live detection and guardrails while code executes
Step-by-Step: Safely Test AI-Generated Scripts
1. Run your agent or script in a hardened container
docker run –rm -it
–security-opt seccomp=default.json
–cap-drop=ALL
-v $(pwd):/workspace
python:3.11-slim
Applies syscall restrictions and drops unnecessary capabilities
Runs with no persistent volume changes
Enables safe, repeatable testing of LLM output
2. Scan the container with Docker Scout
docker scout cves my-agent:latest
Surfaces known CVEs and outdated dependencies
Detects unsafe base images or misconfigured package installs
Available both locally and inside CI/CD workflows
3. Add runtime policy (beta) to block unsafe behavior
scout policy add deny-external-network
–rule "deny outbound to *"
This would catch an AI agent that unknowingly makes an outbound request to an internal system, third-party API, or external data store.
Note: Runtime policy enforcement in Docker Scout is currently in development. CLI and behavior may change upon release.
Best Practices for Securing AI Agent Containers
Practice
Why it matters
Use slim, verified base images
Minimizes attack surface and dependency drift
Avoid downloading from unverified sources
Prevents LLMs from introducing shadow dependencies
Use .dockerignore and secrets management
Keeps secrets out of containers
Run containers with dropped capabilities
Limits impact of unexpected commands
Apply runtime seccomp profiles
Enforces syscall-level sandboxing
Log agent behavior for analysis
Builds observability into experimentation
Integrating Into Your Cloud-Native Workflow
Runtime security for AI tools isn’t just for local testing, it fits cleanly into cloud-native and CI/CD workflows too.
GitHub Actions Integration Example:
jobs:
security-scan:
runs-on: ubuntu-latest
steps:
– uses: actions/checkout@v3
– name: Build container
run: docker build -t my-agent:latest .
– name: Scan for CVEs
run: docker scout cves my-agent:latest
Works across environments:
Local dev via Docker Desktop
Remote CI/CD via GitHub Actions, GitLab, Jenkins
Kubernetes staging environments with policy enforcement and agent isolation
Cloud Development Environments (CDEs) with Docker + secure agent sandboxes
Dev teams using ephemeral workspaces and Docker containers in cloud IDEs or CDEs can now enforce the same policies across local and cloud environments.
Real-World Example: AI-Generated Infra Gone Wrong
A platform team uses an LLM agent to auto-generate Kubernetes deployment templates. A developer reviews the YAML and merges it. The agent-generated config opens an internal-only service to the internet via LoadBalancer. The CI pipeline passes. The deploy works. But a customer database is now exposed.
Had the developer run this template inside a containerized sandbox with outbound policy rules, the attempt to expose the service would have triggered an alert, and the policy would have prevented escalation.
Lesson: You can’t rely on static review alone. You need to see what AI-generated code does, not just what it looks like.
Why This Matters: Secure-by-Default for AI-Native Dev Teams
As LLM-powered tools evolve from suggestion to action, runtime safety becomes a baseline requirement, not an optional add-on.
The future of secure AI development starts in the inner loop, with runtime policies, observability, and smart defaults that don’t slow you down.
Docker’s platform gives you:
Developer-first workflows with built-in security
Runtime enforcement to catch AI mistakes early
Toolchain integration across build, test, deploy
Cloud-native flexibility across local dev, CI/CD, and CDEs
Whether you’re building AI-powered automations, agent-based platforms, or tools that generate infrastructure, you need a runtime layer that sees what AI can’t, and blocks what it shouldn’t do.
What’s Next
Runtime protection is moving left, into your dev environment. With Docker, developers can:
Run LLM-generated code in secure, ephemeral containers
Observe runtime behavior before pushing to CI
Enforce policies that prevent high-risk actions
Reduce the risk of silent security failures in AI-powered apps
Docker is working to bring MCP Defender into our platform to provide this protection out-of-the-box, so hallucinations don’t turn into incidents.
Ready to Secure Your AI Workflow?
Sign up for early access to Docker’s runtime security capabilities
Watch our Tech Talk on “Building Safe AI Agents with Docker”
Explore Docker Scout for real-time vulnerability insights
Join the community conversation on Docker Community Slack or GitHub Discussions
Let’s build fast, and safely.
Quelle: https://blog.docker.com/feed/
Die Jackery Explorer 240 v2 Powerstation ist perfekt für Camping und Notfälle. Bei Amazon ist sie kurzzeitig im Angebot. (Powerstation, Technik/Hardware)
Quelle: Golem
Der SAP-Betriebsrat sieht Vertrauensschwund bei Kunden und Beschäftigten. Ständige Entlassungen würden SAP schaden. (Arbeit, SAP)
Quelle: Golem
Ein Walking Pad der Marke Merach hat bei Amazon einen neuen Tiefstpreis erreicht. Dank Doppelrabatt spart man 77 Euro. (Technik/Hardware, Amazon)
Quelle: Golem
Amazons derzeit meistgekaufte LED-Taschenlampe mit 5.000 Verkäufen im letzten Monat ist befristet im Angebot. (LED, Technik/Hardware)
Quelle: Golem
Schon im November 2024 hat die Ampel-Regierung den Entwurf für ein Kritis-Dachgesetz beschlossen. Nun ging dieser praktisch unverändert durchs Kabinett. (Infrastruktur, Deutsche Bahn)
Quelle: Golem
This blog post is the fourth out of a six-part blog series called Agent Factory which will share best practices, design patterns, and tools to help guide you through adopting and building agentic AI.
Developer experiences as the key to scale
AI agents are moving quickly from experimentation to real production systems. Across industries, we see developers testing prototypes in their Integrated Development Environment (IDE) one week and deploying production agents to serve thousands of users the next. The key differentiator is no longer whether you can build an agent—it’s how fast and seamlessly you can go from idea to enterprise-ready deployment.
Deploy AI agents quickly with Azure AI Foundry
Industry trends reinforce this shift:
In-repo AI development: Models, prompts, and evaluations are now first-class citizens in GitHub repos—giving developers a unified space to build, test, and iterate on AI features.
More capable coding agents: GitHub Copilot’s new coding agent can open pull requests after completing tasks like writing tests or fixing bugs, acting as an asynchronous teammate.
Open frameworks maturing: Communities around LangGraph, LlamaIndex, CrewAI, AutoGen, and Semantic Kernel are rapidly expanding, with “agent templates” on GitHub repos becoming common.
Open protocols emerging: Standards like the Model Context Protocol (MCP) and Agent-to-Agent (A2A) are creating interoperability across platforms.
Developers increasingly expect to stay in their existing workflow—GitHub, VS Code, and familiar frameworks—while tapping into enterprise-grade runtimes and integrations. The platforms that win will be those that meet developers where they are—with openness, speed, and trust.
What a modern agent platform should deliver
From our work with customers and the open-source community, we’ve seen a clear picture emerge of what developers really need. A modern agent platform must go beyond offering models or orchestration—it has to empower teams across the entire lifecycle:
Local-first prototyping: Developers want to stay in their flow. That means designing, tracing, and evaluating AI agents directly in their IDE with the same ease as writing and debugging code. If building an agent requires jumping into a separate UI or unfamiliar environment, iteration slows and adoption drops.
Frictionless transition to production: A common frustration we hear is that an agent that runs fine locally becomes brittle or requires heavy rewrites in production. The right platform provides a single, consistent API surface from experimentation to deployment, so what works in development works in production—with scale, security, and governance layered in automatically.
Open by design: No two organizations use the exact same stack. Developers may start with LangGraph for orchestration, LlamaIndex for data retrieval, or CrewAI for coordination. Others prefer Microsoft’s first-party frameworks like Semantic Kernel or AutoGen. A modern platform must support this diversity without forcing lock-in, while still offering enterprise-grade pathways for those who want them.
Interop by design: Agents are rarely self-contained. They must talk to tools, databases, and even other agents across different ecosystems. Proprietary protocols create silos and fragmentation. Open standards like the Model Context Protocol (MCP) and Agent-to-Agent (A2A) unlock collaboration across platforms, enabling a marketplace of interoperable tools and reusable agent skills.
One-stop integration fabric: An agent’s real value comes when it can take meaningful action: updating a record in Dynamics 365, triggering a workflow in ServiceNow, querying a SQL database, or posting to Teams. Developers shouldn’t have to rebuild connectors for every integration. A robust agent platform provides a broad library of prebuilt connectors and simple ways to plug into enterprise systems.
Built-in guardrails: Enterprises cannot afford agents that are opaque, unreliable, or non-compliant. Observability, evaluations, and governance must be woven into the development loop—not added as an afterthought. The ability to trace agent reasoning, run continuous evaluations, and enforce identity, security, and compliance policies is as critical as the models themselves.
How Azure AI Foundry delivers this experience
Azure AI Foundry is designed to meet developers where they are, while giving enterprises the trust, security, and scale they need. It connects the dots across IDEs, frameworks, protocols, and business channels—making the path from prototype to production seamless.
Build where developers live: VS Code, GitHub, and Foundry
Developers expect to design, debug, and iterate AI agents in their daily tools—not switch into unfamiliar environments. Foundry integrates deeply with both VS Code and GitHub to support this flow.
VS Code extension for Foundry: Developers can create, run, and debug agents locally with direct connection to Foundry resources. The extension scaffolds projects, provides integrated tracing and evaluation, and enables one-click deployment to Foundry Agent Service—all inside the IDE they already use.
Model Inference API: With a single, unified inference endpoint, developers can evaluate performance across models and swap them without rewriting code. This flexibility accelerates experimentation while future-proofing applications against a fast-moving model ecosystem.
GitHub Copilot and the coding agent: Copilot has grown beyond autocomplete into an autonomous coding agent that can take on issues, spin up a secure runner, and generate a pull request, signaling how agentic AI development is becoming a normal part of the developer loop. When used alongside Azure AI Foundry, developers can accelerate agent development by having Copilot generate agent code while pulling in the models, agent runtime, and observability tools from Foundry needed to build, deploy, and monitor production-ready agents.
Use your frameworks
Agents are not one-size-fits-all, and developers often start with the frameworks they know best. Foundry embraces this diversity:
First-party frameworks: Foundry supports both Semantic Kernel and AutoGen, with a convergence into a modern unified framework coming soon. This future-ready framework is designed for modularity, enterprise-grade reliability, and seamless deployment to Foundry Agent Service.
Third-party frameworks: Foundry Agent Service integrates directly with CrewAI, LangGraph, and LlamaIndex, enabling developers to orchestrate multi-turn, multi-agent conversations across platforms. This ensures you can work with your preferred OSS ecosystem while still benefiting from Foundry’s enterprise runtime.
Interoperability with open protocols
Agents don’t live in isolation—they need to interoperate with tools, systems, and even other agents. Foundry supports open protocols by default:
MCP: Foundry Agent Service allows agents to call any MCP-compatible tools directly, giving developers a simple way to connect external systems and reuse tools across platforms.
A2A: Semantic Kernel supports A2A, implementing the protocol to enable agents to collaborate across different runtimes and ecosystems. With A2A, multi-agent workflows can span vendors and frameworks, unlocking scenarios like specialist agents coordinating to solve complex problems.
Ship where the business runs
Building an agent is just the first step—impact comes when users can access it where they work. Foundry makes it easy to publish agents to both Microsoft and custom channels:
Microsoft 365 and Copilot: Using the Microsoft 365 Agents SDK, developers can publish Foundry agents directly to Teams, Microsoft 365 Copilot, BizChat, and other productivity surfaces.
Custom apps and APIs: Agents can be exposed as REST APIs, embedded into web apps, or integrated into workflows using Logic Apps and Azure Functions—with thousands of prebuilt connectors to SaaS and enterprise systems.
Observe and harden
Reliability and safety can’t be bolted on later—they must be integrated into the development loop. As we explored in the previous blog, observability is essential for delivering AI that is not only effective, but also trustworthy. Foundry builds these capabilities directly into the developer workflow:
Tracing and evaluation tools to debug, compare, and validate agent behavior before and after deployment.
CI/CD integration with GitHub Actions and Azure DevOps, enabling continuous evaluation and governance checks on every commit.
Enterprise guardrails—from networking and identity to compliance and governance—so that prototypes can scale confidently into production.
Why this matters now
Developer experience is the new productivity moat. Enterprises need to enable their teams to build and deploy AI agents quickly, confidently, and at scale. Azure AI Foundry delivers an open, modular, and enterprise-ready path—meeting developers in GitHub and VS Code, supporting both open-source and first-party frameworks, and ensuring agents can be deployed where users and data already live.
With Foundry, the path from prototype to production is smoother, faster, and more secure—helping organizations innovate at the speed of AI.
What’s next
In Part 5 of the Agent Factory series, we’ll explore how agents connect and collaborate at scale. We’ll demystify the integration landscape—from agent-to-agent collaboration with A2A, to tool interoperability with MCP, to the role of open standards in ensuring agents can work across apps, frameworks, and ecosystems. Expect practical guidance and reference patterns for building truly connected agent systems.
Did you miss these posts in the series?
Agent Factory: The new era of agentic AI—common use cases and design patterns.
Agent Factory: Building your first AI agent with the tools to deliver real-world outcomes.
Agent Factory: Top 5 agent observability best practices for reliable AI.
Azure AI Foundry
Build AI agents that automate tasks, enhance user experiences, and deliver results.
Learn more
The post Agent Factory: From prototype to production—developer tools and rapid agent development appeared first on Microsoft Azure Blog.
Quelle: Azure