AWS Builder ID now supports Sign in with GitHub and Amazon

AWS Builder ID, your profile for accessing AWS applications including AWS Builder Center, AWS Training and Certification and Kiro, now supports two new social logins: GitHub and Amazon. This expansion of sign-in options builds on the existing Google Apple social sign-in capabilities, providing GitHub and Amazon users with a streamlined way to access AWS resources without managing separate credentials on AWS.
With Sign in with Github and Amazon integration, developers and builders can now enjoy access to their AWS Builder ID profile using their GitHub or Amazon Account credentials. This enhancement eliminates password management complexity, reduces forgotten password issues, and provides a frictionless experience for both new user registration and returning user sign-ins. Whether you’re accessing development resources in AWS Builder Center, enrolling in certification programs or using Kiro to code your next app, your GitHub and Amazon Accounts can now serve as a secure gateway to your builder AWS journey.
To learn more about AWS Builder ID and get started with Sign in with GitHub and Amazon, visit the AWS Builder ID documentation.
Quelle: aws.amazon.com

Amazon Bedrock now supports observability of First Token Latency and Quota Consumption

Amazon Bedrock is a fully managed service for building generative AI applications using high-performing foundation models from leading AI providers. It now supports two new CloudWatch metrics, TimeToFirstToken and EstimatedTPMQuotaUsage, giving you deeper visibility into inference performance and quota consumption.
TimeToFirstToken measures the latency from when a request is sent to when the first token is received, for streaming APIs (ConverseStream and InvokeModelWithResponseStream). You can use this metric to set CloudWatch alarms which monitor latency degradation and establish SLA baselines, without any client-side instrumentation. EstimatedTPMQuotaUsage tracks your estimated Tokens Per Minute (TPM) quota consumption, including cache write tokens and output burndown multipliers, across all inference APIs (Converse, InvokeModel, ConverseStream, and InvokeModelWithResponseStream). You can use this metric to set proactive alarms before reaching your quota limit, track your quota consumption across your models, and request further quota increases before usage is rate limited.
Both metrics are supported in all commercial Bedrock regions for models available via cross-region inference profiles and in-region inference, updated every minute for successfully completed requests. These are available in your CloudWatch out of the box; you pay only for the underlying model inference you consume, with no API changes or opt-in required.
To learn more about TimeToFirstToken and EstimatedTPMQuotaUsage, see our documentation page on Monitoring Amazon Bedrock.
Quelle: aws.amazon.com

Amazon Bedrock AgentCore Runtime now supports stateful MCP server features

Amazon Bedrock AgentCore Runtime now supports stateful Model Context Protocol (MCP) server features, enabling developers to build MCP servers that leverage elicitation, sampling, and progress notifications alongside existing support for resources, prompts, and tools. These capabilities allow MCP servers deployed to AgentCore Runtime to collect user input interactively during tool execution, request LLM-generated content from clients, and provide real-time progress updates for long-running operations. With stateful MCP sessions, each user session runs in a dedicated microVM with isolated resources, and the server maintains session context across multiple interactions using an Mcp-Session-Id header. Elicitation enables server-initiated, multi-turn conversations to gather information such as user preferences. Sampling allows servers to request AI-powered text generation from the client for tasks like personalized recommendations. Progress notifications keep clients informed during operations such as searching for flights or processing bookings. These features work together to support complex, interactive agent workflows that go beyond simple request-response patterns.
Stateful MCP server features are supported in AgentCore Runtime across fourteen AWS Regions: US East (N. Virginia), US East (Ohio), US West (Oregon), Asia Pacific (Mumbai), Canada (Central), Asia Pacific (Seoul), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Europe (Frankfurt), Europe (Ireland), Europe (London), Europe (Paris), and Europe (Stockholm).
To learn more, see Stateful MCP server features in the Amazon Bedrock AgentCore documentation.
Quelle: aws.amazon.com

Introducing GPT-5.4 in Microsoft Foundry

Built for Reliable AI Production: Stronger reasoning, dependable execution, and agentic workflows at scaleToday, we’re announcing OpenAI’s GPT‑5.4 to be generally available soon in Microsoft Foundry: a model designed to help organizations move from planning work to reliably completing it in production environments. As AI agents are applied to longer, more complex workflows; consistency and follow‑through become as important as raw intelligence. GPT‑5.4 combines stronger reasoning with built in computer use capabilities to support automation scenarios, and dependable execution across tools, files, and multi‑step workflows at scale.

GPT-5.4: Enhanced Reliability in Production AIGPT-5.4 is designed for organizations operating AI in real production environments, where consistency, instruction adherence, and sustained context are critical to success. The model brings together advances in reasoning, coding, and agentic workflows to help AI systems not only plan tasks but complete them with fewer interruptions and reduced manual oversight.

Compared with earlier generations, GPT-5.4 emphasizes stability across longer interactions, enabling teams to deploy agentic AI with greater confidence in day-to-day production use.

GPT-5.4 introduces advancements that aim for production grade AI:

More consistent reasoning over time, helping maintain intent across multi‑turn and multi‑step interactionsEnhanced instruction alignment to reduce prompt tuning and oversightLatency improved performance for responsive, real-time workflowsIntegrated computer use capabilities for structured orchestration of tools, file access, data extraction, guarded code execution, and agent handoffsMore dependable tool invocation reducing prompt tuning and human oversightHigher‑quality generated artifacts, including documents, spreadsheets, and presentations with more consistent structureTogether, these improvements support AI systems that behave more predictably as tasks grow in length and complexity.

From capability to real-world outcomesGPT‑5.4 delivers practical value across a wide range of production scenarios where follow‑through and reliability are essential:

Agent‑driven workflows, such as customer support, research assistance, and business process automationEnterprise knowledge work, including document drafting, data analysis, and presentation‑ready outputsDeveloper workflows, spanning code generation, refactoring, debugging support, and UI scaffoldingExtended reasoning tasks, where logical consistency must be preserved across longer interactionsTeams benefit from reduced task drift, fewer mid‑workflow failures, and more predictable outcomes when deploying GPT‑5.4 in production.

GPT-5.4 Pro: Deeper analysis for complex decision workflowsGPT‑5.4 Pro, a premium variant designed for scenarios where analytical depth and completeness are prioritized over latency.

Additional capabilities include:

Multi‑path reasoning evaluation, allowing alternative approaches to be explored before selecting a final responseGreater analytical depth, supporting problems with trade‑offs or multiple valid solutionsImproved stability across long reasoning chains, especially in sustained analytical tasksEnhanced decision support, where rigor and thoroughness outweigh speed considerationsOrganizations typically select GPT‑5.4 Pro when deeper analysis is required such as scientific research and complex problems, while GPT‑5.4 remains the right choice for workloads that prioritize reliable execution and agentic follow‑through.

Microsoft Foundry: Enterprise‑Grade Control from Day OneGPT‑5.4 and GPT‑5.4 Pro are available through Microsoft Foundry, which provides the operational controls organizations need to deploy AI responsibly in production environments. Foundry supports policy enforcement, monitoring, version management, and auditability, helping teams manage AI systems throughout their lifecycle.

By deploying GPT‑5.4 through Microsoft Foundry, organizations can integrate advanced agentic capabilities into existing environments while aligning with security, compliance, and operational requirements from day one.

Customer Spotlight

Get Started with GPT-5.4 in Microsoft FoundryGPT‑5.4 sets a new bar for production‑ready AI by combining stronger reasoning with dependable execution. Through enterprise‑grade deployment in Microsoft Foundry, organizations can move beyond experimentation and confidently build AI systems that complete complex work at scale. Computer use capabilities will be introduced shortly after launch.

GPT‑5.4 in Microsoft Foundry is priced at $2.50 per million input tokens, $0.25 per million cached input tokens, and $15.00 per million output tokens. It is available at launch in Standard Global and Standard Data Zone (US), with additional deployment options coming soon. GPT‑5.4 Pro is priced at $30.00 per million input tokens, and $180.00 per million output tokens, and is available at launch in Standard Global.

Build agents for real-world workloads. Start building with GPT‑5.4 in Microsoft Foundry today.
The post Introducing GPT-5.4 in Microsoft Foundry appeared first on Microsoft Azure Blog.
Quelle: Azure