Azure AI Foundry: What's New From Build 2025


Introduction
Let’s be honest: keeping up with Microsoft Build announcements can feel like drinking from a fire hose - especially when AI is involved! But if, like me, you’re interested in Azure AI Foundry, this year’s updates are worth paying close attention to.
Following on from my last post - Why Use Azure AI Foundry? - I wanted to zoom in on the key announcements from Build 2025 and what they actually mean in practice. Not just the “what”, but the “why” - especially if you’re a builder trying to make sense of where to focus.
I won’t attempt to cover every announcement (there were loads!) - but I have pulled together the ones that stood out most to me from the Foundry perspective.
If you want the full rundown, the Build Book of News is a great companion resource.
Model Router
What it is:
This one is really cool! The Model Router is a new feature that automatically selects the most suitable Azure OpenAI model for your specific prompt, optimising for performance and cost.
By evaluating factors like query complexity, cost, and performance, it intelligently routes requests to the most suitable model.
Why it matters:
Reduces the need for manual model selection
Enhances response quality by choosing the best-fit model
Optimises costs by selecting the most efficient model
Example use cases:
Dynamic selection between GPT-4 and GPT-3.5 based on prompt complexity
Automatically routing image-related prompts to vision models
Selecting lightweight models for simple tasks to save costs
More Models & More Capacity
What it is:
Soooo many models! In Azure AI Foundry, there are models sold directly by Microsoft, and models from third parties. The update here is that you can now use models from AI focussed tech companies xAI, Black Forest Labs, and Hugging Face.
On top of that, Microsoft is extending reserved capacity to cover Azure OpenAI and select Foundry Models (including Black Forest Labs, and xAI). This means you get consistent performance even when demand spikes.
Why it matters:
Huge increase in model choice - over 11,000!
Consistent performance under load with reserved capacity options
Fine-tune and experiment without infrastructure overhead
Example use cases:
Experiment with emerging models in development, then scale to production seamlessly
Ensure consistent response times for AI services during high-traffic periods
Foundry Local
What it is:
Foundry Local brings Azure AI Foundry capabilities directly to your own infrastructure – whether that’s a developer workstation, edge device, or air-gapped data centre. This includes support for offline execution! Imagine interacting with an AI chatbot as a Windows app on your laptop - even when completely offline!
Why it matters:
Enables AI use in scenarios with strict data privacy or sovereignty requirements
Delivers sub-second response times without a cloud round-trip
Supports hybrid and offline environments with no connectivity needed
Example use cases:
Run AI models at the edge in manufacturing, retail, or healthcare settings
Enable offline document processing or vision capabilities on laptops
Deploy secure, private agents in government or defence environments
Fine-Tuning & Developer Tier
What it is:
Fine-tuning has traditionally been tricky – not just technically, but also in terms of cost and where you could actually run it. That’s changing.
Fine-tuning allows you to retrain a model on your own data so it adapts to your domain. This is different from grounding, which connects a model to external data at runtime without changing how the model is trained.
With the public previews of Global Training and the new Developer Tier, Azure AI Foundry is making fine-tuning more accessible than ever. You can now fine-tune the latest Azure OpenAI models in new worldwide regions, with lower pricing designed specifically for experimentation and iteration.
Global Training handles the infrastructure behind the scenes – and the Developer Tier removes the upfront hosting cost, so you only pay when you actually train or use a model.
Why it matters:
Run fine-tuning closer to your data with expanded regional support
Experiment more freely with reduced pricing
Skip the infra setup – Foundry handles it for you
Example use cases:
Trial multiple fine-tuning approaches in a low-cost environment before committing to production
Fine-tune a model in-region to meet data residency requirements for financial or healthcare data
Quickly test how adding domain-specific examples affects summarisation performance - without setting up infrastructure
Build and evaluate early-stage agent prototypes on the Developer Tier, then scale seamlessly to production using the same workflows
Multi-agent Orchestration
What it is:
Azure AI Foundry now includes native tools for designing and coordinating multiple AI agents within a single system. This goes beyond prompt chaining - agents can now have distinct roles, shared memory, and coordinated workflows, all managed within Foundry. I need to write about this, and Agents in general, in a lot more detail!
As part of this, Microsoft also introduced an agent catalogue: a growing library of pre-built, configurable agents for common tasks like retrieval, planning, evaluation, and summarisation. You can use them as-is or customise them to fit your specific needs.
Why it matters:
Enables more advanced, multi-step AI use cases
Removes the need for custom orchestration logic
Encourages modular, maintainable agent design
Supports real collaboration between specialised agents
Example use cases:
A planning agent delegates tasks to retrieval, generation, and validation agents
A multi-agent customer service flow handles triage, resolution, and escalation
A compliance assistant splits work across extraction, analysis, and reporting agents
A document workflow uses separate agents to summarise, translate, and format content
Identity for Agents
What it is:
Microsoft Entra Agent ID is a new capability that brings enterprise-grade identity and access management to AI agents. Just like you’d give an app or service a managed identity, you can now give AI agents their own secure, verifiable identity within your organisation.
This allows agents to authenticate, authorise, and operate securely across your systems, with support for auditing, policy enforcement, and lifecycle management - all integrated with Microsoft Entra.
Why it matters:
Secures agent interactions with APIs, data, and enterprise resources
Enables RBAC and policy enforcement for AI agents
Improves traceability and auditing of agent actions
Aligns agent behaviour with existing identity and governance frameworks
Example use cases:
An AI agent authenticates with Entra to access SharePoint or Microsoft Graph
Different agents have scoped permissions based on function - e.g. read-only vs full write access
Agent activity is logged and monitored alongside human and app identities
Azure AI Foundry Observability
What it is:
Azure AI Foundry Observability is a unified solution for governance, evaluation, tracing, and monitoring. It brings real-time visibility into models, agents, workflows, and user interactions - all from a single view.
In my last post, I briefly described evaluation and monitoring in Foundry. I’ve not had chance to look at this in greater detail as yet, but as I understand this announcement, Foundry Observability combines what existed before (within Foundry itself), such as live request tracing, model metrics, agent behaviours, evaluation results, etc - but now fully integrated with Azure Monitor, Application Insights, and the Foundry portal itself.
Why it matters:
One integrated view - no more tool sprawl
Real-time tracing and evaluation across agents and models
Built-in governance features to support audits and responsible AI
Makes it easier to go from prototype to production with confidence
Example use cases:
Trace a user request across multiple agents in a multi-turn workflow
Set alerts when accuracy or safety thresholds are breached
Track agent behaviour over time to spot drift or unexpected changes
Export evaluation logs for compliance or internal reviews
And The Rest!
There’s a lot I haven’t covered here - from updates to Agentic Retrieval in Azure AI Search, to Agent Evaluators, and improvements to the Foundry API and SDK.
I’ve also noticed growing overlap with tools like Copilot Studio and Azure Logic Apps, which adds more capability - but also more complexity. It’s a lot to take in, but I hope this post has helped you cut through the noise and focus on what’s new and important in Azure AI Foundry.
The link below (click the image) will take you to another excellent write-up, and the Build Book of News is another fantastic resource for exploring the full range of announcements.
If you’d prefer to see it all in action, I highly recommend the Build session: Azure AI Foundry: The AI App and Agent Factory - complete with demos of many of the updates mentioned here.
Disclaimer: The views expressed in this blog are my own and do not necessarily reflect those of my employer or Microsoft.
Subscribe to my newsletter
Read articles from Daniel McLoughlin directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by

Daniel McLoughlin
Daniel McLoughlin
Daniel McLoughlin is a Technology Strategist with extensive experience in Microsoft Azure and the broader Microsoft ecosystem. Working for a leading Microsoft Partner in the UK and Ireland, he provides strategic direction across technical, commercial, and operational domains. His strategic work focuses on Microsoft's Cloud Adoption Framework (CAF) and Well-Architected Framework (WAF), establishing governance models, defining organisational standards, and developing architectural patterns and practices. Based in Yorkshire, UK, Daniel combines his professional role with family life as a husband and father. His interests extend beyond technology to philosophy and language learning, bringing additional perspective to his strategic work. Areas of Focus Cloud Strategy & Governance Microsoft Framework Alignment Partner Programme Leadership Standards & Compliance Strategic Architecture Microsoft Partnership Development Community Engagement Technical Mentoring