This solution accelerator automates the deployment of a complete, production-ready AI application environment in Azure. It provisions Azure AI Foundry, Microsoft Fabric, Azure AI Search, and Microsoft Purview—all pre-wired with private networking, managed identities, and governance controls.
This repository is intended to be used as a solution accelerator following the open-source license terms listed in the GitHub repository. The intended purpose is to demonstrate how organizations can:
- Deploy production-grade AI infrastructure with security controls enabled by default
- Integrate data platforms (Fabric) with AI services (Foundry) for retrieval-augmented generation scenarios
- Implement governance and compliance monitoring through Purview integration
The deployed infrastructure is designed to host AI chat applications that use your organization's documents for grounded responses.
The solution was evaluated for:
- Infrastructure Security: Private endpoints, managed identities, and RBAC configurations were validated against Microsoft's Well-Architected Framework
- Deployment Reliability: Automated testing in multiple Azure regions to ensure consistent provisioning
- Integration Correctness: Validation that Fabric, Search, Foundry, and Purview components are properly connected
- Infrastructure Only: This accelerator deploys infrastructure and basic integrations. You must still provide your own documents, configure AI prompts, and build the end-user application.
- Region Availability: Not all Azure regions support all required services. EastUS2 is recommended.
- Quota Requirements: Azure OpenAI and other services have quota limits that may restrict deployment.
- Network Restrictions: Some Fabric-to-Search private link scenarios are not yet fully supported by Azure.
- English Only: Documentation and sample configurations are provided in English only.
Users can customize several parameters to align the deployment with their organizational requirements:
- Network Isolation Mode: Configure the level of network isolation (private endpoints, public access)
- Model Selection: Choose which Azure OpenAI models to deploy
- Capacity Settings: Adjust Fabric capacity SKU and AI service quotas
- Governance Configuration: Connect to existing Purview accounts for compliance monitoring
Please note that these parameters are provided as guidance to start the configuration. Users should adjust the system to meet their specific security, compliance, and operational requirements.
- Check Quota Before Deployment: Use the quota check guide to ensure sufficient capacity
- Review Post-Deployment Steps: Validate all components are properly configured after deployment
- Customize Prompts: Modify system prompts and configurations for your specific use case
- Human Review: Implement human-in-the-loop validation for AI-generated content in production scenarios
- This accelerator does not include sample data or documents
- Documents you upload to Fabric lakehouses are stored in your Azure subscription
- Data is indexed by Azure AI Search within your subscription
- No data is sent to Microsoft beyond standard Azure service telemetry
This accelerator deploys Azure OpenAI models (GPT-4o by default) within your Azure subscription. All AI model usage:
- Is subject to Azure OpenAI Service terms
- Is metered and billed to your Azure subscription
- Follows your organization's Azure OpenAI content filtering policies