Welcome to the AI Multi-Agent Architecture Workshop, designed for organizations seeking to unlock the power of AI-driven intelligent agents. Over this 3-to-5-day interactive workshop, Microsoft architects will guide you step-by-step to build a private, secure AI system tailored to your business needs.
This workshop will teach you how to develop a multi-agent system capable of comprehending diverse datasets across various locations. These intelligent agents can answer questions with detailed explanations and source references, providing your organization with a powerful, ChatGPT-like experience designed for enterprise use.
This hands-on workshop will walk you through creating a Proof of Concept (POC) for a Generative AI Multi-Agent Architecture using Azure Services. By the end of the workshop, you'll have built:
-
A Scalable Backend
Developed with Bot Framework and FastAPI, the backend serves as the engine connecting AI logic to multiple communication channels, including:- Web Chat
- Microsoft Teams
- SMS
- Slack, and more!
-
A User-Friendly Frontend
Build a web application that combines:- A search engine capable of querying your data intelligently.
- A bot UI for seamless conversational experiences.
-
A RAG-Based Multi-Agent Architecture
Leverage Retrieval-Augmented Generation (RAG) to enable your agents to retrieve precise information and generate accurate responses.
- Step-by-Step Guidance: Each module builds upon the previous one, progressively introducing you to real-world AI architecture concepts.
- Custom Enterprise AI: Create intelligent agents that understand your organization’s data while maintaining privacy and security.
- Multi-Channel Capabilities: Deploy your agents across various platforms for broad accessibility.
- Practical Experience: Learn by doing, with notebooks and code samples tailored for an enterprise setting.
By the end of the workshop, you'll have a working knowledge of how to design, build, and deploy AI agents in a multi-agentic architecture. This hands-on experience will help you understand the value of Azure-powered Generative AI in solving real-world business problems.
This is a customer-funded Value-Based Delivery (VBD). Below, you'll find all the assets and resources needed for a successful workshop delivery.
Item | Description | Link |
---|---|---|
VBD SKU Info and Datasheet | CSAM must dispatch it as "Customer Invested" against credits/hours of Unified Support Contract. Customer decides if 3 or 5 days. | ESXP SKU page |
VBD Accreditation for CSAs | Links for CSAs to get the Accreditation needed to deliver the workshop | Link 1 , Link 2 |
VBD 3-5 day POC Asset (IP) | The MVP to be delivered (this GitHub repo) | Azure-Cognitive-Search-Azure-OpenAI-Accelerator |
VBD Workshop Deck | The deck introducing and explaining the workshop | Intro AOAI GPT Azure Smart Search Engine Accelerator.pptx |
CSA Training Video | 2 Hour Training for Microsoft CSA's | POC VBD Training Recording |
- Azure subscription
- Microsoft members preferably to be added as Guests in clients Azure AD. If not possible, then customers can issue corporate IDs to Microsoft members
- A Resource Group (RG) needs to be set for this Workshop POC, in the customer Azure tenant
- The customer team and the Microsoft team must have Contributor permissions to this resource group so they can set everything up 2 weeks prior to the workshop
- Customer Data/Documents must be uploaded to the blob storage account, at least two weeks prior to the workshop date
- A Single-Tenant App Registration (Service Principal) must be created by the customer (save the Client Id and Secret Value).
- Customer must provide the Microsoft Team , 10-20 questions (easy to hard) that they want the Agent/Bot to respond correctly.
- For IDE collaboration and standarization during workshop, AML compute instances with Jupyper Lab will be used, for this, Azure Machine Learning Workspace must be deployed in the RG
- Note: Please ensure you have enough core compute quota in your Azure Machine Learning workspace
- The user asks a question.
- In the backend app, an Agent determines which source to use based on the user input
- Five types of sources are available:
- 3a. Azure SQL Database - contains COVID-related statistics in the US.
- 3b. API Endpoints - RESTful OpenAPI 3.0 API from a online currency broker.
- 3c. Azure Bing Search API - provides access to the internet allowing scenerios like: QnA on public websites .
- 3d. Azure AI Search - contains AI-enriched documents from Blob Storage:
- Transcripts of the dialogue of all the episodes of the TV Show: FRIENDS
- 90,000 Covid publication abstracts
- 4 lenghty PDF books
- 3f. CSV Tabular File - contains COVID-related statistics in the US.
- The Agent retrieves the result from the correct source and crafts the answer.
- The Agent state is saved to CosmosDB as persistent memory and for further analysis.
- The answer is delivered to the user.
https://gptsmartsearch-frontend.azurewebsites.net
- 100% Python.
- Uses Azure Cognitive Services to index and enrich unstructured documents: OCR over images, Chunking and automated vectorization.
- Uses Hybrid Search Capabilities of Azure AI Search to provide the best semantic answer (Text and Vector search combined).
- Uses LangChain as a wrapper for interacting with Azure OpenAI , vector stores, constructing prompts and creating agents.
- Multi-Agentic Architecture using LangGraph.
- Multi-Lingual (ingests, indexes and understand any language)
- Multi-Index -> multiple search indexes
- Multi-modal input and output (text and audio)
- Tabular Data Q&A with CSV files and SQL flavor Databases
- Uses Azure AI Document Intelligence SDK (former Form Recognizer) to parse complex/large PDF documents
- Uses Bing Search API to power internet searches and Q&A over public websites.
- Connects to API Data sources by converting natural language questions to API calls.
- Uses CosmosDB as persistent memory to save user's conversations.
- Uses Streamlit to build the Frontend web application in python.
- Uses Bot Framework and Bot Service to Host the Bot API Backend and to expose it to multiple channels including MS Teams.
- Uses also FastAPI to deploy an alternative backend API with streaming capabilites
You must have an Azure OpenAI Service already created.
- Fork this repository to your GitHub account.
In Azure OpenAI Studio, deploy the following models:
(Note: Older versions of these models will not work)
gpt-4o
gpt-4o-mini
text-embedding-3-large
tts
whisper
- Create a Resource Group (RG) to house all the assets for this accelerator.
- Note: Azure OpenAI services can exist in a different RG or even a different subscription.
Click the button below to deploy all necessary Azure infrastructure (e.g., Azure AI Search, Cognitive Services, etc.):
Important:
If this is your first time creating an Azure AI Services Multi-Service Account, do the following manually:
- Go to the Azure portal.
- Create the account.
- Read and accept the Responsible AI Terms.
Once done, delete this manually created account and then use the above deployment button.
-
Clone your forked repository to your AML Compute Instance.
- If your repository is private, refer to the Troubleshooting section for guidance on cloning private repos.
-
Install the dependencies in a Conda environment. Run the following commands on the Python 3.12 Conda environment you plan to use for the notebooks:
conda create -n GPTSearch python=3.12 conda activate GPTSearch pip install -r ./common/requirements.txt conda install ipykernel python -m ipykernel install --user --name=GPTSearch --display-name "GPTSearch (Python 3.12)"
- Create a Python virtual environment (.venv):
- When creating the virtual environment, select the
./common/requirements.txt
file. - Alternatively, install dependencies manually:
pip install -r ./common/requirements.txt
- When creating the virtual environment, select the
- Activate the virtual environment:
.venv\scripts\activate
- Install
ipykernel
:pip install ipykernel
Edit the credentials.env
file with the appropriate values from the services created in Step 4.
- To obtain
BLOB_SAS_TOKEN
andBLOB_CONNECTION_STRING
, navigate to:
Storage Account > Security + Networking > Shared Access Signature > Generate SAS
- Execute the notebooks in order, as they build on top of each other.
- Use the appropriate kernel:
- For AML, select:
GPTSearch (Python 3.12)
- For VS Code, select the
.venv
kernel.
- For AML, select:
- If cloning a private repository: Refer to the detailed guide here.
- For issues with dependency installation: Ensure your Python version matches the required version.
Troubleshooting
Steps to clone a private repo:
- On your Terminal, Paste the text below, substituting in your GitHub email address. Generate a new SSH key.
ssh-keygen -t ed25519 -C "your_email@example.com"
- Copy the SSH public key to your clipboard. Add a new SSH key.
cat ~/.ssh/id_ed25519.pub
# Then select and copy the contents of the id_ed25519.pub file
# displayed in the terminal to your clipboard
- On GitHub, go to Settings-> SSH and GPG Keys-> New SSH Key
- In the "Title" field, add a descriptive label for the new key. "AML Compute". In the "Key" field, paste your public key.
- Clone your private repo
git clone git@github.com:YOUR-USERNAME/YOUR-REPOSITORY.git
This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com.
When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.
This project may contain trademarks or logos for projects, products, or services. Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft's Trademark & Brand Guidelines. Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship. Any use of third-party trademarks or logos are subject to those third-party's policies.