azure-functions-openai-rag-cosmosdb-go

このプロジェクトは、Azure FunctionsとOpenAIを活用して、Cosmos DBと連携するリッチなアプリケーションを構築するためのBicepテンプレートです。データの取得と処理を自動化し、効率的なワークフローを実現します。

GitHubスター

0

ユーザー評価

未評価

お気に入り

0

閲覧数

17

フォーク

0

イシュー

0

README
Azure Functions
Using Azure Functions OpenAI trigger and bindings extension to import data and query with Azure Open AI and Azure Cosmos DB and exposed as an MCP tool

This sample contains an Azure Function using OpenAI bindings extension to highlight OpenAI retrieval augmented generation with Cosmos DB.

You can learn more about the OpenAI trigger and bindings extension in the GitHub documentation and in the Official OpenAI extension documentation

Information about Model Context Protocol tools in Azure Functions is available on in the Azure Functions MCP blog

Prerequisites
Prepare your local environment
Create Azure OpenAI and Azure Cosmos DB resources for local and cloud dev-test

Once you have your Azure subscription, run the following in a new terminal window to create Azure OpenAI, Azure Cosmos DB and other resources needed: You will be asked if you want to enable a virtual network that will lock down your OpenAI and Cosmos DB so they are only available from the deployed function app over private endpoints. To skip virtual network integration, select true. If you select networking, your local IP will be added to the OpenAI and Cosmos DB services so you can debug locally.

azd init --template https://github.com/eamonoreilly/azure-functions-openai-cosmosdb-mcp-go

Make sure to run this before calling azd to provision resources so azd can run scripts required to setup permissions

Mac/Linux:

chmod +x ./infra/scripts/*.sh 

Windows:

set-executionpolicy remotesigned

Run the follow command to provision resoruces in Azure

azd provision

If you don't run azd provision, you can create an OpenAI resource and an Cosmos DB resource in the Azure portal to get your endpoints. After it deploys, click Go to resource and view the Endpoint value. You will also need to deploy a model, e.g. with name chat with model gpt-4o and embeddings with model text-embedding-3-small

Create local.settings.json (Should be in the same folder as host.json. Automatically created if you ran azd provision)
{
  "IsEncrypted": false,
  "Values": {
    "AzureWebJobsStorage": "UseDevelopmentStorage=true",
    "FUNCTIONS_WORKER_RUNTIME": "custom",
    "AZURE_OPENAI_ENDPOINT": "<paste from above>",
    "CHAT_MODEL_DEPLOYMENT_NAME": "chat",
    "cosmosDBNoSqlEndpoint__Endpoint": "<paste from above>",
    "EMBEDDING_MODEL_DEPLOYMENT_NAME": "embeddings",
    "SYSTEM_PROMPT": "You must only use the provided documents to answer the question"
    }
}
Permissions
Add your account (your account email, for example: contoso@microsoft.com) with the following permissions to the Azure OpenAI and Azure Cosmos DB resources when testing locally.

If you used azd provision this step is already done - your logged in user and your function's managed idenitty already have permissions granted.

  • Cognitive Services OpenAI User (OpenAI resource)
  • Cosmoso DB Data Contributor (Cosmos DB resource)
Access to Azure OpenAI and Azure Cosmos DB with virtual network integration

If you selected virtual network integration, access to Azure OpenAI and Azure Cosmos DB is limited to the Azure Function app through private endpoints and cannot be reached from the internet. To allow testing from your local machine, you need to go to the networking tab in Azure OpenAI and Azure Cosmos DB and add your client ip to the allowed list. If you used azd provision this step is already done.

Run your app using Visual Studio Code
  1. Open a new terminal window in the ./app folder and enter func start
  2. Run F5 from the root of the repo and it should ask to attach to a process. Pick main
  3. Using your favorite REST client, e.g. RestClient in VS Code, PostMan, curl, make a post. test.http has been provided to run this quickly.
Add the Functions endpoint as an MCP Server

Open up the .vscode/mcp.json and start the local MCP server that points to the running function app

Deploy to Azure

Run this command to provision the function app, with any required Azure resources, and deploy your code:

azd up

You're prompted to supply these required deployment parameters:

Parameter Description
Environment name An environment that's used to maintain a unique deployment context for your app. You won't be prompted if you created the local project using azd init.
Azure subscription Subscription in which your resources are created.
Azure location Azure region in which to create the resource group that contains the new Azure resources. Only regions that currently support the Flex Consumption plan are shown.

After publish completes successfully, azd provides you with the URL endpoints of your new functions, but without the function key values required to access the endpoints. To learn how to obtain these same endpoints along with the required function keys, see Invoke the function on Azure in the companion article Quickstart: Create and deploy functions to Azure Functions using the Azure Developer CLI.

Test MCP server in Azure Functions

Open up the .vscode/mcp.json and start the remote function app. It will ask for the function app name and also a key.
The system key can be obtained from the portal (under keys) or the CLI (az functionapp keys list --resource-group --name )

Redeploy your code

You can run the azd up command as many times as you need to both provision your Azure resources and deploy code updates to your function app.

[!NOTE]
Deployed code files are always overwritten by the latest deployment package.

Clean up resources

When you're done working with your function app and related resources, you can use this command to delete the function app and its related resources from Azure and avoid incurring any further costs (--purge does not leave a soft delete of AI resource and recovers your quota):

azd down --purge