Azure AI Foundry integration
Connect your agents to LLMs deployed in Azure AI Inference, Azure OpenAI Service, and Azure AI Services.
The Azure AI Foundry integration allows Beamlit users to call models deployments from Azure AI Foundry services (Azure AI Inference, Azure OpenAI Service, and Azure AI Services) through a Beamlit endpoint that unifies access control, credentials and observability management.
There are 2 types of integrations related to this service:
- Azure AI Inference: connect to a model endpoint deployed as an “Azure AI Services” model on Azure. This typically includes OpenAI models.
- Azure AI Marketplace: connect to a model deployed from the Azure Marketplace. This typically includes Llama models.
The integration must be set up by an admin in the Integrations section in the workspace settings.
Azure AI Inference
Set up the integration
In order to use this integration, you must register an Azure AI Inference endpoint and access key into your Beamlit workspace settings.
First, go to the Azure AI Foundry console, and open your project. Select the “Azure AI Inference” capability, and retrieve both:
- the API key
- the Azure AI model inference endpoint
On Beamlit, in Workspace Settings > Azure AI Inference integration, create a new connection and paste the endpoint and the Access token there.
Connect to a model
Once you’ve set up the integration in the workspace, any workspace member can use it to reference an “Azure AI Services” model as an external model API.
When creating a model API, select “Azure AI Inference”. Then, input the name of your model just as it is deployed on Azure.
After the model API is created, you will receive a dedicated global Beamlit endpoint to call the model. Beamlit will forward inference requests to Azure, using your Azure credentials for authentication and authorization.
Azure AI Marketplace
Set up the integration & connect to a model
In order to use this integration, you must register an Azure endpoint and access token into your Beamlit workspace settings.
First, go to the Azure AI Foundry console, and open your project. Go to your models, and open the model you want to connect to. Retrieve:
- the API key
- the Azure AI model inference endpoint
On Beamlit, in Workspace Settings > Azure Marketplace integration, create a new connection and paste this token into the “Access token” section.
Once you’ve set up the integration in the workspace, any workspace member can use it to connect to the model as an external model API.
When creating a model API, select Azure Marketplace, input the name of the endpoint as you want it on Beamlit, and finish creating the model.
After the model API is created, you will receive a dedicated global Beamlit endpoint to call the model. Beamlit will forward inference requests to Azure, using your Azure credentials for authentication and authorization.