You can bring agents developed in Python in any framework (LangChain, CrewAI, or any custom framework) and deploy them on Beamlit by integrating a few lines of the Beamlit SDK and leveraging our other developer tools (the Beamlit CLI, GitHub action, etc.).

Check out this Getting Started tutorial in order to develop and deploy your first Hello World AI agent globally in less than 5 minutes.

Beamlit agent file structure

Overview

To deploy a custom AI agent on Beamlit, you need the following file in your repository:

  • agent.py (or src/agent.py): This file must contain an async main function that runs your agent. The agent.main function serves as the default entry point during serving and deployment.
    • You can specify a different module to serve using the flag bl serve --module your-function-path.your-function-name
    • The function accepts these arguments, which Beamlit passes when a request is made to your deployed agent:
      • request: The FastAPI Request that triggers the agent’s execution.
      • agent: (Optional) The core agent algorithm. By default, it uses a LangChain ReAct agent.
      • model: (Optional) The LLM chat API that powers the agent. By default, it uses the model API specified during agent creation, passed in LangChain format ChatModel
      • functions: (Optional) The functions available for tool calling. By default, it uses the functions from the src/functions folder, passed in LangChain format BaseTool.

Mark your agent’s main function with the @agent decorator to serve it with Beamlit. You can override the default argument values that Beamlit passes upon requesting the agent by using the decorator parameters (see below).

Some folders are optional but designed to handle specific parts of the agent code on Beamlit:

  • src/functions: all files listed in this folder will be interpreted as custom tools that will all be made available to the agent automatically when serving or deploying the agent. The tool’s Python function itself is identified with the @function decorator.

Quickstart

It is required to have uv installed to use the following command.

You can quickly initialize a new project from scratch by using the CLI command bl create-agent-app. This will create a pre-scaffolded local repo where your entire code can be added. By default, this command creates a boilerplate agent that has:

  • A LangChain ReAct agent as the core agent logic. This agent can be overridden by passing a custom agent as the override-agent parameter to the @agent decorator in /src/agent.py

  • A hello-world tool (a function) that returns a placeholder text string. This function can be customized by changing the code in the helloworld function in /src/functions/helloworld.py, and additional functions can be created in the /src/functions folder.

The @agent decorator

The @agent decorator identifies the main function that handles your core agent logic such as how to handle queries and responses, tool calls and such.

@agent()
async def main(request, agent):
	# Your main agent logic
	...

Inside the decorator, you can pass several parameters:

  • agent: this can be either the Beamlit SDK’s Agent object or equivalent dict, and allows to specify all the parameters for the deployment of your agent on Beamlit. Examples include:
    • metadata.name: the name of the agent on Beamlit
    • spec.description: the description of the agent on Beamlit
    • spec.model: the name of a model API deployed in your Beamlit workspace, to use as core chat model for the agent
  • override_agent: You can override the default LangChain ReAct agent by passing any custom agent instance. When the deployed agent receives a request, Beamlit will pass this as the agent argument to the agent’s main() function.
  • override_model: You can query an LLM provider directly instead of routing through Beamlit by passing a ChatModel object from LangChain. When the deployed agent receives a request, Beamlit will pass this as the llm to the default LangChain ReAct agent. This is used as a more granular alternative to override_agent.
  • remote_functions: You can pass on additional Beamlit functions that are already deployed so your agent can access them too, in addition to the functions from src/functions/. This parameter takes in an array of string representing the names of the functions on Beamlit. This is used as a more granular alternative to override_agent.

Example

Here’s an example that demonstrates overriding the default agent object.

from fastapi import Request
from langchain_openai import ChatOpenAI
from langgraph.checkpoint.memory import MemorySaver
from langgraph.prebuilt import create_react_agent

from beamlit.agents import agent
from customfunctions.helloworld import helloworld

chat = ChatOpenAI()
memory = MemorySaver()
custom_agent = create_react_agent(chat, tools=[helloworld], checkpointer=memory)

@agent(override_agent=custom_agent)
async def main(request: Request, agent):
	# Your main agent logic
	...

When developing custom agents, you may also need to retrieve instantiated model APIs and functions in your code. The following section shows you how to.

Retrieve existing functions and model APIs

Get model APIs

You can use model APIs deployed on Beamlit by calling get_chat_model(). This function returns a model API in LangChain format ChatModel

from beamlit.agents.chat import get_chat_model

model = get_chat_model("model-name-on-beamlit")

custom_agent = create_react_agent(model, tools=[])

Get functions

You can use functions defined both in your code and deployed on Beamlit by calling get_functions(). This function returns a list of tools in LangChain format BaseTool.

from beamlit.functions import get_functions
from langgraph.prebuilt import create_react_agent

functions = get_functions(warning=False)

custom_agent = create_react_agent(chat, tools=functions)

By default, this returns the list of functions defined in the local folder src/functions/. You can pass several parameters:

  • remote_functions: to retrieve additional Beamlit functions that are already deployed so your agent can access them too, in addition to the functions from src/functions/. This parameter takes in an array of string representing the names of the functions on Beamlit

Deploy an agent

Learn how to deploy your custom AI agents on Beamlit as a serverless endpoint.