Skip to main content
Configure agents to use an external LLM instead of our default models. Maintain full control over the language model while leveraging Beyond Presence’s optimized avatar and agent management stack.

Use Cases

External LLM configuration is ideal for:
  • Custom models: Use fine-tuned or specialized language models
  • Data control: Route LLM requests through your own infrastructure or preferred providers
  • Regulatory compliance: Meet specific data residency requirements

Requirements

Your external LLM API must implement a single endpoint following the OpenAI-compatible Chat Completions API format. At minimum, the endpoint must support:
  • Standard chat completion requests with messages array
  • Streaming responses with server-sent events
Additional features like function calling or vision are optional and only needed if your specific use case requires them.
While not strictly required, we recommend supporting token authentication for secure access.
Many LLM providers offer OpenAI-compatible endpoints that work out of the box: xAI (Grok): Get API key
  • Base URL: https://api.x.ai/v1
  • Models: grok-4, grok-3-beta, grok-code-fast-1
Google (Gemini): Get API key
  • Base URL: https://generativelanguage.googleapis.com/v1beta/openai
  • Models: gemini-2.0-flash, gemini-1.5-pro, gemini-1.5-flash
Using these providers may incur additional costs based on their pricing models.

Configuration

You can configure external LLMs through our API. Dashboard support is planned for a future release.
  1. Register your LLM API with the Create External API Configuration endpoint
  2. Reference the configuration when creating an agent with the Create Agent endpoint
While we require you to specify a token for the API, you can pass an empty string if your endpoint does not require authentication.

Complete Example

See our external LLM example for full implementation including API registration, agent creation, and optional deployment of a compatible LLM API.

Next Steps