Skip to main content
Every persona needs a language model to power its conversations. Anam provides several built-in LLMs you can use without any additional setup — just reference the model’s ID in your persona configuration.

Built-in models

LLM IDModelBest for
0934d97d-0c3a-4f33-91b0-5e136a0ef466OpenAI GPT-4.1 MiniRecommended for most projects
a7cf662c-2ace-4de1-a21e-ef0fbf144bb7GPT OSS 120BHigh throughput reasoning, great at tool calling
27cbd128-f1e6-4b67-8ab3-9123659be08cGemini 3 Flash PreviewFast reasoning with predictable tool calling
9d8900ee-257d-4401-8817-ba9c835e9d36Gemini 2.5 FlashOur fastest model
88190a76-3e87-4935-ab39-f4f73038815aKimi k2Great at agentic tasks
ANAM_LLAMA_v3_3_70B_V1Llama 3.3 70BOpen-source preference, larger context

Using a built-in LLM

Set the llmId field in your persona configuration to the ID of the model you want to use:
const personaConfig = {
  name: "Cara",
  avatarId: "30fa96d0-26c4-4e55-94a0-517025942e18",
  voiceId: "6bfbe25a-979d-40f3-a92b-5394170af54b",
  llmId: "0934d97d-0c3a-4f33-91b0-5e136a0ef466",
  systemPrompt: "You are Cara, a helpful customer service representative.",
};
You can also select a model from the dropdown in Anam Lab when creating or editing a persona.

Choosing a model

For most use cases, GPT-4.1 Mini is a good starting point — it balances speed, cost, and quality. If your persona uses tools heavily, consider GPT OSS 120B or Gemini 3 Flash Preview for more reliable tool calling. If latency is your top priority, Gemini 2.5 Flash is the fastest option.

Greeting behavior

When using a built-in LLM, the persona greets the user with an opening message when the session starts. The content of this greeting is controlled by the system prompt. To skip the greeting entirely, set skipGreeting to true:
const personaConfig = {
  name: "Cara",
  avatarId: "30fa96d0-26c4-4e55-94a0-517025942e18",
  voiceId: "6bfbe25a-979d-40f3-a92b-5394170af54b",
  llmId: "0934d97d-0c3a-4f33-91b0-5e136a0ef466",
  systemPrompt: "You are Cara, a helpful customer service representative.",
  skipGreeting: true,
};
This is useful when you want the user to initiate the conversation, or when the persona is responding to an event rather than starting a dialogue.

Bring your own LLM

If the built-in models don’t fit your needs, you can connect your own:
  • Server-side custom LLMs — Register your model with Anam and we call it from our servers, keeping latency low.
  • Client-side custom LLMs — Handle LLM calls yourself in your client code using CUSTOMER_CLIENT_V1 as the LLM ID.
  • LiveKit — Use Anam as a face layer in your existing LiveKit agent pipeline with any LLM.