Skip to main content

Voice Agents

Voice agents are the core of HubTalk AI. Each agent combines a language model (LLM), speech recognition (ASR), and text-to-speech (TTS) to hold real-time conversations with callers.

Agent List

The Agents section displays all agents in your organization. You can search by agent name or company name. Each agent shows its current configuration and status.

Creating an Agent

To create a new agent, click Create Agent in the top right corner. The agent editor has two main areas:

Component Selection

At the top of the agent editor, select the three core components:
ComponentPurpose
LLMThe language model that powers the agent’s reasoning and responses
ASRAutomatic Speech Recognition — converts caller’s speech to text
TTSText-to-Speech — converts the agent’s text responses to spoken audio

System Prompt

Below the component selector is the main prompt editor. HubTalk AI uses a single prompt model — you write one comprehensive system prompt that defines the agent’s personality, instructions, constraints, and behavior.
Write your prompt in the language you want the agent to use by default. The agent can switch languages during conversation if instructed.

Greeting Configuration

Configure how the agent greets callers:
  • Fixed greeting — A static message spoken at the start of every call
  • Dynamic greeting — A greeting that incorporates variables (caller name, time of day, campaign data)
The greeting is synthesized via TTS and played immediately when the call connects, before the LLM processes any input. This reduces perceived latency.

Right Panel Settings

The right panel provides access to advanced configuration:
SettingDescription
FunctionsAdd operator transfer, call ending, and custom API functions
Post-call AnalysisConfigure variables extracted after each call
VAD SettingsFine-tune voice activity detection (interruption, silence, speech thresholds)
Speech SettingsConfigure agent behavior during caller silence
WebhookSet the URL where call data is sent after each conversation
Knowledge BaseAttach document collections the agent uses to answer questions

VAD Settings

Voice Activity Detection (VAD) controls how the agent handles interruptions, pauses, and speech boundaries.
ParameterDescription
Interruption threshold (first/last slider)Controls how the agent reacts when the caller starts speaking over the agent
Max speech pauseHow long the agent waits silently if the caller pauses mid-dialog
Min speech counterMinimum speech length for the system to count it as “real” speech
Interruption min speech counterHow much “confident speech” is required before the caller can interrupt the agent
Silence counter for buffer resetAfter prolonged silence, the system clears the accumulated audio buffer
All parameters are measured in audio chunks, except for the interruption settings.

Speech Settings

Speech settings control how the agent responds to caller silence. For example, you can configure the agent to ask “Are you still there?” every 7 seconds of silence, up to 3 times.
ParameterDescription
Silence intervalSeconds of silence before the agent speaks
Max promptsNumber of times the agent prompts the caller before taking action

Chat Testing

Every agent has a built-in chat testing feature accessible from the agent editor. Use it to:
  • Test prompt behavior with simulated conversations
  • Verify function triggers and responses
  • Check knowledge base retrieval
  • Validate multilingual switching
Chat testing simulates the text layer only. For full voice testing (ASR + TTS), make a real test call through the telephony system.