Add metadata to your Knowledge Base sources to deliver more relevant, localized, and precise answers, helping customers find what they need faster and improving overall resolution speed.

1. Adding metadata on knowledge import

When uploading files, URLs, or tabular data to the Knowledge Base, you can attach metadata at import time. This metadata is stored with each document or data chunk, enabling structured filtering and contextual retrieval later. For example, when importing car rental policies, you might tag each file with metadata like "locale": "US, CA, EU", or "serviceType": "car_rental, equipment_rental". This ensures that when the agent queries using metadata filters (static or dynamic), it only retrieves content relevant to the user’s local region or service context.

2. Dynamically, or statically apply metadata at runtime

From the Knowledge Base tool in your agent, define the metadata your agent should use when querying the tool. You can specify a static value (or variable) to consistently filter results, or let the agent dynamically assign metadata at runtime — allowing it to query the Knowledge Base contextually based on each unique conversation.

Example – Car Booking Service

If your Knowledge Base includes information for multiple locales (e.g., US, CA, EU), you can set a metadata field like locale. Instead of hardcoding a single locale, the agent can dynamically apply the user’s locale at runtime — for example:

If a user says “I want to book a car in New York,” the agent automatically filters Knowledge Base results with locale: US, ensuring responses only reference policies, pricing, and availability relevant to that locale.


We’ve added a set of built-in time variables that make it easier to access and use time within your agents—no external API calls or workarounds required. Perfect for agents that depend on current or relative time inputs.

Variable nameDescriptionExample
vf_nowThe current date and time formatted in a human-readable way. You can modify the timezone in project settings.Jan 1, 2025, 16:37
vf_dateThe current date formatted in a human-readable way.Jan 1, 2025
vf_timeThe current time formatted in a human-readable way.16:37
vf_monthThe current month.January
vf_dayThe current day of the month.1
vf_yearThe current year.2025
vf_user_timezoneThe user's timezone in the format. If unavailable, defaults to the project's timezone.America/Toronto

Project timezone can be set in project/behaviour settings:



We've added Deepgram Flux, their ASR newest model built specifically for Voice AI.

Flux is the first conversational speech recognition model built specifically for voice agents. Unlike traditional STT that just transcribes words, Flux understands conversational flow and automatically handles turn-taking.

Flux tackles the most critical challenges for voice agents today: knowing when to listen, when to think, and when to speak. The model features first-of-its-kind model-integrated end-of-turn detection, configurable turn-taking dynamics, and ultra-low latency optimized for voice agent pipelines, all with Nova-3 level accuracy.

Flux is Perfect for: turn-based voice agents, customer service bots, phone assistants, and real-time conversation tools.

Key Benefits:

  • Smart turn detection — Knows when speakers finish talking
  • Ultra-low latency — ~260ms end-of-turn detection
  • Early LLM responses — EagerEndOfTurn events for faster replies
  • Turn-based transcripts — Clean conversation structure
  • Natural interruptions — Built-in barge-in handling
  • Nova-3 accuracy — Best-in-class transcription quality

Converts text to speech in real time and keeps the spoken audio perfectly aligned with the displayed text. This ensures call transcripts are an accurate, word-for-word representation of what was actually said.

This setting lets you define how long a conversation can stay inactive before the transcript automatically ends.

This is different from session timeout — the session stays open, but the transcript closes after the set inactivity period, enabling more accurate reporting and evaluations.

Important: ending the transcript does not end the user’s ability to re-engage. If the user responds again, a new transcript will begin within the same session.


We’ve added a new Priority Processing setting for OAI-supported models. When enabled, your requests will be given higher processing priority for faster response times and reduced latency. Note: this will consume more credits.

MCP tools

by Michael Hood

Supercharge your agents by connecting directly to MCP servers.

  • 🔌 Connect to MCP servers in just a few clicks
  • 📥 Add MCP server tools to your agents
  • 🔄 Sync MCP servers to stay up-to-date

Bring in any tool, expand what your agents can do, and take your workflows to the next level.

Documentation

You can now enable your agents to forward calls to a different number, SIP address, or extension.

  • 📞 Seamlessly transfer callers to the right person or agent
  • 🔀 Supports phone numbers, SIP addresses, and extensions
  • 🛠️ Configure forwarding directly in your agent’s tools

This makes it easier to connect customers with the right destination without breaking the flow of the conversation.

Shareable links have been upgraded to better reflect the agent you’re building. Each link now points to a hosted version of your AI agent that mirrors your selected environment (dev, staging, production) and interface, so what you share is exactly what others will experience. Password protection is also available for secure access.

  • 🔗 Sharable links now mirror your actual AI agent
  • 🛠️ Environment-specific links (dev, staging, production)
  • 🎨 Customize the look and feel via the Interfaces tab
  • 🔒 Optional password protection for secure sharing