Telnyx telephony integration
by Michael HoodYou can now connect your Telnyx account to import and manage phone numbers directly in Voiceflow, enabling Telnyx as your telephony provider for both inbound and outbound calls..
You can now connect your Telnyx account to import and manage phone numbers directly in Voiceflow, enabling Telnyx as your telephony provider for both inbound and outbound calls..
Added native support for DTMF keypad input in phone conversations. Users can now enter digits via their phone keypad, sending a DTMF trace to the runtime. Configure timeout and delimiters (#, *) to control when input is processed. See documentation here.
Add metadata to your Knowledge Base sources to deliver more relevant, localized, and precise answers, helping customers find what they need faster and improving overall resolution speed.
1. Adding metadata on knowledge import
When uploading files, URLs, or tabular data to the Knowledge Base, you can attach metadata at import time. This metadata is stored with each document or data chunk, enabling structured filtering and contextual retrieval later. For example, when importing car rental policies, you might tag each file with metadata like "locale": "US, CA, EU", or "serviceType": "car_rental, equipment_rental". This ensures that when the agent queries using metadata filters (static or dynamic), it only retrieves content relevant to the user’s local region or service context.
2. Dynamically, or statically apply metadata at runtime
From the Knowledge Base tool in your agent, define the metadata your agent should use when querying the tool. You can specify a static value (or variable) to consistently filter results, or let the agent dynamically assign metadata at runtime — allowing it to query the Knowledge Base contextually based on each unique conversation.
Example – Car Booking Service
If your Knowledge Base includes information for multiple locales (e.g., US, CA, EU), you can set a metadata field like locale. Instead of hardcoding a single locale, the agent can dynamically apply the user’s locale at runtime — for example:
If a user says “I want to book a car in New York,” the agent automatically filters Knowledge Base results with locale: US, ensuring responses only reference policies, pricing, and availability relevant to that locale.
We’ve added a set of built-in time variables that make it easier to access and use time within your agents—no external API calls or workarounds required. Perfect for agents that depend on current or relative time inputs.
| Variable name | Description | Example |
|---|---|---|
| vf_now | The current date and time formatted in a human-readable way. You can modify the timezone in project settings. | Jan 1, 2025, 16:37 |
| vf_date | The current date formatted in a human-readable way. | Jan 1, 2025 |
| vf_time | The current time formatted in a human-readable way. | 16:37 |
| vf_month | The current month. | January |
| vf_day | The current day of the month. | 1 |
| vf_year | The current year. | 2025 |
| vf_user_timezone | The user's timezone in the format. If unavailable, defaults to the project's timezone. | America/Toronto |
Project timezone can be set in project/behaviour settings:
We've added Deepgram Flux, their ASR newest model built specifically for Voice AI.
Flux is the first conversational speech recognition model built specifically for voice agents. Unlike traditional STT that just transcribes words, Flux understands conversational flow and automatically handles turn-taking.
Flux tackles the most critical challenges for voice agents today: knowing when to listen, when to think, and when to speak. The model features first-of-its-kind model-integrated end-of-turn detection, configurable turn-taking dynamics, and ultra-low latency optimized for voice agent pipelines, all with Nova-3 level accuracy.
Flux is Perfect for: turn-based voice agents, customer service bots, phone assistants, and real-time conversation tools.
Key Benefits:
Converts text to speech in real time and keeps the spoken audio perfectly aligned with the displayed text. This ensures call transcripts are an accurate, word-for-word representation of what was actually said.

This setting lets you define how long a conversation can stay inactive before the transcript automatically ends.
This is different from session timeout — the session stays open, but the transcript closes after the set inactivity period, enabling more accurate reporting and evaluations.
Important: ending the transcript does not end the user’s ability to re-engage. If the user responds again, a new transcript will begin within the same session.
We’ve added a new Priority Processing setting for OAI-supported models. When enabled, your requests will be given higher processing priority for faster response times and reduced latency. Note: this will consume more credits.

Supercharge your agents by connecting directly to MCP servers.
Bring in any tool, expand what your agents can do, and take your workflows to the next level.

You can now enable your agents to forward calls to a different number, SIP address, or extension.
This makes it easier to connect customers with the right destination without breaking the flow of the conversation.
