Standard ElevenLabs voice agents operate in a single modality — voice (and optionally text). The user speaks (or types), the agent processes the audio/text input, and responds by voice. Multimodal agents extend this by accepting file inputs alongside voice — the user can share an image or PDF document, and the agent processes the visual content as part of the conversation context.
The April 1, 2026 API update formalises multimodal support with a dedicated configuration field and client-side upload capability. This update positions ElevenLabs voice agents as a complete multimodal interface rather than a voice-only tool — an important differentiation as enterprise voice agent deployments increasingly encounter scenarios where users need to share visual information.
Configuration
Enable File Input in Agent Configuration
Add the file_input field to your agent’s ConversationConfig when creating or updating the agent. The file_input object contains: enabled (boolean — true to allow file uploads in conversations) and max_files_per_conversation (integer — the maximum number of files a user can attach in a single conversation session). Example API request body fragment: ‘conversation_config’: {‘file_input’: {‘enabled’: true, ‘max_files_per_conversation’: 3}}.
LLM Requirement: Vision-Capable Model
Multimodal file input requires a vision-capable LLM configured for the agent. Supported vision-capable LLMs available through ElevenAgents include: GPT-4o (OpenAI), Claude 3.5 Sonnet and Claude 3.5 Haiku (Anthropic), Gemini 1.5 Pro and Gemini 1.5 Flash (Google). If the agent’s LLM does not support vision input, enabling file_input in the configuration will not produce useful multimodal responses — the LLM will be unable to process the image content.
Client SDK Integration
For web applications, the @elevenlabs/client and @elevenlabs/react SDKs support the multimodal message WebSocket event type (added in the April 2026 SDK update — v2.41.0). The useConversationControls hook exposes sendMultimodalMessage, and the MultimodalMessageInput type is exported from @elevenlabs/client. In a React application, sending a file during a conversation: import the sendMultimodalMessage function from the conversation controls hook, create a MultimodalMessageInput object with the file data and MIME type, and call sendMultimodalMessage with the object. The file is transmitted to the agent as part of the conversation context.
Supported File Types
| File Type | MIME Type | Use Cases |
| Images (JPEG) | image/jpeg | Product photos, screenshots, error messages, product labels |
| Images (PNG) | image/png | Screenshots, diagrams, app UI errors, documents as images |
| Images (WebP) | image/webp | Web-captured images, optimised photo formats |
| Images (GIF) | image/gif | Animated screenshots (static frame analysed) |
| PDF documents | application/pdf | Manuals, contracts, invoices, forms, reports |
The LLM processes image files through its vision capability — the image is sent to the LLM alongside the conversation context and the LLM generates a response based on both the image content and the conversation history. PDF processing converts the PDF pages to images for vision processing — text extraction may not be as complete as RAG-based Knowledge Base document processing for text-heavy PDFs.
Use Cases
Visual Customer Support
The most immediately high-value multimodal agent use case is customer support where users share images of their issue. A user having trouble with a product setting can take a screenshot and share it directly in the voice conversation — the agent sees the screenshot and provides specific guidance based on what it sees rather than asking the user to describe it verbally. A user reporting a damaged product can share a photo — the agent documents the damage, initiates a return process, and provides next steps, all without requiring the user to navigate a separate form or portal.
Document Review by Voice
Users can share PDF documents — contracts, invoices, manuals, reports — and ask questions about them by voice. An agent configured as a financial services assistant can receive an uploaded invoice, read the amounts and vendor information, and answer questions about it verbally: ‘What is the total on this invoice?’ ‘Is this charge from this vendor consistent with our contract terms?’ Document review by voice makes information from documents accessible without requiring users to read and locate specific information themselves.
Product Identification and Recommendations
Retail and e-commerce voice agents can receive product images and make purchase recommendations, compatibility checks, or similar item suggestions based on what they see. A user photographing a furniture piece and asking ‘does this match anything in your catalogue?’ triggers visual similarity search and agent response based on the image. A user photographing a component and asking ‘is this compatible with my model X device?’ receives a specific compatibility answer based on visual identification.
Technical Troubleshooting
Technical support agents can receive screenshots of error messages, configuration screens, or device states and provide specific troubleshooting guidance. Rather than the agent asking ‘what does the error message say?’ — and the user reading it aloud imperfectly — the user shares a screenshot and the agent reads the exact error and responds specifically. This eliminates a major friction point in voice-based ElevenLabs multimodal agents technical support where precision matters.
Related: For the complete ElevenLabs Conversational AI setup guide, see our Conversational AI guide 2026
Server-Side File Injection
Before the April 2026 client-side file upload update, the existing conversation file uploads API (POST /v1/convai/conversations/{conversation_id}/files, added January 2026) allowed server-side injection of files into ongoing conversations. This approach is used when the server — not the user — provides the visual context: customer account information (a server-side image of the customer’s order, subscription status, or account history), personalised document context (a server-side PDF of the customer’s contract or agreement), or pre-loaded product information (images of the product the user called about, provided by the CRM integration before the agent’s first response).
Both approaches — client-side user uploads and server-side file injection — can be used in the same agent and the same conversation, enabling fully contextualised multimodal interactions where both the server and the user contribute visual context.
Three Insights Most Multimodal Agent Coverage Misses
1. Max Files Per Conversation Is a Conversation Design Decision, Not a Technical Limit
The max_files_per_conversation parameter appears to be a technical constraint — how many files can be uploaded. It is more accurately a conversation design decision. For a support agent handling a single issue per conversation, max 1-2 files is appropriate. For a document review agent where users may share multiple pages, max 5-10 files may be more suitable. For a product recommendation agent where users may share multiple product photos, max 3-5 is a reasonable limit. Setting this parameter thoughtfully — rather than leaving it at a default or setting it arbitrarily high — is part of designing the conversation experience.
2. PDF Multimodal Processing Is Different from Knowledge Base Document Processing
ElevenLabs offers two ways for agents to work with PDF content: Knowledge Base document upload (where the PDF is indexed for RAG retrieval before the conversation) and multimodal file input (where the PDF is shared by the user during the conversation). These serve different use cases: Knowledge Base is for your fixed reference content (product docs, policy documents, FAQs). Multimodal file input is for user-provided documents specific to their conversation (their invoice, their contract, their manual). The processing also differs — Knowledge Base RAG extracts and indexes text for semantic search. Multimodal converts PDF pages to images for vision model processing, which is less accurate for text-heavy PDFs but handles visual document layouts better.
3. Multimodal Agents Change the Voice-First Design Paradigm
Voice-first interface design has historically been constrained by the assumption that voice is the only input modality — every user interaction must be expressible in speech. Multimodal agents break this constraint. Complex visual information (error messages, product photos, document contents) that is impractical to describe verbally can now be shared as a file. This changes what is appropriate to design as a voice agent use case: complex technical support, document-centred workflows, and visually-driven product interactions that were previously unsuitable for voice-first design are now viable. The design question shifts from ‘can this be expressed in voice?’ to ‘which parts of this interaction are voice-natural and which benefit from visual input?’
Multimodal Agents in 2027
The multimodal agent capability will expand along two dimensions. Input modalities — the April 2026 launch supports images and PDFs. Future updates will likely add audio file input (users share a recording for the agent to transcribe and respond to), video file input (users share a short video for the agent to describe and respond to), and screen sharing (the agent sees the user’s screen in real time). Output modalities — current multimodal agents respond by voice and optionally text. Future development will likely extend to agents that generate images or diagrams in response to user queries — a truly bidirectional multimodal conversation interface.
Key Takeaways
- ElevenLabs multimodal agents (April 2026) allow users to share images and PDFs during voice conversations — the agent sees the file and responds by voice based on its contents.
- Configuration: add file_input field to ConversationConfig with enabled: true and max_files_per_conversation limit. Requires a vision-capable LLM (GPT-4o, Claude 3.5, Gemini 1.5 Pro).
- Client-side upload via @elevenlabs/client SDK. Server-side injection via POST /v1/convai/conversations/{id}/files. Both can be used in the same conversation.
- Top use cases: visual customer support, document review by voice, product identification, technical troubleshooting with screenshots.
- Multimodal agents expand voice-first design to cover interactions that previously required visual interfaces — complex technical support, document-centred workflows, visually-driven product assistance.
Conclusion
ElevenLabs multimodal agents address the most significant limitation of voice-only AI interfaces: the inability to share visual information without describing it in words. For customer service, technical support, document review, and product assistance — the highest-value enterprise voice agent use cases — visual context is often essential to providing accurate, specific responses. The April 2026 multimodal update makes this context available through a natural file-sharing interface that users already understand. For developers building voice agents in these domains, enabling multimodal support is a configuration addition of a few minutes that meaningfully expands the range of interactions the agent can handle effectively. Enable it, test with representative user scenarios, and ensure your LLM selection supports vision input — the rest is conversation design.
Frequently Asked Questions
What are ElevenLabs multimodal agents?
ElevenLabs voice agents configured to accept image and PDF file uploads from users during conversations. The agent sees the uploaded file and responds by voice based on its contents, enabling visual customer support, document review by voice, and image-based question answering.
What LLMs support multimodal input in ElevenLabs?
Vision-capable LLMs available through ElevenAgents include GPT-4o (OpenAI), Claude 3.5 Sonnet and Claude 3.5 Haiku (Anthropic), and Gemini 1.5 Pro and Gemini 1.5 Flash (Google). The agent’s LLM must support vision input for multimodal file uploads to produce useful responses.
When did ElevenLabs add multimodal support?
Client-side multimodal file upload in conversations was added in the April 1, 2026 API update. Server-side conversation file upload (POST /v1/convai/conversations/{id}/files) was added in January 2026. The April 2026 update adds user-facing file sharing in the conversation interface.
How many files can a user upload in one conversation?
The max_files_per_conversation parameter in the agent’s ConversationConfig controls this limit. Set it based on your use case — 1-2 for single-issue support, 5-10 for document review workflows. There is no published absolute maximum beyond what the LLM’s context window can process.
How is multimodal file input different from the Knowledge Base?
Knowledge Base loads your fixed reference documents for RAG retrieval — your product docs, policies, FAQs. Multimodal file input receives user-provided documents specific to their conversation — their invoice, their screenshot, their contract. They serve different purposes and can both be active in the same agent.
Methodology
Multimodal agent features from ElevenLabs API changelog (April 1, 2026) — file_input ConversationConfig field, conversation file upload support, and multimodal message WebSocket event. SDK support from ElevenLabs changelog (April 2026) — @elevenlabs/client multimodal message support. Server-side file upload from ElevenLabs changelog (January 2026). LLM vision support from ElevenLabs ElevenAgents official documentation. Use case context from ElevenLabs documentation and editorial team assessment. This article was drafted with AI assistance and reviewed by the editorial team at ElevenLabsMagazine.com.
References
ElevenLabs. (April 1, 2026). Changelog — conversation file uploads and multimodal support. https://elevenlabs.io/docs/changelog/2026/4/1
ElevenLabs. (2026). ElevenAgents SDK documentation. https://elevenlabs.io/docs/eleven-agents
ElevenLabs. (2026). @elevenlabs/client changelog. https://elevenlabs.io/docs/changelog
