MongoDB Sets a New Standard for Retrieval Accuracy with Voyage 4 Models for Production-Ready AI Applications
Tavily and TinyFish among customers using
To support developers moving AI applications into production,
"The biggest challenge customers face with AI isn't experimentation, it's operating reliably at scale," said
Transforming data into AI intelligence
As projects move into production, many organizations are discovering that their existing data stacks were never designed to support context-aware, retrieval-intensive workloads at scale. Developers are left managing fragmented combinations of operational databases, vector stores, and model APIs, which introduces complexity, latency, and operational risk at the exact moment speed and reliability matter most. This fragmentation has become a primary barrier to AI innovations, translating into real customer-facing impact.
- State-of-the-art accuracy with models from Voyage AI: The general availability of the new Voyage 4 series continues giving developers high performing embedding models—which outperform Gemini and Cohere on the public RTEB leaderboard—for more accurate retrieval at lower cost. The Voyage 4 series includes the general-purpose voyage-4 embedding model, which strikes a balance between retrieval accuracy, cost, and latency, the flagship voyage-4-large model for the highest retrieval accuracy, voyage-4-lite for optimized latency and cost, and an open-weights voyage-4-nano for local development and testing, or on-device applications.
- Facilitated context extraction from video, images, and text: The general availability of the new voyage-multimodal-3.5 model expands support for interleaved text and images to now include video. Voyage AI's voyage-multimodal-3 was the first production-grade embedding model to handle interleaved text and images, voyage-multimodal-3.5 advances this unified processing approach, more effectively vectorizing multimodal data together to best capture key semantic meaning from tables, graphics, figures, slides, PDFs, and more. This helps developers eliminate the significant effort required for complex document parsing, which can reduce retrieval accuracy and lead to less trustworthy applications.
-
Automated Embedding for MongoDB Vector Search: Automatically generate and store high-fidelity embeddings using Voyage AI whenever data is inserted, updated, or queried. By handling embedding generation natively within the database,
MongoDB removes the need for separate embedding pipelines or external model services. Embeddings stay fresh as data changes, helping retrieval to remain accurate and AI applications to maintain reliable context. The result is a simpler architecture with fewer moving parts, making it easier for teams to build and run AI-enabled applications in production. Automated Embedding is available in public preview with support in our drivers (e.g. Javascript, Python, Java, etc) and AI Frameworks like LangChain and LangGraph (Python). Available today forMongoDB Community , and coming soon on MongoDB Atlas.
"We were looking for extremely accurate embedding models, and Voyage AI provided accuracy at scale," says
"Today, companies need to move extremely fast, and at very lean startups, you need to only focus on what you are building," said
For the first time, developers can build and run AI applications with operational data, semantic understanding, and retrieval in a single system.
To learn more about these new capabilities and to get started, please find the wrap blog here.
About
Headquartered in
Press Contact:
press@mongodb.com
View original content to download multimedia:https://www.prnewswire.com/news-releases/mongodb-sets-a-new-standard-for-retrieval-accuracy-with-voyage-4-models-for-production-ready-ai-applications-302662558.html
SOURCE