Copied


Startup Revolutionizes Retrieval-Augmented Generation for Enterprises with RAG 2.0

Joerg Hiller   Aug 30, 2024 08:36 0 Min Read


Contextual AI, a Silicon Valley-based startup, has introduced a groundbreaking platform called RAG 2.0, which promises to revolutionize retrieval-augmented generation (RAG) for enterprises. According to the NVIDIA Blog, RAG 2.0 achieves approximately 10x better parameter accuracy and performance compared to competing offerings.

Background and Development

Douwe Kiela, CEO of Contextual AI, has been an influential figure in the field of large language models (LLMs). Inspired by seminal papers from Google and OpenAI, Kiela and his team recognized early on the limitations of LLMs in dealing with real-time data. This understanding led to the development of the first RAG architecture in 2020.

RAG is designed to continuously update foundation models with new, relevant information. This approach addresses the data freshness issues inherent in LLMs, making them more useful for enterprise applications. Kiela's team realized that without efficient and cost-effective access to real-time data, even the most sophisticated LLMs would fall short in delivering value to enterprises.

RAG 2.0: The Next Evolution

Contextual AI's latest offering, RAG 2.0, builds upon the original architecture to deliver enhanced performance and accuracy. The platform integrates real-time data retrieval with LLMs, enabling a 70-billion-parameter model to run on infrastructure designed for just 7 billion parameters without compromising accuracy. This optimization opens up new possibilities for edge use cases, where smaller, more efficient computing resources are essential.

“When ChatGPT was released, it exposed the limitations of existing LLMs,” explained Kiela. “We knew that RAG was the solution to many of these problems, and we were confident we could improve upon our initial design.”

Integrated Retrievers and Language Models

One of the key innovations in RAG 2.0 is the close integration of its retriever architecture with the LLM. The retriever processes user queries, identifies relevant data sources, and feeds this information back to the LLM, which then generates a response. This integrated approach ensures higher precision and response quality, reducing the likelihood of “hallucinated” data.

Contextual AI differentiates itself by refining its retrievers through back propagation, aligning both retriever and generator components. This unification allows for synchronized adjustments, leading to significant gains in performance and accuracy.

Tackling Complex Use Cases

RAG 2.0 is designed to be LLM-agnostic, compatible with various open-source models like Mistral and Llama. The platform leverages NVIDIA’s Megatron LM and Tensor Core GPUs to optimize its retrievers. Contextual AI employs a “mixture of retrievers” approach to handle data in various formats, such as text, video, and PDF.

This hybrid method involves deploying different types of RAGs and a neural reranking algorithm to prioritize the most relevant information. This approach ensures that the LLM receives the best possible data to generate accurate responses.

“Our hybrid retrieval strategy maximizes performance by leveraging the strengths of different RAG types,” Kiela said. “This flexibility allows us to tailor solutions to specific use cases and data formats.”

The optimized architecture of RAG 2.0 reduces latency and lowers compute demands, making it suitable for a wide range of industries, from fintech and manufacturing to medical devices and robotics. The platform can be deployed in the cloud, on-premises, or in fully disconnected environments, offering versatility to meet diverse enterprise needs.

“We are focused on solving the most challenging use cases,” Kiela added. “Our aim is to enhance high-value, knowledge-intensive roles, enabling companies to save money and boost productivity.”


Read More
The Hong Kong Monetary Authority has issued a warning about a fraudulent website posing as OCBC Bank (Hong Kong) Limited, urging public vigilance.
BitMEX has changed the Mark Method for NILUSDTH25 and REDUSDTZ25 to Fair Price marking, effective March 25, 2025, enhancing price accuracy.
BitMEX introduces NILUSDT perpetual swaps, offering traders up to 50x leverage. This new listing enhances trading options on the platform.
BitMEX announces the introduction of NILUSDT perpetual swap listing, offering traders up to 50x leverage. The NIL token will be available for trading starting March 25, 2024.
Cronos (CRO) Labs has appointed Mirko Zhao as its new leader, succeeding Ken Timsit. Zhao aims to enhance the blockchain’s growth and community engagement.
Cronos (CRO) Labs announces Mirko Zhao as the new Head of Product and Engineering, succeeding Ken Timsit, to lead the blockchain ecosystem's innovative growth.
Filecoin (FIL) introduces ProPGF, an on-chain funding program aimed at supporting public goods development within its ecosystem, enhancing transparency and community involvement.
Linea is set to unveil a significant announcement during a livestream event, as the company encourages resilience amidst current market challenges.