RAG 2.0: Retrieval Augmented Language Models

Vishal Rajput
AIGuys
Published in
12 min readApr 16, 2024

--

Language models have led to amazing progress, but they also have important shortcomings. One solution for many of these shortcomings is retrieval augmentation. There have been plenty of articles written about Retrieval Augmented Generation (RAG) pipelines, which as a technology is quite cool. But today, we are taking it one step further and truly exploring what’s next for the technology of RAG. What if we can create models with trainable retrievers, or in short, the entire RAG pipeline is customizable like fine-tuning an LLM? The problem with current RAGs is that they are not fully in tune within it’s submodules, it’s like a Frankenstein monster, it somehow works, but the parts are not in harmony and perform quite suboptimally together. So, to tackle all the issues with Frankenstein RAG, let’s take a deep dive into RAG 2.0.

Table of Contents

  • What Are RAGs?
  • What RAG 2.0 Achieves?
  • How Does RAG Solve Issues of Intelligence?
  • Better Retrieval Strategies
  • SOTA Retrieval Algorithms
  • Contextualizing the Retriever for the Generator
  • Combined Contextualized Retriever and Generator
  • SOTA Contextualizaton
  • Conclusion

Please take a moment to share the article and follow me, only if you see the hard work. It takes weeks to write such articles, some love would be highly appreciated.

--

--

AIGuys
AIGuys

Published in AIGuys

Deflating the AI hype and bringing real research and insights on the latest SOTA AI research papers. We at AIGuys believe in quality over quantity and are always looking to create more nuanced and detail oriented content.