Close Menu
  • AI
  • Content Creation
  • Tech
  • Robotics
AI-trends.todayAI-trends.today
  • AI
  • Content Creation
  • Tech
  • Robotics
Trending
  • Cursor Releases TypeScript-based SDKs for Building Coding Agents with Sandboxed Cloud Virtual Machines, Subagents Hooks and Token Based Pricing
  • The smol Audio Notebook: An Adaptive Collection of Notebooks for Whisper, Parakeet Voxtral Granite Speech and Audio Flamingo 3.
  • Elon Musk squeezed OpenAI, and they’re ‘going to want to kill me’
  • Taylor Swift is attempting to trademark her image. TikTok Deepfake Advertisements Show Why
  • Qwen Team Releases FlashQLA: a High-Performance Linear Attention Kernel Library That Achieves Up to 3× Speedup on NVIDIA Hopper GPUs
  • Emergency First Responders Say Waymos Are Getting Worse
  • Google One, YouTube and YouTube One drive Google 25M subscribers in Q1
  • The Top 10 Compression techniques for LLM inference using KV cache: Reduced memory overhead across evictions, low-rank methods, and quantization
AI-trends.todayAI-trends.today
Home»Tech»Google DeepMind finds a fundamental bug in RAG: embedding limits break retrieval at scale

Google DeepMind finds a fundamental bug in RAG: embedding limits break retrieval at scale

Tech By Gavin Wallace04/09/20254 Mins Read
Facebook Twitter LinkedIn Email
Step-by-Step Guide to Creating Synthetic Data Using the Synthetic Data
Step-by-Step Guide to Creating Synthetic Data Using the Synthetic Data
Share
Facebook Twitter LinkedIn Email

The RAG system relies on dense embeddings that translate documents and questions into fixed dimensional vector space. This approach is the standard for most AI applications. However, recent research by Google DeepMind team shows that this method may not be optimal. fundamental architectural limitation It is not possible to solve the problem by simply using larger models and better training.

What is the theoretical limit of embedding dimensions?

It is important to consider the ability of embedded embeddings with fixed sizes to represent a given object. A dimension embedding D Once the database reaches a certain size, it is impossible to represent every possible combination of documents. The results of communication complexity theory and sign-rank theories support this conclusion.

  • The retrieval of embeddings up to 512 pixels is around 500K documents.
  • Limits for 1024 dimensions extend to approximately 4 million documents.
  • There is a theoretical limit of 4096 pixels. 250 million documents.

This is a best estimate based on the data available. Free embedding OptimizationVectors can be optimized directly against the test labels. Even embeddings that are language constrained in real-world situations fail earlier.

https://arxiv.org/pdf/2508.21038

What Problems Does LIMIT Benchmark Show?

Google DeepMind Team developed LIMIT to empirically test the limitation. LIMIT is a benchmark dataset that was designed for embedders. LIMIT is available in two versions:

  • LIMIT FULL (50K Documents) Even strong embedders can collapse in this setup. Recall@100 is often affected. Minimum 20%.
  • LIMIT small: The models are too small to do the job, despite their simplicity. Although performance is variable, it remains unreliable:
    • Promptriever Llama3 8B: 54.3% recall@2 (4096d)
    • GritLM7B 38.4% recall@2 (4096d)
    • E5-Mistral 7B: 29.5% recall@2 (4096d)
    • Gemini Embed: 33.7% recall@2 (3072d)

No embedder can reach full recall with only 46 documents. This shows that it is not the dataset size but rather the architecture of the embedding vectors themselves which limits the embedder.

In contrast, BM25This limitation does not affect a sparse, classical lexical model. They can capture more combinations than dense embeddings because they operate on unbounded spaces.

https://arxiv.org/pdf/2508.21038

What is the significance of this for RAG

Current RAG implementations assume that embedded embeddings will scale infinitely as more data is added. Google DeepMind researchers explain why this is not true: Inherently, the size of an embedded object limits retrieval ability. It affects

  • Find out more about the search engines that are available Handling millions of documents
  • Agentic systems These queries are complex and require logical reasoning.
  • Instruction-following retrieval tasksThe relevance of queries is determined dynamically.

Even advanced benchmarks such as MTEB do not capture these limitations, because they test a limited part/section/subset of query-document combination.

What Alternatives Exist to the Single-Vector Embeddings Method?

According to the research team, scalable retrieval requires moving beyond single vector embeddings.

  • Cross-EncodersYou can achieve perfect recall of LIMIT queries by scoring directly query-document pair, at the expense of a high latency for inference.
  • Multi-Vector Models (e.g., ColBERT)The retrieval is more expressive by using multiple vectors for each sequence. This improves performance when performing LIMIT tasks.
  • Sparse Models (BM25, TF-IDF, neural sparse retrievers)The search engine is better at scaling in high-dimensional searches, but it lacks semantic generalization.

It is important to note that Architectural innovation is neededNot just larger embedders

The Key Takeaway:

Analysis by the research team shows that dense embeddedings are still bound to a The mathematical limitThey cannot capture all relevant combinations when corpus size exceeds limits related to embedding dimensionity. The LIMIT benchmark shows this failure in concrete terms:

  • The following are some of the ways to get in touch with us. LIMIT FULL (50K documents) Recall@100 falls below 20%
  • The following are some of the ways to get in touch with us. LIMIT small (46) docs: The best of the models can only recall 54% @2.

The use of classical techniques, such as BM25 and newer architectures, like cross-encoders and multi-vector retrievers, is still essential when building retrieval engines.


Click here to find out more PAPER here. Please feel free to browse our GitHub Page for Tutorials, Codes and Notebooks. Also, feel free to follow us on Twitter Join our Facebook group! 100k+ ML SubReddit Subscribe Now our Newsletter.


Asif Razzaq serves as the CEO at Marktechpost Media Inc. As an entrepreneur, Asif has a passion for harnessing Artificial Intelligence to benefit society. Marktechpost is his latest venture, a media platform that focuses on Artificial Intelligence. It is known for providing in-depth news coverage about machine learning, deep learning, and other topics. The content is technically accurate and easy to understand by an audience of all backgrounds. This platform has over 2,000,000 monthly views which shows its popularity.

deepmind Google
Share. Facebook Twitter LinkedIn Email
Avatar
Gavin Wallace

Related Posts

Cursor Releases TypeScript-based SDKs for Building Coding Agents with Sandboxed Cloud Virtual Machines, Subagents Hooks and Token Based Pricing

30/04/2026

The smol Audio Notebook: An Adaptive Collection of Notebooks for Whisper, Parakeet Voxtral Granite Speech and Audio Flamingo 3.

30/04/2026

Qwen Team Releases FlashQLA: a High-Performance Linear Attention Kernel Library That Achieves Up to 3× Speedup on NVIDIA Hopper GPUs

29/04/2026

The Top 10 Compression techniques for LLM inference using KV cache: Reduced memory overhead across evictions, low-rank methods, and quantization

29/04/2026
Top News

The AI Slur ‘Clanker’ Has Become a Cover for Racist TikTok Skits

The new Bernie Sanders AI Safety Bill would halt data center construction

Anthropic Claude Cowork is an AI agent that actually works.

Microsoft, Meta, and Google Triple Up on AI Spending

AI Drafting My Stories? Is AI Drafting My Stories?

Load More
AI-Trends.Today

Your daily source of AI news and trends. Stay up to date with everything AI and automation!

X (Twitter) Instagram
Top Insights

LlamaIndex releases LiteParse, a CLI- and TypeScript native library for spatial PDF parsing within AI Agent Workflows

20/03/2026

The First AI Bust: How will we refer to it?

18/11/2025
Latest News

Cursor Releases TypeScript-based SDKs for Building Coding Agents with Sandboxed Cloud Virtual Machines, Subagents Hooks and Token Based Pricing

30/04/2026

The smol Audio Notebook: An Adaptive Collection of Notebooks for Whisper, Parakeet Voxtral Granite Speech and Audio Flamingo 3.

30/04/2026
X (Twitter) Instagram
  • Privacy Policy
  • Contact Us
  • Terms and Conditions
© 2026 AI-Trends.Today

Type above and press Enter to search. Press Esc to cancel.