TGS-RAG Claims 80% Token Reduction vs LightRAG at Higher Accuracy

Researchers at Southeast University Nanjing have published TGS-RAG (Text-Graph Synergy RAG), a hybrid retrieval framework claiming ~217M tokens consumed on HotpotQA versus ~757M for LightRAG and ~646M for GraphRAG — a ~70–80% token reduction — while also reporting the highest accuracy in the benchmark comparison. The core innovation: beam-search graph pruning is treated as "move to memory" rather than delete, allowing the vector retrieval channel to revive discarded graph nodes when textual evidence later implicates them. The mechanism avoids new database queries during the orphan-bridging step, which is how the token savings accumulate.

Why It Matters

A 3–4× cost reduction with accuracy parity or better is a Pareto improvement — the enterprise case for adopting TGS-RAG over LightRAG/GraphRAG is one-sided if the results replicate. The infrastructure footprint (pgvector, 0.6B Qwen embedding, single 32GB Linux box) also reads as a deliberate build-it-yourself signal.