Friday, May 16, 2025
News PouroverAI
Visit PourOver.AI
No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
News PouroverAI
No Result
View All Result

DRAGIN: A Novel Machine Learning Framework for Dynamic Retrieval Augmentation in Large Language Models and Outperforming Conventional Methods

April 2, 2024
in AI Technology
Reading Time: 4 mins read
0 0
A A
0
Share on FacebookShare on Twitter


The Dynamic Retrieval Augmented Generation (RAG) paradigm aims to improve the performance of LLMs by determining when to retrieve external information and what to retrieve during text generation. Current methods often rely on static rules to decide when to recover and limit retrieval to recent sentences or tokens, which may not capture the full context. This approach risks introducing irrelevant data and increasing computation costs unnecessarily. Effective strategies for optimal retrieval timing and crafting relevant queries are essential to enhance LLM generation while mitigating these challenges.

Researchers from Tsinghua University and the Beijing Institute of Technology have developed DRAGIN, a Dynamic Retrieval Augmented Generation framework tailored to LLMs. DRAGIN dynamically determines when and what to retrieve based on real-time information needs during text generation. It introduces RIND for timing retrieval, considering LLM uncertainty and token importance, and QFS for query formulation, leveraging self-attention across the context. DRAGIN outperforms existing methods across four knowledge-intensive datasets without requiring additional training or prompt engineering.

Single-round retrieval-augmented methods enhance LLMs by incorporating external knowledge retrieved using the initial input as a query. Previous studies extensively explore this approach, such as REPLUG, which uses LLMs to generate training data for retrieval models, and UniWeb, which self-assesses the need for retrieval. However, multi-round retrieval becomes essential for complex tasks requiring extensive external knowledge. Methods like RETRO and IC-RALM trigger retrieval at fixed intervals, but FLARE innovatively triggers retrieval upon encountering uncertain tokens, improving retrieval relevance by considering the LLM’s real-time information needs.

The DRAGIN framework comprises two key components: Real-time Information Needs Detection (RIND) and Query Formulation based on Self-attention (QFS). RIND evaluates tokens’ uncertainty, semantic significance, and impact on subsequent context to trigger retrieval dynamically. QFS formulates queries by analyzing the LLM’s self-attention mechanism, prioritizing tokens based on their relevance to the current context. After retrieval, the framework truncates the output at the identified token, integrates retrieved knowledge using a designed prompt template, and generates resumes. This iterative process ensures the LLM seamlessly incorporates relevant external information, enhancing its output’s quality and relevance.

The performance of DRAGIN was evaluated against various baseline methods across four datasets, and the experimental results were compared. DRAGIN consistently outperformed other methods, demonstrating its effectiveness in enhancing LLMs. Efficiency analysis revealed that DRAGIN required fewer retrieval calls than some baselines, indicating its efficiency. Timing analysis showed DRAGIN’s superiority in determining optimal retrieval moments based on real-time information needs. DRAGIN’s query formulation method outperformed other frameworks, emphasizing its ability to select tokens representing LLM’s information needs accurately. Furthermore, BM25 outperformed SGPT as a retrieval method, suggesting the continued effectiveness of lexicon-based approaches in RAG tasks.

In conclusion, DRAGIN is a framework addressing limitations in dynamic RAG methods for LLMs. DRAGIN improves retrieval activation timing with RIND and enhances query formulation precision using QFS, leading to better performance on knowledge-intensive tasks. Despite its reliance on Transformer-based LLMs’ self-attention mechanism, DRAGIN demonstrates effectiveness. Future work aims to overcome limitations related to self-attention accessibility. DRAGIN integrates external knowledge by truncating LLM output for retrieval augmentation and incorporating retrieved information using a prompt template. The impact of query formulation techniques is evaluated, with DRAGIN surpassing other methods like FLARE, FL-RAG, and FS-RAG.

Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

If you like our work, you will love our newsletter.

Don’t Forget to join our 39k+ ML SubReddit

Sana Hassan, a consulting intern at Marktechpost and dual-degree student at IIT Madras, is passionate about applying technology and AI to address real-world challenges. With a keen interest in solving practical problems, he brings a fresh perspective to the intersection of AI and real-life solutions.

🐝 Join the Fastest Growing AI Research Newsletter Read by Researchers from Google + NVIDIA + Meta + Stanford + MIT + Microsoft and many others…



Source link

Tags: AugmentationConventionalDRAGINdynamicFrameworklanguageLargeLearningMachineMethodsmodelsOutperformingretrieval
Previous Post

Singapore Expands Crypto Regulation, Introduces Stricter User Protection Requirements

Next Post

The winning combination for real-time insights: Messaging and event-driven architecture

Related Posts

How insurance companies can use synthetic data to fight bias
AI Technology

How insurance companies can use synthetic data to fight bias

June 10, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset
AI Technology

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
Decoding Decoder-Only Transformers: Insights from Google DeepMind’s Paper
AI Technology

Decoding Decoder-Only Transformers: Insights from Google DeepMind’s Paper

June 9, 2024
How Game Theory Can Make AI More Reliable
AI Technology

How Game Theory Can Make AI More Reliable

June 9, 2024
Buffer of Thoughts (BoT): A Novel Thought-Augmented Reasoning AI Approach for Enhancing Accuracy, Efficiency, and Robustness of LLMs
AI Technology

Buffer of Thoughts (BoT): A Novel Thought-Augmented Reasoning AI Approach for Enhancing Accuracy, Efficiency, and Robustness of LLMs

June 9, 2024
Deciphering Doubt: Navigating Uncertainty in LLM Responses
AI Technology

Deciphering Doubt: Navigating Uncertainty in LLM Responses

June 9, 2024
Next Post
The winning combination for real-time insights: Messaging and event-driven architecture

The winning combination for real-time insights: Messaging and event-driven architecture

Anthropic Explores Many-Shot Jailbreaking: Exposing AI’s Newest Weak Spot

Anthropic Explores Many-Shot Jailbreaking: Exposing AI's Newest Weak Spot

Influencer marketing in India to become more lucrative, may reach Rs 3,375 crore by 2026: Report

Influencer marketing in India to become more lucrative, may reach Rs 3,375 crore by 2026: Report

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

  • Trending
  • Comments
  • Latest
Is C.AI Down? Here Is What To Do Now

Is C.AI Down? Here Is What To Do Now

January 10, 2024
23 Plagiarism Facts and Statistics to Analyze Latest Trends

23 Plagiarism Facts and Statistics to Analyze Latest Trends

June 4, 2024
Porfo: Revolutionizing the Crypto Wallet Landscape

Porfo: Revolutionizing the Crypto Wallet Landscape

October 9, 2023
A Complete Guide to BERT with Code | by Bradney Smith | May, 2024

A Complete Guide to BERT with Code | by Bradney Smith | May, 2024

May 19, 2024
How To Build A Quiz App With JavaScript for Beginners

How To Build A Quiz App With JavaScript for Beginners

February 22, 2024
Saginaw HMI Enclosures and Suspension Arm Systems from AutomationDirect – Library.Automationdirect.com

Saginaw HMI Enclosures and Suspension Arm Systems from AutomationDirect – Library.Automationdirect.com

December 6, 2023
Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

June 10, 2024
AI Compared: Which Assistant Is the Best?

AI Compared: Which Assistant Is the Best?

June 10, 2024
How insurance companies can use synthetic data to fight bias

How insurance companies can use synthetic data to fight bias

June 10, 2024
5 SLA metrics you should be monitoring

5 SLA metrics you should be monitoring

June 10, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

June 10, 2024
Facebook Twitter LinkedIn Pinterest RSS
News PouroverAI

The latest news and updates about the AI Technology and Latest Tech Updates around the world... PouroverAI keeps you in the loop.

CATEGORIES

  • AI Technology
  • Automation
  • Blockchain
  • Business
  • Cloud & Programming
  • Data Science & ML
  • Digital Marketing
  • Front-Tech
  • Uncategorized

SITEMAP

  • Disclaimer
  • Privacy Policy
  • DMCA
  • Cookie Privacy Policy
  • Terms and Conditions
  • Contact us

Copyright © 2023 PouroverAI News.
PouroverAI News

No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing

Copyright © 2023 PouroverAI News.
PouroverAI News

Welcome Back!

Login to your account below

Forgotten Password? Sign Up

Create New Account!

Fill the forms bellow to register

All fields are required. Log In

Retrieve your password

Please enter your username or email address to reset your password.

Log In