Sunday, June 8, 2025
News PouroverAI
Visit PourOver.AI
No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
News PouroverAI
No Result
View All Result

Large language models use a surprisingly simple mechanism to retrieve some stored knowledge | MIT News

March 25, 2024
in AI Technology
Reading Time: 4 mins read
0 0
A A
0
Share on FacebookShare on Twitter



Large language models, such as those powering popular artificial intelligence chatbots like ChatGPT, are extremely intricate. Despite their widespread use in various applications, such as customer support, code generation, and language translation, the inner workings of these models remain somewhat mysterious.

To gain a better understanding of how these massive machine-learning models retrieve stored knowledge, researchers at MIT and other institutions delved into the mechanisms involved.

Their findings were surprising: Large language models (LLMs) often employ a simple linear function to extract and interpret stored facts. Additionally, the model uses the same decoding function for similar types of facts. Linear functions, which involve only two variables and no exponents, capture the direct, linear relationship between variables.

The researchers demonstrated that by pinpointing linear functions for different facts, they can investigate the model to uncover its knowledge about new subjects and where that knowledge is stored within the model.

Using a novel technique they developed to estimate these straightforward functions, the researchers discovered that even when a model provides an incorrect response to a prompt, it often retains the correct information. In the future, this approach could be used to identify and rectify inaccuracies within the model, potentially reducing instances of incorrect or nonsensical responses.

“Although these models are complex, nonlinear functions trained on vast amounts of data and difficult to comprehend, there are instances of remarkably simple mechanisms at work within them. This is one such example,” says Evan Hernandez, an electrical engineering and computer science (EECS) graduate student and co-lead author of a paper detailing these findings.

Hernandez collaborated with co-lead author Arnab Sharma, a computer science graduate student at Northeastern University; their advisor, Jacob Andreas, an EECS associate professor and member of the Computer Science and Artificial Intelligence Laboratory (CSAIL); senior author David Bau, an assistant professor of computer science at Northeastern; as well as other researchers from MIT, Harvard University, and the Israeli Institute of Technology. The research will be presented at the International Conference on Learning Representations.

Uncovering Facts

Most large language models, also known as transformer models, are neural networks. These networks, inspired by the human brain, consist of billions of interconnected nodes or neurons organized into multiple layers for encoding and processing data.

The knowledge stored in a transformer can often be represented as relations linking subjects and objects. For example, the relation “Miles Davis plays the trumpet” connects the subject, Miles Davis, with the object, trumpet.

As a transformer accumulates more knowledge, it stores additional facts about a particular subject across various layers. When a user inquires about that subject, the model must decode the most relevant fact to respond to the query.

For example, if someone prompts a transformer with “Miles Davis plays the. . .”, the model should respond with “trumpet” rather than “Illinois” (the state where Miles Davis was born).

“Somewhere in the network’s computation, there must be a mechanism that retrieves the fact that Miles Davis plays the trumpet and then utilizes that information to generate the next word. We aimed to understand this mechanism,” Hernandez explains.

The researchers conducted a series of experiments to investigate LLMs and found that, despite their complexity, the models utilize a simple linear function to decode relational information. Each function is tailored to the type of fact being retrieved.

For instance, the transformer employs one decoding function when outputting the instrument a person plays and a different function when outputting the state where a person was born.

The researchers devised a method to estimate these basic functions and computed functions for 47 different relations, such as “capital city of a country” and “lead singer of a band.”

While there are countless possible relations, the researchers focused on this specific subset because they are representative of the kinds of facts that can be expressed in this manner.

They tested each function by altering the subject to determine if it could retrieve the correct object information. For example, the function for “capital city of a country” should retrieve Oslo if the subject is Norway and London if the subject is England.

The functions successfully retrieved the correct information over 60% of the time, indicating that some information in a transformer is encoded and retrieved using this method.

“However, not all information is linearly encoded. For certain facts, even though the model is aware of them and generates text consistent with these facts, we are unable to identify linear functions for them. This suggests that the model employs a more sophisticated method to store that information,” Hernandez notes.

Visualizing a Model’s Knowledge

The researchers also utilized the functions to determine what a model understands to be true about different subjects.

In one experiment, they initiated with the prompt “Bill Bradley was a” and utilized the decoding functions for “plays sports” and “attended university” to ascertain if the model recognizes that Sen. Bradley was a basketball player who attended Princeton.

“We can demonstrate that, although the model may emphasize different information in its text production, it does encode all of that information,” Hernandez states.

They leveraged this probing technique to develop an “attribute lens,” a grid that visualizes where specific information regarding a particular relation is stored within the transformer’s numerous layers.

Attribute lenses can be automatically generated, offering a streamlined approach to assist researchers in gaining a deeper understanding of a model. This visualization tool could empower scientists and engineers to correct stored knowledge and prevent an AI chatbot from disseminating false information.

In the future, Hernandez and his collaborators aim to further explore cases where facts are not stored linearly. They also hope to conduct experiments with larger models and evaluate the accuracy of linear decoding functions.

“This work is exciting as it uncovers a missing piece in our comprehension of how large language models recall factual knowledge during inference. Prior research indicated that LLMs construct information-rich representations of given subjects, from which specific attributes are extracted during inference. This study reveals that the complex nonlinear computation of LLMs for attribute extraction can be effectively approximated with a simple linear function,” says Mor Geva Pipek, an assistant professor in the School of Computer Science at Tel Aviv University, who was not involved in this study.

This research received partial support from Open Philanthropy, the Israeli Science Foundation, and an Azrieli Foundation Early Career Faculty Fellowship.



Source link

Tags: AI InterpretabilitychatgptEvan HernandezJacob AndreasKnowledgelanguageLargelarge language modelsLLMsMechanismMITmodelsNewsretrieveSimplestoredsurprisingly
Previous Post

Analysis-South Korea’s push to make its markets global dogged by FX history By Reuters

Next Post

Trust Fintech IPO opens tomorrow: Check issue size, price band, GMP and other details

Related Posts

How insurance companies can use synthetic data to fight bias
AI Technology

How insurance companies can use synthetic data to fight bias

June 10, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset
AI Technology

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
How Game Theory Can Make AI More Reliable
AI Technology

How Game Theory Can Make AI More Reliable

June 9, 2024
Decoding Decoder-Only Transformers: Insights from Google DeepMind’s Paper
AI Technology

Decoding Decoder-Only Transformers: Insights from Google DeepMind’s Paper

June 9, 2024
Buffer of Thoughts (BoT): A Novel Thought-Augmented Reasoning AI Approach for Enhancing Accuracy, Efficiency, and Robustness of LLMs
AI Technology

Buffer of Thoughts (BoT): A Novel Thought-Augmented Reasoning AI Approach for Enhancing Accuracy, Efficiency, and Robustness of LLMs

June 9, 2024
Deciphering Doubt: Navigating Uncertainty in LLM Responses
AI Technology

Deciphering Doubt: Navigating Uncertainty in LLM Responses

June 9, 2024
Next Post
Trust Fintech IPO opens tomorrow: Check issue size, price band, GMP and other details

Trust Fintech IPO opens tomorrow: Check issue size, price band, GMP and other details

Sora: First Impressions

Sora: First Impressions

A Study by Google DeepMind on Evaluating Frontier Machine Learning Models for Dangerous Capabilities

A Study by Google DeepMind on Evaluating Frontier Machine Learning Models for Dangerous Capabilities

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

  • Trending
  • Comments
  • Latest
23 Plagiarism Facts and Statistics to Analyze Latest Trends

23 Plagiarism Facts and Statistics to Analyze Latest Trends

June 4, 2024
Accenture creates a regulatory document authoring solution using AWS generative AI services

Accenture creates a regulatory document authoring solution using AWS generative AI services

February 6, 2024
Managing PDFs in Node.js with pdf-lib

Managing PDFs in Node.js with pdf-lib

November 16, 2023
Graph neural networks in TensorFlow – Google Research Blog

Graph neural networks in TensorFlow – Google Research Blog

February 6, 2024
13 Best Books, Courses and Communities for Learning React — SitePoint

13 Best Books, Courses and Communities for Learning React — SitePoint

February 4, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

June 10, 2024
AI Compared: Which Assistant Is the Best?

AI Compared: Which Assistant Is the Best?

June 10, 2024
How insurance companies can use synthetic data to fight bias

How insurance companies can use synthetic data to fight bias

June 10, 2024
5 SLA metrics you should be monitoring

5 SLA metrics you should be monitoring

June 10, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

June 10, 2024
Facebook Twitter LinkedIn Pinterest RSS
News PouroverAI

The latest news and updates about the AI Technology and Latest Tech Updates around the world... PouroverAI keeps you in the loop.

CATEGORIES

  • AI Technology
  • Automation
  • Blockchain
  • Business
  • Cloud & Programming
  • Data Science & ML
  • Digital Marketing
  • Front-Tech
  • Uncategorized

SITEMAP

  • Disclaimer
  • Privacy Policy
  • DMCA
  • Cookie Privacy Policy
  • Terms and Conditions
  • Contact us

Copyright © 2023 PouroverAI News.
PouroverAI News

No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing

Copyright © 2023 PouroverAI News.
PouroverAI News

Welcome Back!

Login to your account below

Forgotten Password? Sign Up

Create New Account!

Fill the forms bellow to register

All fields are required. Log In

Retrieve your password

Please enter your username or email address to reset your password.

Log In