LLM Grounding Analysis
LLM grounding vs. factual recall
Large language models (LLMs) can memorize and apply new information. However, it's unclear how they balance this new context with their pre-existing knowledge. This research analyzes how LLMs manage this conflict using a new counterfactual dataset.
The study investigates LLMs using Fakepedia, a dataset presenting contradictions between known facts and new information. Through Masked Grouped Causal Tracing (MGCT), the research deciphers LLMs' grounding mechanisms by contrasting neural activation patterns. Findings help understand the co-functioning of grounding with recall capabilities within LLMs.
active
—
entered showcase: 2024-05-03
—
entry updated: 2024-05-03
This project has not yet been evaluated by the C4DT Factory team.
We will be happy to evaluate it upon request.
Experiments
Python
Apache-2.0