Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Evaluating the Effectiveness of Retrieval-Augmented Large Language Models in Scientific Document Reasoning

Conference ·
Despite the dramatic progress in Large Language Model (LLM) development, LLMs often provide seemingly plausible but not factual information, often referred as hallucinations. Retrieval-augmented LLMs provide a non-parametric approach to solve these issues by retrieving relevant information from external data sources and augment the training process. These models helps to trace evidence from an externally provided knowledge base allowing the model predictions to be better interpreted and verified. In this work, we critically evaluate these models in their ability to perform in scientific document reasoning tasks. To this end, we tuned multiple such model variants with science-focused instructions and evaluated them on a scientific document reasoning benchmark for the usefulness of the retrieved document passages. Our findings suggest that models justify predictions in science tasks with fabricated evidence and leveraging scientific corpus as pretraining data does not alleviate the risk of evidence fabrication.
Research Organization:
Pacific Northwest National Laboratory (PNNL), Richland, WA (United States)
Sponsoring Organization:
USDOE
DOE Contract Number:
AC05-76RL01830
OSTI ID:
2484353
Report Number(s):
PNNL-SA-189029
Country of Publication:
United States
Language:
English

Similar Records

Opportunities for retrieval and tool augmented large language models in scientific facilities
Journal Article · Mon Nov 04 23:00:00 EST 2024 · npj Computational Materials · OSTI ID:2476271

LLaMP v0.1.0
Software · Wed Apr 23 20:00:00 EDT 2025 · OSTI ID:code-170337