Assessment of fine-tuned large language models for real-world chemistry and material science applications
- Ecole Polytechnique Federale Lausanne (EPFL), Sion (Switzerland)
- Ecole Polytechnique Federale Lausanne (EPFL), Sion (Switzerland); Consejo Superior de Investigaciones Cientificas (CSIC), Oviedo (Spain). Instituto de Ciencia y TecnologÍa del Carbono (INCAR)
- Ecole Polytechnique Federale Lausanne (EPFL), Sion (Switzerland); Friedrich Schiller University, Jena (Germany); Helmholtz Institute for Polymers in Energy Applications Jena (HIPOLE Jena) (Germany)
- University of Cambridge (United Kingdom)
- Technical University of Denmark, Lyngby (Denmark); University of Oxford (United Kingdom)
- University of Chicago, IL (United States); Argonne National Laboratory (ANL), Argonne, IL (United States)
- Politecnico di Torino (Italy)
- Ecole Polytechnique Federale Lausanne (EPFL) (Switzerland)
- Koc University, Istanbul (Turkey)
- Heriot-Watt University, Edinburgh (United Kingdom)
- BIGCHEM GmbH, Unterschleißheim (Germany)
- University of Cambridge (United Kingdom); National Institutes of Health (NIH), Bethesda, MD (United States)
- Helmholtz Zentrum Hereon, Geesthacht (Germany)
- Monash University, Clayton, VIC (Australia)
- University of Waterloo, ON (Canada)
- University of Toronto, ON (Canada)
- University of Pisa (Italy)
- Consejo Superior de Investigaciones Cientificas (CSIC), Oviedo (Spain). Instituto de Ciencia y TecnologÍa del Carbono (INCAR)
- University of Mohaghegh Ardabili (Iran)
- University of Tehran (Iran)
- University of Chicago, IL (United States)
- Massachusetts Inst. of Technology (MIT), Cambridge, MA (United States); University of Notre Dame, IN (United States)
- Technical University of Vienna (Austria)
- BIGCHEM GmbH, Unterschleißheim (Germany); Helmholtz Munich - Deutsches Forschungszentrum für Gesundheit und Umwelt (GmbH), Neuherberg (Germany)
- University of Notre Dame, IN (United States)
- Ecole Polytechnique Federale Lausanne (EPFL), Sion (Switzerland); Technical University of Denmark, Lyngby (Denmark)
The current generation of large language models (LLMs) has limited chemical knowledge. Recently, it has been shown that these LLMs can learn and predict chemical properties through fine-tuning. Using natural language to train machine learning models opens doors to a wider chemical audience, as field-specific featurization techniques can be omitted. In this work, we explore the potential and limitations of this approach. We studied the performance of fine-tuning three open-source LLMs (GPT-J-6B, Llama-3.1-8B, and Mistral-7B) for a range of different chemical questions. We benchmark their performances against “traditional” machine learning models and find that, in most cases, the fine-tuning approach is superior for a simple classification problem. Depending on the size of the dataset and the type of questions, we also successfully address more sophisticated problems. The most important conclusions of this work are that, for all datasets considered, their conversion into an LLM fine-tuning training set is straightforward and that fine-tuning with even relatively small datasets leads to predictive models. These results suggest that the systematic use of LLMs to guide experiments and simulations will be a powerful technique in any research study, significantly reducing unnecessary experiments or computations.
- Research Organization:
- Argonne National Laboratory (ANL), Argonne, IL (United States)
- Sponsoring Organization:
- USDOE Office of Science (SC)
- Grant/Contract Number:
- AC02-06CH11357
- OSTI ID:
- 2586547
- Journal Information:
- Chemical Science, Journal Name: Chemical Science Journal Issue: 2 Vol. 16; ISSN 2041-6539; ISSN 2041-6520
- Publisher:
- Royal Society of Chemistry (RSC)Copyright Statement
- Country of Publication:
- United States
- Language:
- English
Similar Records
CACTUS: Chemistry Agent Connecting Tool Usage to Science