The current generation of large language models (LLMs) has limited chemical knowledge. Recently, it has been shown that these LLMs can learn and predict chemical properties through fine-tuning. Using natural language to train machine learning models opens doors to a wider chemical audience, as field-specific featurization techniques can be omitted. In this work, we explore the potential and limitations of this approach. We studied the performance of fine-tuning three open-source LLMs (GPT-J-6B, Llama-3.1-8B, and Mistral-7B) for a range of different chemical questions. We benchmark their performances against “traditional” machine learning models and find that, in most cases, the fine-tuning approach is superior for a simple classification problem. Depending on the size of the dataset and the type of questions, we also successfully address more sophisticated problems. The most important conclusions of this work are that, for all datasets considered, their conversion into an LLM fine-tuning training set is straightforward and that fine-tuning with even relatively small datasets leads to predictive models. These results suggest that the systematic use of LLMs to guide experiments and simulations will be a powerful technique in any research study, significantly reducing unnecessary experiments or computations.
Assessment of fine-tuned large language models for real-world chemistry and material science applications / Van Herck, Joren; Victoria Gil, María; Maik Jablonka, Kevin; Abrudan, Alex; Sode Anker, Andy; Asgari, Mehrdad; Blaiszik, Ben; Buffo, Antonio; Choudhury, Leander; Corminboeuf, Clemence; Daglar, Hilal; Mohammad Elahi, Amir; Foster, Ian; Garcia, Susana; Garvin, Matthew; Godin, Guillaume; Good, Lydia L.; Gu, Jianan; Xiao Hu, Noémie; Jin, Xin; Junkers, Tanja; Keskin, Seda; Knowles, Tuomas; Laplaza, Rubén; Lessona, Michele; Majumdar, Sauradeep; Mashhadimoslem, Hossein; Mcintosh, Ruaraidh; Mohamad Moosavi, Seyed; Mouriño, Beatriz; Nerli, Francesca; Pevida, Covadonga; Poudineh, Neda; Rajabi Kochi, Mahyar; Saar, Kadi L.; Hooriabad Saboor, Fahimeh; Sagharichiha, Morteza; Schmidt, Kj; Shi, Jiale; Simone, Elena; Svatunek, Dennis; Taddei, Marco; Tetko, Igor V.; Tolnai, Domonkos; Vahdatifar, Sahar; Whitmer, Jonathan K.; Wieland, Florian; Willumeit, Regine; Zuttel, Andreas; Smit, Berend. - In: CHEMICAL SCIENCE. - ISSN 2041-6520. - 16:2(2025), pp. 670-684. [10.1039/d4sc04401k]
Assessment of fine-tuned large language models for real-world chemistry and material science applications
Antonio Buffo;Michele Lessona;Elena Simone;
2025
Abstract
The current generation of large language models (LLMs) has limited chemical knowledge. Recently, it has been shown that these LLMs can learn and predict chemical properties through fine-tuning. Using natural language to train machine learning models opens doors to a wider chemical audience, as field-specific featurization techniques can be omitted. In this work, we explore the potential and limitations of this approach. We studied the performance of fine-tuning three open-source LLMs (GPT-J-6B, Llama-3.1-8B, and Mistral-7B) for a range of different chemical questions. We benchmark their performances against “traditional” machine learning models and find that, in most cases, the fine-tuning approach is superior for a simple classification problem. Depending on the size of the dataset and the type of questions, we also successfully address more sophisticated problems. The most important conclusions of this work are that, for all datasets considered, their conversion into an LLM fine-tuning training set is straightforward and that fine-tuning with even relatively small datasets leads to predictive models. These results suggest that the systematic use of LLMs to guide experiments and simulations will be a powerful technique in any research study, significantly reducing unnecessary experiments or computations.File | Dimensione | Formato | |
---|---|---|---|
d4sc04401k.pdf
accesso aperto
Tipologia:
2a Post-print versione editoriale / Version of Record
Licenza:
Creative commons
Dimensione
1.8 MB
Formato
Adobe PDF
|
1.8 MB | Adobe PDF | Visualizza/Apri |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/2997101