Song Zihan, Hwang Gyo-Yeob, Zhang Xin, Huang Shan, Park Byung-Kwon
Dong-A University, Busan, 49315, Republic of Korea.
Sci Rep. 2025 Jan 10;15(1):1608. doi: 10.1038/s41598-025-85715-7.
The exponential growth of scientific articles has presented challenges in information organization and extraction. Automation is urgently needed to streamline literature reviews and enhance insight extraction. We explore the potential of Large Language Models (LLMs) in key-insights extraction from scientific articles, including OpenAI's GPT-4.0, MistralAI's Mixtral 8 × 7B, 01AI's Yi, and InternLM's InternLM2. We have developed an article-level key-insight extraction system based on LLMs, calling it ArticleLLM. After evaluating the LLMs against manual benchmarks, we have enhanced their performance through fine-tuning. We propose a multi-actor LLM approach, merging the strengths of multiple fine-tuned LLMs to improve overall key-insight extraction performance. This work demonstrates not only the feasibility of LLMs in key-insight extraction, but also the effectiveness of cooperation of multiple fine-tuned LLMs, leading to efficient academic literature survey and knowledge discovery.
科学文章的指数级增长给信息组织和提取带来了挑战。迫切需要自动化来简化文献综述并增强见解提取。我们探索了大语言模型(LLMs)在从科学文章中提取关键见解方面的潜力,包括OpenAI的GPT - 4.0、MistralAI的Mixtral 8×7B、01AI的Yi和澜舟科技的InternLM2。我们基于大语言模型开发了一个文章级关键见解提取系统,将其命名为ArticleLLM。在根据人工基准评估大语言模型后,我们通过微调提高了它们的性能。我们提出了一种多模型大语言模型方法,融合多个微调大语言模型的优势以提高整体关键见解提取性能。这项工作不仅证明了大语言模型在关键见解提取方面的可行性,还证明了多个微调大语言模型合作的有效性,从而实现高效的学术文献调研和知识发现。