Framework and construction methodology of underground engineering domain knowledge large language model: UndergrGPT
1 State Key Laboratory for Tunnel Engineering, Jinan, China
2 School of Qilu Transportation, Shandong University, Jinan, China
3 School of Computer Science and Technology, Shandong University, Qingdao, China
  • Volume
  • Citation
    Xu Z, Wang Z, Ren P, Zhang X, Li T. Framework and construction methodology of underground engineering domain knowledge large language model: UndergrGPT. Smart Constr. 2024(2):0012, https://doi.org/10.55092/sc20240012. 
  • DOI
    10.55092/sc20240012
  • Copyright
    Copyright2024 by the authors. Published by ELSP.
Abstract

Large Language Models (LLMs) have achieved tremendous success in natural language processing tasks. However, in vertical domains such as underground engineering, the lack of large-scale structured corpora and the requirement for high-accuracy responses significantly increase the difficulty of constructing domain knowledge LLMs. Simultaneously, these challenges amplify the inherent flaws of models, such as machine hallucination and outdated content. To address the aforementioned challenges, firstly, we proposed a framework for constructing underground engineering domain knowledge LLMs, which includes human-machine collaborative knowledge extraction, parameter efficient fine-tuning, and retrieval-augmented generation. The framework aims to provide a universal construction paradigm for underground engineering domain knowledge LLMs. Secondly, within the proposed framework, we established an underground engineering domain knowledge database (UndergrData), introduced a domain knowledge extraction approach based on a few-shot prompt learning data augmentation strategy, resulting in the formation of the training dataset; we applied an underground engineering domain knowledge LLMs fine-tuning method under low-resource and constructed the first underground engineering domain knowledge LLM base (UndergrGPT); we adopted the approach of underground engineering domain knowledge retrieval-augmented generation, enabling high-quality response generation. Finally, compared with general LLMs, UndergrGPT effectively reduced hallucinations, demonstrating a significant advantage in response accuracy. It’s worth noting that the proposed framework can be extended to other vertical domains.

Keywords

domain knowledge large language model; domain knowledge database; construction framework; knowledge extraction; parameter efficient fine-tuning; retrieval-augmented generation

Preview