Wuxin: Architecture Design and Empirical Study for Vertical-Domain Large Language Model System
CSTR:
Author:
Affiliation:

1.North Automatic Control Technology Institute, Taiyuan 030006, China;2.Key Laboratory of CTC &IE (Engineering University of PAP),Ministry of Education, Xi’an 710086, China

Clc Number:

TP391.1

Fund Project:

  • Article
  • |
  • Figures
  • |
  • Metrics
  • |
  • Reference
  • |
  • Related
  • |
  • Cited by
  • |
  • Materials
  • |
  • Comments
    Abstract:

    In customized scenarios, it is urgent to enhance the understanding and generation capabilities of large language models (LLMs) in specific vertical domains. We propose a paradigm for developing vertical-domain LLM system named “Wuxin”, which covers a series of development methods for LLM systems, including architecture, data, model, and training. Wuxin utilizes human-in-the-loop data augmentation to improve the quality of military training injury question and answer datasets, and employs the GaLore strategy to perform efficient full-parameter fine-tuning on small LLMs. Experimental results show that the adopted full-parameter fine-tuning method outperforms LoRA fine-tuning in terms of convergence and accuracy. Furthermore,Wuxin demonstrates significant advantages in understanding professional military training injury knowledge, as well as overcoming hallucinations. Our achievements can provide references for the design and application of question-answering LLM systems in vertical domains.

    Reference
    Related
    Cited by
Get Citation

ZHU Xinli, GAO Zhiqiang, JI Weitong, LI Shaohua, LI Songjie. Wuxin: Architecture Design and Empirical Study for Vertical-Domain Large Language Model System[J].,2025,40(3):637-646.

Copy
Related Videos

Share
Article Metrics
  • Abstract:
  • PDF:
  • HTML:
  • Cited by:
History
  • Received:January 26,2025
  • Revised:March 21,2025
  • Adopted:
  • Online: June 13,2025
  • Published:
Article QR Code