Journal of Frontiers of Computer Science and Technology ›› 2024, Vol. 18 ›› Issue (9): 2361-2369.DOI: 10.3778/j.issn.1673-9418.2406067
• Special Issue on Constructions and Applications of Large Language Models in Specific Domains • Previous Articles Next Articles
JI Guiyang, WANG Peiyan, YU Zhuo
Online:
2024-09-01
Published:
2024-09-01
纪贵阳,王裴岩,余卓
JI Guiyang, WANG Peiyan, YU Zhuo. Research on Knowledge Injection Method for Large Language Model Oriented to Process Specification Texts[J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(9): 2361-2369.
纪贵阳, 王裴岩, 余卓. 面向工艺规范文本的大语言模型知识注入方法研究[J]. 计算机科学与探索, 2024, 18(9): 2361-2369.
[1] 全国科学技术名词审定委员会. 机械工程名词第二分册[M]. 2版. 北京: 科学出版社, 2021. China National Committee for Terminology in Science and Technology. Mechanical engineering terminology volume II[M]. 2nd ed. Beijing: China Science Publishing & Media Ltd., 2021. [2] 邢磊, 张驰, 张柯尧, 等. 浅谈航空装备管理非结构化数据治理实施策略[J]. 航空标准化与质量, 2022(2): 26-31. XING L, ZHANG C, ZHANG K Y, et al. Implementation strategy of unstructured data governance in aerospace equip-ment management: a discussion[J]. Aeronautic Standardization & Quality, 2022(2): 26-31. [3] THIRUNAVUKARASU A J, TING D S J, ELANGOVAN K, et al. Large language models in medicine[J]. Nature Medicine, 2023, 29(8): 1930-1940. [4] ZHANG X, YANG Q. Xuanyuan 2.0: a large Chinese financial chat model with hundreds of billions parameters[C]//Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. New York: ACM, 2023: 4435-4439. [5] 丁忠军, 杨增辉, 张晋军, 等. 基于岗位知识管理的工艺规范体系建设[J]. 航天工业管理, 2016(5): 58-61. DING Z J, YANG Z H, ZHANG J J, et al. Construction of process specification system based on job knowledge manage-ment[J]. Aerospace Industry Management, 2016(5): 58-61. [6] HOULSBY N, GIURGIU A, JASTRZEBSKI S, et al. Parameter-efficient transfer learning for NLP[C]//Proceedings of the 36th International Conference on Machine Learning, Long Beach, Jun 9-15, 2019: 2790-2799. [7] LI X L, LIANG P. Prefix-Tuning: optimizing continuous prompts for generation[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing. Stroudsburg: ACL, 2021: 4582-4597. [8] LESTER B, AL-RFOU R, CONSTANT N. The power of scale for parameter-efficient prompt tuning[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Stroudsburg: ACL, 2021: 3045-3059. [9] LIU X, JI K, FU Y, et al. P-Tuning: prompt tuning can be comparable to fine-tuning across scales and tasks[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics. Stroudsburg: ACL, 2022: 61-68. [10] HU E J, WALLIS P, ALLEN-ZHU Z, et al. LoRA: low-rank adaptation of large language models[C]//Proceedings of the 10th International Conference on Learning Representations, Apr 25-29, 2022. [11] ZAKEN E B, GOLDBERG Y, RAVFOGEL S. BitFit: simple parameter-efficient fine-tuning for transformer-based masked language-models[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics. Strouds-burg: ACL, 2022: 1-9. [12] SU Y, HAN X, ZHANG Z, et al. CokeBERT: contextual knowledge selection and embedding towards enhanced pre-trained language models[J]. AI Open, 2021, 2: 127-134. [13] WANG X, GAO T, ZHU Z, et al. KEPLER: a unified model for knowledge embedding and pre-trained language representation[J]. Transactions of the Association for Computational Linguistics, 2021, 9: 176-194. [14] LIU W, ZHOU P, ZHAO Z, et al. K-BERT: enabling language representation with knowledge graph[C]//Proceedings of the 2020 AAAI Conference on Artificial Intelligence. Menlo Park: AAAI, 2020: 2901-2908. [15] FU P, ZHANG Y, WANG H, et al. Revisiting the knowledge injection frameworks[C]//Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. Stroudsburg: ACL, 2023: 10983-10997. [16] JIANG Z, ZHONG L, SUN M, et al. Efficient knowledge infusion via KG-LLM alignment[EB/OL]. [2024-07-03]. https://arxiv.org/abs/2406.03746. [17] GUO T, YANG Q, WANG C, et al. Knowledgenavigator: lever-aging large language models for enhanced reasoning over knowledge graph[EB/OL]. [2024-04-14]. https://arxiv.org/abs/2312.15880. [18] WU H, ZHANG Y, HAN Z, et al. Quartet Logic: a four-step rea-soning (QLFR) framework for advancing short text classification[EB/OL]. [2024-04-14]. https://arxiv.org/abs/2401.03158. [19] YE D, LIN Y, LI P, et al. A simple but effective pluggable entity lookup table for pre-trained language models[C]//Proceedings of the 60th Annual Meeting of the Association for Com-putational Linguistics. Stroudsburg: ACL, 2022: 523-529. [20] ZHANG Z, ZENG Z, LIN Y, et al. Plug-and-play knowledge injection for pre-trained language models[C]//Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics. Stroudsburg: ACL, 2023: 10641-10658. [21] LEWIS P, PEREZ E, PIKTUS A, et al. Retrieval-augmented generation for knowledge-intensive NLP tasks[C]//Advances in Neural Information Processing Systems 33, Dec 6-12, 2020: 9459-9474. [22] JIANG Z, XU F F, GAO L, et al. Active retrieval augmented generation[C]//Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. Stroudsburg: ACL, 2023: 7969-7992. [23] WANG Y, LI P, SUN M, et al. Self-knowledge guided retrieval augmentation for large language models[C]//Procee-dings of the 2023 Conference on Empirical Methods in Natural Language Processing. Stroudsburg: ACL, 2023: 10303-10315. [24] JEONG S, BAEK J, CHO S, et al. Adaptive-RAG: learning to adapt retrieval-augmented large language models through question complexity[C]//Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Stroudsburg: ACL, 2024: 7029-7043. [25] XU S, PANG L, YU M, et al. Unsupervised information refinement training of large language models for retrieval-augmented generation[EB/OL]. [2024-04-14]. https://arxiv.org/abs/2402.18150. [26] LIU J, HUANG X, CHEN Z, et al. DRAK: unlocking molecular insights with domain-specific retrieval-augmented knowledge in LLMs[EB/OL]. [2024-07-03]. https://arxiv.org/abs/ 2406.18535. [27] WU C, LIN W, ZHANG X, et al. PMC-LLaMA: toward building open-source language models for medicine[EB/OL]. [2024-04-14]. https://arxiv.org/abs/ 2304.14454. [28] SHI W, MIN S, LOMELI M, et al. In-context pretraining: language modeling beyond document boundaries[C]//Proceedings of the 12th International Conference on Learning Representations, Vienna, May 7-11, 2024. [29] WANG J, WANG C, TAN C, et al. Knowledgeable in-context tuning: exploring and exploiting factual knowledge for in-context learning[C]//Proceedings of the 2024 Annual Conference of the North American Chapter of the Association for Computational Linguistics. Stroudsburg: ACL, 2024: 3261-3280. [30] DERNBACH S, AGARWAL K, ZUNIGA A, et al. GLaM: fine-tuning large language models for domain knowledge graph alignment via neighborhood partitioning and generative subgraph encoding[C]//Proceedings of the AAAI 2024 Spring Symposium Series, Stanford, Mar 25-27, 2024. Menlo Park: AAAI, 2024: 82-89. [31] CHENG D, HUANG S, WEI F. Adapting large language models via reading comprehension[C]//Proceedings of the 12th International Conference on Learning Representations, Vienna, May 7-11, 2024. [32] PAPINENI K, ROUKOS S, WARD T, et al. BLEU: a method for automatic evaluation of machine translation[C]//Procee-dings of the 40th Annual Meeting of the Association for Com-putational Linguistics. Stroudsburg: ACL, 2002: 311-318. [33] LIN C Y. Rouge: a package for automatic evaluation of sum-maries[C]//Proceedings of the 2004 Workshop on Text Summarization Branches Out. Stroudsburg: ACL, 2004: 74-81. |
[1] | XIANG Xiaowei, SHEN Yanguang, HU Minghao, YAN Tianwei, LUO Wei, LUO Zhunchen. Research on Science and Technology Policy and Regulation Q&A System Driven by Large Models [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(9): 2349-2360. |
[2] | LI Yifei, ZHANG Lingling, DONG Yuxuan, WANG Jiaxin, ZHONG Yujie, WEI Bifan. Large Language Model Augmentation and Feature Alignment Method for Few-Shot Continual Relation Extraction [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(9): 2326-2336. |
[3] | CHEN Longfei, GAO Xin, HOU Haotian, YE Chuyang, LIU Ya'ou, ZHANG Meihui. Application of Generative Large Language Models in Chinese Radiology Domain [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(9): 2337-2348. |
[4] | LUO Shijie, JIN Rize, HAN Shuzhen. Research on University Basic Knowledge Question-Answering Using Low-Rank Encoding to Optimize Large Language Model [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(8): 2156-2168. |
[5] | SHENG Lei, CHEN Xiliang, LAI Jun. Offline Multi-agent Reinforcement Learning Method Based on Latent State Distribution GPT [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(8): 2169-2179. |
[6] | ZHANG Qi, ZHONG Hao. Submodular Optimization Approach for Entity Summarization in Knowledge Graph Driven by Large Language Models [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(7): 1806-1813. |
[7] | FENG Jun, CHANG Yanghong, LU Jiamin, TANG Hailin, LYU Zhipeng, QIU Yuchun. Construction and Application of Knowledge Graph for Water Engineering Scheduling Based on Large Language Model [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(6): 1637-1647. |
[8] | WANG Runzhou, ZHANG Xinsheng. Medical Knowledge Graph Question-Answering System Based on Hybrid Dynamic Masking and Multi-strategy Fusion [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(10): 2770-2786. |
[9] | PEI Bingsen, LI Xin, JIANG Zhangtao, LIU Mingshuai. Research on Public Security Professional Small Sample Knowledge Extraction Method Based on Large Language Model [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(10): 2630-2642. |
[10] | LI Li, SHI Rongliang, GUO Xu, JIANG Hongxin. Diagnosis of Power System Defects by Large Language Models and Graph Neural Networks [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(10): 2643-2655. |
[11] | ZHANG Caike, LI Xiaolong, ZHENG Sheng, CAI Jiajun, YE Xiaozhou, LUO Jing. Research on Construction and Application of Knowledge Graph Based on Large Language Model [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(10): 2656-2667. |
[12] | LIANG Jia, ZHANG Liping, YAN Sheng, ZHAO Yubo, ZHANG Yawen. Research Progress of Named Entity Recognition Based on Large Language Model [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(10): 2594-2615. |
[13] | JI Xiangyu, WANG Xin, ZHANG Heyi, MENG Zhaopeng, ZHANG Junhua, ZHUANG Pengwei, JIA Yongzhe, XU Dawei. Knowledge Augmentation on Traditional Chinese Medicine Language Model [J]. Journal of Frontiers of Computer Science and Technology, 2024, 18(10): 2616-2629. |
[14] | ZHANG Heyi, WANG Xin, HAN Lifan, LI Zhao, CHEN Zirui, CHEN Zhe. Research on Question Answering System on Joint of Knowledge Graph and Large Language Models [J]. Journal of Frontiers of Computer Science and Technology, 2023, 17(10): 2377-2388. |
[15] | PENG Huang, ZENG Weixin, ZHOU Jie, TANG Jiuyang, ZHAO Xiang. Contrast Research of Representation Learning in Entity Alignment Based on Graph Neural Network [J]. Journal of Frontiers of Computer Science and Technology, 2023, 17(10): 2343-2357. |
Viewed | ||||||||||||||||||||||||||||||||||||||||||||||
Full text 100
|
|
|||||||||||||||||||||||||||||||||||||||||||||
Abstract 125
|
|
|||||||||||||||||||||||||||||||||||||||||||||
/D:/magtech/JO/Jwk3_kxyts/WEB-INF/classes/