[1] OPENAI. GPT-4 technical report[R/OL]. [2025-05-14]. https:// arxiv.org/abs/2303.08774.
[2] DEEPSEEK-AI. DeepSeek-V3 technical report[R/OL]. [2025-05-14]. https://arxiv.org/abs/2412.19437.
[3] QWEN. Qwen2.5 technical report[R/OL]. [2025-05-14]. https://arxiv.org/abs/2412.15115.
[4] CUI J X, NING M N, LI Z J, et al. Chatlaw: a multi-agent collaborative legal assistant with knowledge graph enhanced mixture-of-experts large language model[EB/OL]. [2025-05-14]. https://arxiv.org/abs/2306.16092.
[5] HUANG J J, ZHU H R, XU C, et al. AuditWen: an open-source large language model for audit[C]//Proceedings of the 23rd Chinese National Conference on Computational Linguistics. Singapore: Springer, 2024: 505-521.
[6] ZHANG H B, CHEN J Y, JIANG F, et al. HuatuoGPT, towards taming language model to be a doctor[C]//Findings of the Association for Computational Linguistics: EMNLP 2023. Stroudsburg: ACL, 2023: 10859-10885.
[7] XIE Q Q, HAN W G, ZHANG X, et al. PIXIU: a large language model, instruction data and evaluation benchmark for finance[EB/OL]. [2025-05-14]. https://arxiv.org/abs/2306. 05443.
[8] ZHOU C T, LIU P F, XU P X, et al. LIMA: less is more for alignment[C]//Advances in Neural Information Processing Systems 36, 2023.
[9] CHEN L C, LI S Y, YAN J, et al. AlpaGasus: training a better Alpaca with fewer data[C]//Proceedings of the 12th International Conference on Learning Representations, 2024.
[10] CAO Y H, KANG Y B, WANG C, et al. Instruction mining: instruction data selection for tuning large language models[EB/OL]. [2025-05-14]. https://arxiv.org/abs/2307.06290.
[11] LIU W, ZENG W H, HE K Q, et al. What makes good data for alignment? A comprehensive study of automatic data selection in instruction tuning[C]//Proceedings of the 12th International Conference on Learning Representations, 2024.
[12] LI M, ZHANG Y, LI Z T, et al. From quantity to quality: boosting LLM performance with self-guided data selection for instruction tuning[C]//Proceedings of the 2024 Conference of the North American Chapter of the Association for Compu-tational Linguistics: Human Language Technologies. Stroudsburg: ACL, 2024: 7602-7635.
[13] DU Q L, ZONG C Q, ZHANG J J. MoDS: model-oriented data selection for instruction tuning[EB/OL]. [2025-05-14]. https://arxiv.org/abs/2311.15653.
[14] GE Y, LIU Y L, HU C, et al. Clustering and ranking: diversity-preserved instruction selection through expert-aligned quality estimation[C]//Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. Stroudsburg: ACL, 2024: 464-478.
[15] YANG Y M, NAN Y, YE J J, et al. Measuring data diversity for instruction tuning: a systematic analysis and a reliable metric[C]//Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics. Stroudsburg: ACL, 2025: 18530-18549.
[16] LIN C Y. ROUGE: a package for automatic evaluation of summaries[C]//Proceedings of the Annual Meeting of the Association for Computational Linguistics. Stroudsburg: ACL, 2004: 74-81.
[17] ZHANG T Y, KISHORE V, WU F, et al. BERTScore: evaluating text generation with BERT[EB/OL]. [2025-05-14]. https://arxiv.org/abs/1904.09675.
[18] YAN Y L, CHEN M, SHYU M L, et al. Deep learning for imbalanced multimedia data classification[C]//Proceedings of the 2015 IEEE International Symposium on Multimedia. Piscataway: IEEE, 2015: 483-488.
[19] REZVANI S, WANG X Z. A broad review on class imbalance learning techniques[J]. Applied Soft Computing, 2023, 143: 110415.
[20] MOHAMMED R, RAWASHDEH J, ABDULLAH M. Machine learning with oversampling and undersampling techniques: overview study and experimental results[C]//Proceedings of the 11th International Conference on Information and Communication Systems. Piscataway: IEEE, 2020: 243-248.
[21] YUAN Y G, WEI J N, HUANG H S, et al. Review of resampling techniques for the treatment of imbalanced industrial data classification in equipment condition monitoring[J]. Engineering Applications of Artificial Intelligence, 2023, 126: 106911.
[22] RIVERA W A, XANTHOPOULOS P. A priori synthetic over-sampling methods for increasing classification sensitivity in imbalanced data sets[J]. Expert Systems with Applications, 2016, 66: 124-135.
[23] AHSAN M M, ALI M S, SIDDIQUE Z. Enhancing and improving the performance of imbalanced class data using novel GBO and SSG: a comparative analysis[J]. Neural Networks, 2024, 173: 106157.
[24] SUN Z B, SONG Q B, ZHU X Y, et al. A novel ensemble method for classifying imbalanced data[J]. Pattern Recognition, 2015, 48(5): 1623-1637.
[25] CHAMLAL H, KAMEL H, OUADERHMAN T. A hybrid multi-criteria meta-learner based classifier for imbalanced data[J]. Knowledge-Based Systems, 2024, 285: 111367.
[26] DOUCETTE J, HEYWOOD M I. GP classification under imbalanced data sets: active sub-sampling and AUC approximation[C]//Proceedings of the 11th European Conference on Genetic Programming. Berlin, Heidelberg: Springer, 2008: 266-277.
[27] JEATRAKUL P, WONG K W, FUNG C C. Classification of imbalanced data by combining the complementary neural network and SMOTE algorithm[C]//Proceedings of the 17th International Conference on Neural Information Processing. Models and Applications. Berlin, Heidelberg: Springer, 2010: 152-159.
[28] YE J S, LIU P J, SUN T X, et al. Data mixing laws: optimizing data mixtures by predicting language modeling performance[EB/OL]. [2025-05-15]. https://arxiv.org/abs/2403.16952.
[29] DONG G T, YUAN H Y, LU K M, et al. How abilities in large language models are affected by supervised fine-tuning data composition[C]//Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics. Stroudsburg: ACL, 2024: 177-198.
[30] 黄佳佳, 朱浩然, 徐超, 等. 大语言模型在审计判断中的应用研究[J]. 财会月刊, 2025(6): 21-28.
HUANG J J, ZHU H R, XU C, et al. Application of large language models in audit judgment[J]. Finance and Accoun-ting Monthly, 2025(6): 21-28.
[31] ZHENG Y W, ZHANG R C, ZHANG J H, et al. LlamaFactory: unified efficient fine-tuning of 100+ language models[C]//Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics. Stroudsburg: ACL, 2024: 400-410.
[32] HU E J, SHEN Y L, WALLIS P, et al. LoRA: low-rank adaptation of large language models[EB/OL]. [2025-05-15]. https://arxiv.org/abs/2106.09685.
[33] ZHENG L M, CHIANG W L, SHENG Y, et al. Judging LLM-as-a-judge with MT-bench and chatbot arena[C]//Proceedings of the 37th International Conference on Neural Infor-mation Processing Systems, 2023: 46595-46623.
[34] YUE S B, LIU S J, ZHOU Y X, et al. LawLLM: intelligent legal system with legal reasoning and verifiable retrieval[C]//Proceedings of the 29th International Conference on Database Systems for Advanced Applications. Singapore: Springer, 2024: 304-321. |