Design and Evaluation of Knowledge-Distilled LLM for Improving the Efficiency of School Administrative Document Processing
Abstract
1. Introduction
- The design and implementation of a Knowledge-Distilled LLM architecture using actual datasets.
- The verification of the applicability of the proposed model for improving the efficiency of administrative document processing in public institutions (schools, district offices, neighborhood offices, etc.) in an on-premise environment.
2. Related Work
2.1. Text Mining Technology Trends and Application Cases
2.2. Knowledge Distillation-Based LLM Lightweight Technology
2.3. On-Premise AI System Implementation Technology
3. Methodology
3.1. Design of the Proposed Model OP-LLM-SA
3.2. Knowledge Distillation Pipeline
4. Performance Evaluation of the Implementation Model
4.1. Experimental Environment
4.2. Text Mining Performance Evaluation
- Token Accuracy: Measures word- and phrase-level similarity between the original text and the generated administrative document.
- Sentence Naturalness: This is an indicator that determines how natural the generated administrative documents are, and it was measured at 99.10%, which is almost identical to the original text in terms of grammar and fluency [33].
- Completed Sentence Rate: This metric measures the average sentence length and the percentage of complete sentences in generated administrative documents. Sentence completeness was measured using a GPT model, yielding an average sentence length of 15.68 characters and a complete sentence rate of 97.19% when compared to existing official documents.
- Format Conformity: This metric measures the degree of adherence to official document formatting standards. Measured using the GPT model, it achieved 92.85%, indicating a high rate of format element reproduction.
4.3. System Efficiency Evaluation
4.4. LLM Performance Evaluation
- The format and structure of the proposed model are suitable, and the ratio of completed sentences is also excellent.
- While there are concerns about the leakage of terms used in schools when using ChatGPT4o, the proposed model appears ready for immediate use as it eliminates the risk of external leakage.
- It is regrettable that only text generation is possible. Please enable the use of photos or graphs.
5. Conclusions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Choi, H.-Y.; Choi, Y. Diagnosis of Public Language Used in Administrative Agencies for Develop ‘How to Write Official Documents’ Teaching Materials. J. Lang. Lit. 2014, 58, 53–75. [Google Scholar] [CrossRef]
- Cho, Y. Analysis of Teachers’ Job Burden and Influencing Factors by Job Area. J. Educ. Res. 2024, 22, 89–114. [Google Scholar] [CrossRef]
- Lee, H.-S. The Administrative Work Burden of Childcare Teachers and Improvement Measures. Master’s Thesis. Kookmin University Graduate School of Education, Seoul, Republic of Korea, 2018. Available online: https://www.riss.kr/link?id=T14897125 (accessed on 7 July 2025).
- Mandvikar, S. Augmenting Intelligent Document Processing (IDP) Workflows with Contemporary Large Language Models (LLMs). Int. J. Comput. Trends Technol. 2023, 71, 80–91. [Google Scholar] [CrossRef]
- Oh, Y. Issues and Prospects of AI Digital English Education: Text Analysis Based on Natural Language Processing. Korea J. Engl. Lang. Linguist. 2025, 25, 330–366. [Google Scholar] [CrossRef]
- Jo, W. The Possibility of Collective Emotion Narrative Research Using Self-Attention. Korean J. Sociol. 2024, 58, 315–350. [Google Scholar] [CrossRef]
- Miranda, M.; Beugnot, G.; Heitz, M.; Vert, J.-P. Preserving Privacy in Large Language Models: A Survey on Current Threats and Solutions. arXiv 2024, arXiv:2408.05212. [Google Scholar] [CrossRef]
- Personal Information Protection Act. Article 32-2; Amended 8 March 2023. Available online: https://www.law.go.kr/LSW/lsInfoP.do?lsiSeq=250244&efYd=20230308 (accessed on 19 July 2025).
- Liu, C.; Zhang, H.; Zhao, K.; Ju, X.; Yang, L. LLMEmbed: Rethinking Lightweight LLM’s Genuine Function in Text Classification. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024), Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Stroudsburg, PA, USA, 2024; pp. 7997–8011. Available online: https://aclanthology.org/2024.acl-long.433.pdf (accessed on 20 April 2025).
- Aleqabie, H.J.; Sfoq, M.S.; Albeer, R.A.; Abd, E.H. A Review of Text Mining Techniques: Trends, and Applications in Various Domains. Iraqi J. Comput. Sci. Math. 2024, 5, 9. [Google Scholar] [CrossRef]
- O’Mara-Eves, A.; Thomas, J.; McNaught, J.; Miwa, M.; Ananiadou, S. Using Text Mining for Study Identification in Systematic Reviews: A Systematic Review of Current Approaches. Syst. Rev. 2015, 4, 5. [Google Scholar] [CrossRef] [PubMed]
- Gupta, A.; Lamba, H.; Kumaraguru, P.; Joshi, A. Comprehensive Review of Text Mining Applications in Finance. Financ. Innov. 2020, 6, 39. [Google Scholar] [CrossRef]
- Gupta, T.; Himmelstein, D.S.; Baranzini, S.E.; Greene, C.S. MatSciBERT: A Materials Domain Language Model for Text Mining and Information Extraction. npj Comput. Mater. 2022, 8, 217. [Google Scholar] [CrossRef]
- Taha, K.; Yoo, P.D.; Yeun, C.; Homouz, D.; Taha, A. A Comprehensive Survey of Text Classification Techniques and Their Research Applications: Observational and Experimental Insights. Comput. Sci. Rev. 2024, 54, 100664. [Google Scholar] [CrossRef]
- Shin, S.-Y.; Choi, H.; Kim, D. Analyzing Contents of Open Data in Korean Local Governments: A Text-Mining Approach. Korean J. Public Adm. 2021, 30, 129–171. [Google Scholar]
- Han, S.; Lee, H. An Analysis of News Articles Related to Public Records Using Text Mining. J. Korean Soc. Archival Rec. Manag. 2025, 25, 103–127. [Google Scholar] [CrossRef]
- Lee, J.-S.; Jung, J.-H. A Study on the Conceptual Expansion of Public Design Research through Text Mining. J. Public Des. 2025, 16, 57–66. [Google Scholar] [CrossRef]
- Phuong, M.; Lampert, C.H. Towards Understanding Knowledge Distillation. In Proceedings of the 36th International Conference on Machine Learning (ICML 2019), Long Beach, CA, USA, 9–15 June 2019; PMLR: Cambridge, MA, USA, 2019; Volume 97, pp. 5142–5151. Available online: https://proceedings.mlr.press/v97/phuong19a.html (accessed on 28 March 2025).
- Mansourian, A.M.; Saadatfar, H.; Torkamani, M.; Kazemi, H.; Zolbanin, H.M.; Moghaddam, H.A. A Comprehensive Survey on Knowledge Distillation. arXiv 2025, arXiv:2503.12067. [Google Scholar] [CrossRef]
- Xu, X.; Li, M.; Tao, C.; Shen, T.; Cheng, R.; Li, J.; Xu, C.; Tao, D.; Zhou, T. A Survey on Knowledge Distillation of Large Language Models. arXiv 2024, arXiv:2402.13116. [Google Scholar]
- Yang, C.; Lu, W.; Zhu, Y.; Wang, Y.; Chen, Q.; Gao, C.; Yan, B.; Chen, Y. Survey on Knowledge Distillation for Large Language Models: Methods, Evaluation, and Application. arXiv 2024, arXiv:2407.01885. [Google Scholar] [CrossRef]
- Gu, Y.; Dong, L.; Wei, F.; Huang, M. MiniLLM: Knowledge Distillation of Large Language Models. In Proceedings of the International Conference on Learning Representations (ICLR 2024), Vienna, Austria, 5–9 May 2024; Available online: https://openreview.net/forum?id=5h0qf7IBZZ (accessed on 7 June 2025).
- Fortuna, C.; Mušič, D.; Čerar, G.; Čampa, A.; Kapsalis, P.; Mohorčič, M. On-Premise Artificial Intelligence as a Service for Small and Medium Size Setups. arXiv 2022, arXiv:2210.06956. [Google Scholar] [CrossRef]
- Tachu, E. A Quantitative Study of the Relationship between Cloud Flexibility and On-Premise Flexibility. Issues Inf. Syst. 2022, 23, 214–238. [Google Scholar]
- Pillai, P. Cloud vs. On-Premise Data Warehousing: A Strategic Analysis for Financial Institutions. J. Comput. Sci. Technol. Stud. 2025, 7, 503–513. [Google Scholar] [CrossRef]
- Luka, C. Hybrid Integration Model: Integrating On-Premise Legacy Systems with Azure and AI Technologies. ResearchGate Preprint. 2025. Available online: https://www.researchgate.net (accessed on 2 June 2025).
- Gautam, A. Optimizing Data Storage for AI, Generative AI, and Machine Learning: Challenges, Architectures, and Future Direction. Int. J. Comput. Appl. 2025, 186, 29–33. [Google Scholar] [CrossRef]
- Yu, Y.; Kim, N. Text Classification Using Heterogeneous Knowledge Distillation. J. Korea Soc. Comput. Inf. 2022, 27, 29–41. [Google Scholar] [CrossRef]
- Zou, Y.; Xu, Z.; Zhang, Q.; Lin, Z.; Wang, T.; Liu, Z.; Li, D. Few-Shot Learning With Manifold-Enhanced LLM for Handling Anomalous Perception Inputs in Autonomous Driving. IEEE Trans. Intell. Transp. Syst. Early Access 2025, 1–18. [Google Scholar] [CrossRef]
- Saxena, P.; Janzen, S.; Maaß, W. Streamlining LLMs: Adaptive Knowledge Distillation for Tailored Language Models. In Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 4: Student Research Workshop), Albuquerque, NM, USA, 29 April–4 May 2025; Association for Computational Linguistics: Stroudsburg, PA, USA, 2025; pp. 448–455. [Google Scholar] [CrossRef]
- Fu, Z.; Yang, H.; So, A.M.-C.; Lam, W.; Bing, L.; Collier, N. On the Effectiveness of Parameter-Efficient Fine-Tuning. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington DC, USA, 7–14 February 2023; Volume 37, pp. 12799–12807. [Google Scholar] [CrossRef]
- Xia, C.; Xing, C.; Du, J.; Yang, X.; Feng, Y.; Xu, R.; Yin, W.; Xiong, C. FOFO: A Benchmark to Evaluate LLMs’ Format-Following Capability. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Stroudsburg, PA, USA, 2024; pp. 680–699. [Google Scholar] [CrossRef]
- Bahl, L.R.; Jelinek, F.; Mercer, R.L. A Maximum Likelihood Approach to Continuous Speech Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 1983, 5, 179–190. [Google Scholar] [CrossRef] [PubMed]
- Shlegeris, B.; Roger, F.; Chan, L.; McLean, E. Language Models Are Better Than Humans at Next-Token Prediction. Trans. Mach. Learn. Res. 2024. Available online: https://openreview.net/forum?id=RNsnSLdmV7 (accessed on 13 July 2025).
- Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.-J. BLEU: A Method for Automatic Evaluation of Machine Translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL 2002), Stroudsburg, PA, USA, 7–12 July 2002; Association for Computational Linguistics: Philadelphia, PA, USA, 2002; pp. 311–318. [Google Scholar] [CrossRef]
- Lin, C.-Y. ROUGE: A Package for Automatic Evaluation of Summaries. In Proceedings of the Workshop on Text Summarization Branches Out (WAS 2004), Barcelona, Spain, 25–26 July 2004; Association for Computational Linguistics: Stroudsburg, PA, USA, 2004; pp. 74–81. Available online: https://aclanthology.org/W04-1013.pdf (accessed on 5 July 2025).
- Zhang, T.; Kishore, V.; Wu, F.; Weinberger, K.Q.; Artzi, Y. BERTScore: Evaluating Text Generation with BERT. In Proceedings of the International Conference on Learning Representations (ICLR 2020), Addis Ababa, Ethiopia, 26–30 April 2020; OpenReview: Alameda, CA, USA, 2020. Available online: https://openreview.net/forum?id=SkeHuCVFDr (accessed on 22 July 2025).
- Tsai, C.-C.; Lin, Y.-X.; Lee, C.-I. Enhancing education quality: Exploring teachers’ attitudes and intentions towards intelligent MR devices. Eur. J. Educ. 2024, 59, e12692. [Google Scholar] [CrossRef]
Paper (Source) | Field of Study | Applied Method | Main Objective and Outcome |
---|---|---|---|
O’Mara-Eves et al. (2015) [11] | Systematic Literature Review (SR) | Text mining-based literature selection techniques | Expected improvement in the quality of grammatical particles in research papers and reduction in review time |
Gupta et al. (2020) [12] | Financial Data Analysis | Various text mining techniques | Evaluation of the applicability of text mining in the financial industry |
Gupta et al. (2022) [13] | Materials Engineering Paper Analysis | Domain-specific BERT model (MatSciBERT) | Improvement in data extraction accuracy in materials engineering papers |
Taha et al. (2024) [14] | Text Classification Technology | Comprehensive review and experimental analysis of text classification algorithms | Latest text classification techniques and application examples |
Shin et al. (2021) [15] | Local Government Public Data | Text mining (keyword frequency analysis, topic analysis) | Understanding the status and characteristics of data openness |
Han, S. (2025) [16] | Public Records and News Articles | Keyword analysis, topic modeling | Analysis of major issues and social perceptions in news articles |
Lee, J.-S.; Jung, J.-H. (2025) [17] | Public Design Research | Frequency analysis, keyword network analysis | Analysis of the relationship between main concepts and keywords in public design research |
This paper | Educational Administration Document Processing | Text mining-based knowledge-distilled LLM | Automation of school administration documents and establishment of an efficient system |
Block | Description | Key Outputs/Inputs |
---|---|---|
Administrative Document | Original PDF documents from real-world administrative use cases | Raw text |
Preprocessing | OCR-based parsing and segmentation of questions and answers | Cleaned text |
Dataset | (a) Instruction–Output pairs; (b) instruction-only prompts | (a) Teacher SFT, KD training and evaluation data |
Fine-tuned Teacher | LoRA-based or locally SFT-completed model | High-quality responses |
Distilled Dataset | Instruction–Output pairs generated by the teacher model | Student model training data |
Student Model | Parameter-optimized lightweight model (e.g., 1B~3B) | Model capacity |
Trained Student | Sequential KD-trained model | Final output |
Evaluation Dataset | Real-world public query samples | Evaluation metrics (ROUGE, BERTScore) |
Measure Performance | Comparative analysis between teacher and student models | Performance evaluation report (quantitative + qualitative) |
Specifications | |
---|---|
Model Inference Environment |
|
Metric | Percentage (%) | Description |
---|---|---|
Token Accuracy [32] | 92.36% | The percentage of tokens that are identical to or semantically consistent with the original text. |
Completed Sentence Rate | 97.19% | The percentage of the model’s output that matches the sentence-ending structure, relative to the original text (100%). |
Sentence Naturalness [33] | 99.10% | The fluency score based on the language model’s Perplexity, expressed as a percentage converted from the original text (100%). |
Format Suitability | 92.85% | The reproduction rate of official and administrative document format elements (e.g., title, item, label, date, and amount notation), as a percentage relative to the original text. |
Average Usage (%) | Teacher Model Training | Teacher Model KD | Student Model Training | Student Model Inference |
---|---|---|---|---|
CPU | 0.34% | 2.86% | 0.06% | 0.07% |
RAM | 3.15% | 3.16% | 3.2% | 3.38% |
GPU | 64.51% (126,804 MB) | 36.92% (72,575.84 MB) | 10.80% (4709.3 MB) | 29.28% (4583.11 MB) |
Model | #Params | Method | BLEU | ROUGE-1 | ROUGE-L | BERT_Score |
---|---|---|---|---|---|---|
llama-3.2-Korean-Blossom | 70B | Fine-tuned teacher | 97.20 | 99.05 | 99.04 | 98.29 |
3B | Vanilla Student | 87.50 | 95.04 | 94.68 | 96.09 | |
3B | OP-LLM-SA | 94.30 | 98.09 | 98.18 | 98.55 | |
llama-3.2-instruct | 3B | Fine-tuned teacher | 58.54 | 69.42 | 68.41 | 86.78 |
1B | Vanilla Student | 43.57 | 76.15 | 73.73 | 86.36 | |
1B | OP-LLM-SA | 45.31 | 77.47 | 76.04 | 86.94 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hong, Y. Design and Evaluation of Knowledge-Distilled LLM for Improving the Efficiency of School Administrative Document Processing. Electronics 2025, 14, 3860. https://doi.org/10.3390/electronics14193860
Hong Y. Design and Evaluation of Knowledge-Distilled LLM for Improving the Efficiency of School Administrative Document Processing. Electronics. 2025; 14(19):3860. https://doi.org/10.3390/electronics14193860
Chicago/Turabian StyleHong, Younhee. 2025. "Design and Evaluation of Knowledge-Distilled LLM for Improving the Efficiency of School Administrative Document Processing" Electronics 14, no. 19: 3860. https://doi.org/10.3390/electronics14193860
APA StyleHong, Y. (2025). Design and Evaluation of Knowledge-Distilled LLM for Improving the Efficiency of School Administrative Document Processing. Electronics, 14(19), 3860. https://doi.org/10.3390/electronics14193860