System Design and Evaluation of RAG-Enhanced Digital Humans in Design Education: Analyzing Cognitive Load and Instructional Efficiency
Abstract
1. Introduction
1.1. Research Background and Significance
1.2. Research Questions and Motivation
- (1)
- System Design and Integration (RQ1): How can RAG technology and digital human interfaces be effectively synthesized to create a responsive, intelligent teaching system tailored to the specific requirements of design education?
- (2)
- Educational Impact (RQ2): To what extent does this integrated system influence university design students’ learning outcomes and classroom engagement compared to traditional instructional methods?
- (3)
- Mechanism of Support (RQ3): In what ways does the RAG component enhance the digital human’s capacity to address domain-specific inquiries and provide context-aware scaffolding in design learning scenarios?
2. Literature Review
2.1. Potential and Applications of RAG Technology in Education
2.2. Applications of Digital Humans in Design Education
2.3. Research Gaps and Contributions of This Study
3. Research Methodology
3.1. System Design and Implementation: Synthesizing RAG and Digital Humans (RQ1)
3.1.1. RAG Model Selection and System Construction
3.1.2. Digital Human Design: Implementation Based on LiveTalking
3.1.3. System Integration and Local Deployment
3.1.4. Pedagogical Framework and Curriculum Integration
3.2. Experimental Design and Implementation
3.2.1. Experimental Subjects and Setting
3.2.2. Group Assignment and Control
3.2.3. Experimental Procedure and Data Collection
3.3. Data Analysis Methods
4. Results
4.1. Quantitative Analysis: Assessing Educational Impact (RQ2)
4.2. Qualitative Analysis: Unpacking the Mechanism of Support (RQ3)
“Instead of flipping through pages for ages to find the definition of ‘Art Nouveau,’ I got a concise, context-aware answer instantly. It saved my brain power for actually understanding the concept rather than just searching for it.”(Student C, University A)
“It definitely made studying less boring. Having the avatar ‘look’ at you and respond felt like a real tutoring session. It pushed me to ask more questions just to see how it would react.”(Student F, University B)
“In class, I’m afraid of looking stupid if I ask about a simple term. With the system, I could ask the same definition three times until I truly got it. It judged my questions, but it didn’t judge me.”(Student D, University C)
“It’s great for facts, but less for interpretation. If I asked ‘Why does this design feel sad?’, the answer was a bit generic. Also, sometimes the 5-s thinking pause made me wonder if it crashed.”(Student H, University A)
5. Discussion
5.1. Interpreting the Findings Through Theoretical Lenses
5.2. Theoretical and Empirical Contributions
5.3. Practical Strategies for Using RAG-Enabled Digital Humans in Education
5.4. Research Limitations
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Appendix A
- import streamlit as stfrom sentence_transformers import SentenceTransformerimport faissimport numpy as npimport requestsimport jsonimport time# --- Configurations ---# Define the names/paths of the key models and dataEMBEDDING_MODEL_NAME = ’BAAI/bge-large-zh’ # Specify the embedding modelKNOWLEDGE_INDEX_PATH = "design_knowledge_index.faiss" # Path to the FAISS indexKNOWLEDGE_EMBEDDINGS_PATH = "design_knowledge_embeddings.npy" # Path to the knowledge base embeddingsKNOWLEDGE_TEXTS_PATH = "design_knowledge_texts.npy" # Path to the knowledge base texts
- OPENWEBUI_API_URL = ’http://localhost:8080/api/v1/chat’ # URL for OpenWebUI APIOPENWEBUI_MODEL_NAME = ’qwen’ # Specify the language model name (must match OpenWebUI)
- STABLE_DIFFUSION_API_URL = ’http://localhost:7860/sdapi/v1/txt2img’ # URL for Stable Diffusion API (optional image generation)LIVETALKING_STREAM_URL = "http://localhost:8000/video_feed" # URL for LiveTalking video stream# --- Load Models and Data ---@st.cache_resource # Cache the models to avoid reloadingdef load_models():"""Loads the embedding model, FAISS index, and knowledge base."""embedding_model = SentenceTransformer(EMBEDDING_MODEL_NAME)index = faiss.read_index(KNOWLEDGE_INDEX_PATH)knowledge_base = np.load(KNOWLEDGE_EMBEDDINGS_PATH)knowledge_texts = np.load(KNOWLEDGE_TEXTS_PATH)return embedding_model, index, knowledge_base, knowledge_texts
- embedding_model, index, knowledge_base, knowledge_texts = load_models()# --- Functions ---def rag_generate_with_openwebui(prompt):"""Generates text using the RAG pipeline and OpenWebUI."""headers = {’Content-Type’: ’application/json’}data = {"model": OPENWEBUI_MODEL_NAME,"messages": [{"role": "user", "content": prompt}],"stream": False}try:response = requests.post(OPENWEBUI_API_URL, headers=headers, data=json.dumps(data), timeout=10)response.raise_for_status() # Raise HTTPError for bad responses (4xx or 5xx)return response.json()[’choices’][0][’message’][’content’]except requests.exceptions.RequestException as e:st.error(f"Failed to communicate with the language model: {e}")return None
- def rag_pipeline(query):"""Retrieval-Augmented Generation pipeline."""query_embedding = embedding_model.encode(query) # Encode the query using the embedding modelquery_embedding = np.array([query_embedding]).astype(’float32’) # Convert to a numpy arrayD, I = index.search(query_embedding, k=3) # Search the FAISS index for the top 3 relevant documentscontext = [knowledge_texts[i] for i in I[0]] # Retrieve the context texts
- prompt = f"Answer the question based on the following background knowledge:\n{’ ’.join(context)}\nQuestion: {query}" # Construct the promptreturn rag_generate_with_openwebui(prompt) # Generate the answer
- def generate_lecture_content(topic):"""Generates lecture content for a given topic using the RAG pipeline."""prompt = f"Explain the following design topic in detail: {topic}"return rag_generate_with_openwebui(prompt)
- def generate_image(prompt):"""Generates an image using Stable Diffusion (optional)."""headers = {’Content-Type’: ’application/json’}payload = {"prompt": prompt,"negative_prompt": "ugly, deformed, distorted","steps": 20,"sampler_index": "DPM++ 2M Karras"}try:response = requests.post(STABLE_DIFFUSION_API_URL, headers=headers, data=json.dumps(payload), timeout=60)response.raise_for_status()r = response.json()for i in r[’images’]:return f"data:image/png;base64,{i}"except requests.exceptions.RequestException as e:st.error(f"Failed to generate image: {e}")return None# --- Streamlit App ---st.title("AI-Powered Design Teaching Assistant")# --- Sidebar Controls ---with st.sidebar:st.header("System Controls")if st.button("Start Automated Lecture"):st.session_state.lecture_mode = Truest.session_state.lecture_index = 0st.session_state.lecture_paused = False
- if st.session_state.get("lecture_mode", False):col1, col2 = st.columns(2)if col1.button("Pause Lecture"):st.session_state.lecture_paused = Trueif col2.button("Resume Lecture"):st.session_state.lecture_paused = False
- col3, col4 = st.columns(2)if col3.button("Skip Topic"):st.session_state.lecture_index += 1st.session_state.lecture_paused = Falseif col4.button("Repeat Topic"):st.session_state.lecture_paused = False
- if st.session_state.lecture_index > 0 and st.button("Previous Topic"):st.session_state.lecture_index -= 1st.session_state.lecture_paused = False
- if st.button("End Lecture"):st.session_state.lecture_mode = Falsest.session_state.lecture_index = 0st.session_state.lecture_paused = False# --- Main Interface ---col_video, col_content = st.columns([0.7, 0.3])
- with col_video:st.header("Digital Human Instructor")st.markdown(f’<iframe src="{LIVETALKING_STREAM_URL}" height="480" width="640" frameborder="0" scrolling="no"></iframe>’, unsafe_allow_html=True)
- with col_content:st.header("Interaction & Feedback")query = st.text_input("Ask a question:")
- if st.button("Ask") and query:with st.spinner("Thinking…"):response = rag_pipeline(query)if response:st.write(f"**Instructor:** {response}")
- st.subheader("Provide Feedback")feedback = st.text_area("Your feedback:")if st.button("Submit Feedback") and feedback:# TODO: Implement feedback submission mechanism (e.g., store in database, send via email)st.success("Feedback submitted!")
- st.header("Automated Lecture")if st.session_state.get("lecture_mode", False):if st.session_state.lecture_index < len(COURSE_OUTLINE):current_topic = COURSE_OUTLINE[st.session_state.lecture_index]st.subheader(f"Current Topic: {current_topic}")
- if not st.session_state.get("lecture_paused", False):with st.spinner(f"Lecturing on: {current_topic}"):lecture_script = generate_lecture_content(current_topic)if lecture_script:st.write(f"**Instructor:** {lecture_script}")
- image_prompt = f"design illustration for {current_topic}"image_url = generate_image(image_prompt)if image_url:st.image(image_url, caption=f"Image related to {current_topic}", use_column_width=True)
- st.session_state.lecture_index += 1else:st.info("Lecture Paused")else:st.success("Lecture Completed!")st.session_state.lecture_mode = False# --- State Initialization ---if "lecture_mode" not in st.session_state:st.session_state.lecture_mode = Falseif "lecture_index" not in st.session_state:st.session_state.lecture_index = 0if "lecture_paused" not in st.session_state:st.session_state.lecture_paused = False# --- Example COURSE_OUTLINE ---COURSE_OUTLINE = ["Introduction to Modern World Design","The Bauhaus Movement"
References
- Cross, N. Design Thinking: Understanding How Designers Think and Work; Berg: Oxford, UK, 2011; pp. 1–234. [Google Scholar]
- Almufarreh, A.; Arshad, M. Promising emerging technologies for teaching and learning: Recent developments and future challenges. Sustainability 2023, 15, 6917. [Google Scholar] [CrossRef]
- Laurillard, D. Teaching as a Design Science: Building Pedagogical Patterns for Learning and Technology; Routledge: London, UK, 2012; pp. 1–224. [Google Scholar]
- Waldrop, M.M. The science of teaching science. Nature 2015, 523, 272. [Google Scholar] [CrossRef] [PubMed]
- Neshaei, S.P.; Tashkovska, M.; Mejia-Domenzain, P.; Wambsganss, T.; Käser, T. User-centric Reflective Writing Assistance: Leveraging RAG for Enhanced Personalized Support. In Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 26 April–1 May 2025; pp. 1–8. [Google Scholar]
- Yu, Y.; Liang, M.; Yin, M.; Lu, K.; Du, J.; Xue, Z. Unsupervised Multimodal Graph Contrastive Semantic Anchor Space Dynamic Knowledge Distillation Network for Cross-Media Hash Retrieval. In Proceedings of the 2024 IEEE 40th International Conference on Data Engineering (ICDE), Utrecht, The Netherlands, 13–16 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 4699–4708. [Google Scholar]
- Kolb, D.A. Experiential Learning: Experience as the Source of Learning and Development, 2nd ed.; FT Press: Upper Saddle River, NJ, USA, 2015; pp. 1–400. [Google Scholar]
- Zhou, X.; Kim, S.; Chen, Y. Generative AI visual creativity system combined with knowledge retrieval. J. Comput. Methods Sci. Eng. 2025, 14727978251346065. [Google Scholar] [CrossRef]
- Notarangelo, L.D.; Kim, M.S.; Walter, J.E.; Lee, Y.N. Human RAG mutations: Biochemistry and clinical implications. Nat. Rev. Immunol. 2016, 16, 234–246. [Google Scholar] [CrossRef]
- Arslan, M.; Ghanem, H.; Munawar, S.; Cruz, C. A Survey on RAG with LLMs. Procedia Comput. Sci. 2024, 246, 3781–3790. [Google Scholar] [CrossRef]
- Bailenson, J.N. Virtual interpersonal touch: Haptic interaction and social presence in immersive virtual environments. Presence Teleoper. Virtual Environ. 2006, 15, 586–603. [Google Scholar]
- Sung, E.C.; Han, D.I.D.; Bae, S.; Kwon, O. What drives technology-enhanced storytelling immersion? The role of digital humans. Comput. Hum. Behav. 2022, 132, 107246. [Google Scholar] [CrossRef]
- Demirel, H.O.; Ahmed, S.; Duffy, V.G. Digital human modeling: A review and reappraisal of origins, present, and expected future methods for representing humans computationall. Int. J. Hum.–Comput. Interact. 2022, 38, 897–937. [Google Scholar] [CrossRef]
- Mikropoulos, T.A.; Natsis, A. Educational virtual environments: A ten-year review of empirical research (1999–2009). Comput. Educ. 2011, 56, 769–780. [Google Scholar] [CrossRef]
- Woodgate, D. Immersive spatial narratives as a framework for augmenting creativity in foresight-based learning systems. Horiz. Int. J. Learn. Futures 2019, 27, 57–71. [Google Scholar] [CrossRef]
- Marinelli, A.; Iannacci, F.; Papile, F.; Diamanti, M.V.; Sponchioni, M.; Del Curto, B. Higher education classroom of the future: An EU-funded project integrating physical lectures, virtual reality, and artificial intelligence. In Proceedings of the 13th International Materials Education Symposium, Cambridge, UK, 4–5 April 2024; p. 38. [Google Scholar]
- Pedro, F.; Subosa, M.; Rivas, A.; Valverde, P. Artificial Intelligence in Education: Challenges and Opportunities for Sustainable Development; United Nations Educational, Scientific and Cultural Organization (UNESCO): Paris, France, 2019; pp. 1–59. [Google Scholar]
- Lang, Q.; Wang, M.; Yin, M.; Liang, S.; Song, W. Transforming education with generative AI (GAI): Key insights and future prospects. IEEE Trans. Learn. Technol. 2025, 18, 230–242. [Google Scholar] [CrossRef]
- Filipović, A.M.; Fastić-Pajk, I.; Puljiz, H.; Šabić, I. AI-Driven Customization of Adaptive Learning Content: Lessons from Applying LLMs. In Proceedings of the Digital Transformation in Education and Artificial Intelligence Application: Third International Conference, MoStart 2025, Mostar, Bosnia and Herzegovina, 23–25 April 2025; Volume 202, p. 67. [Google Scholar]
- Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; Wainwright, C.L.; Mishkin, P.; Lowe, R. Training language models to follow instructions with human feedback. Adv. Neural Inf. Process. Syst. 2022, 35, 27730–27744. [Google Scholar]
- Sadat, M.; Zhou, Z.; Lange, L.; Araki, J.; Gundroo, A.; Wang, B.; Menon, R.; Parvez, R.; Feng, Z. DelucionQA: Detecting hallucinations in domain-specific question answering. In Findings of the Association for Computational Linguistics: EMNLP 2023; Association for Computational Linguistics: Singapore, 2023; pp. 822–835. [Google Scholar]
- Izacard, G.; Lewis, P.; Grave, E.; Martinet, L.; Kessler, G.; Usunier, N.; Joulin, A. Distilling knowledge from reader to retriever for question answering. arXiv 2021, arXiv:2101.00408. [Google Scholar]
- Henkel, O.; Levonian, Z.; Li, C.; Postle, M. Retrieval-augmented generation to improve math question-answering: Trade-offs between groundedness and human preference. In Proceedings of the 17th International Conference on Educational Data Mining, Paris, France, 1–4 July 2024; pp. 315–320. [Google Scholar]
- Thway, M.; Recatala-Gomez, J.; Lim, F.S.; Hippalgaonkar, K.; Ng, L.W. Harnessing GenAI for Higher Education: A Study of a Retrieval Augmented Generation Chatbot’s Impact on Human Learning. arXiv 2024, arXiv:2406.07796. [Google Scholar] [CrossRef]
- Chen, L.; Chen, P.; Lin, Z. Artificial intelligence in education: A review. IEEE Access 2020, 8, 75264–75278. [Google Scholar] [CrossRef]
- Ali, S.; Fatima, F.; Hussain, J.; Qureshi, M.I.; Fatima, S.; Zahoor, A. Exploring Student’s Experiences and Problems in Online Teaching and Learning During COVID-19 and Improvement of Current LMS Through Human-Computer Interaction (HCI) Approaches. Int. J. Interact. Mob. Technol. 2023, 17, 4–21. [Google Scholar] [CrossRef]
- Kahl, S.; Löffler, F.; Maciol, M.; Ridder, F.; Schmitz, M.; Spanagel, J.; Schilling, M. Enhancing AI Tutoring in Robotics Education: Evaluating the Effect of Retrieval-Augmented Generation and Fine-Tuning on Large Language Models; University of Münster: Münster, Germany, 2024. [Google Scholar]
- Holly, M.; Pirker, J.; Resch, S.; Brettschuh, S. Designing VR experiences–expectations for teaching and learning in VR. Educ. Technol. Soc. 2021, 24, 107–119. [Google Scholar]
- Dede, C. Immersive interfaces for engagement and learning. Science 2009, 323, 66–69. [Google Scholar] [CrossRef]
- Selwyn, N. Should Robots Replace Teachers? AI and the Future of Education; Polity Press: Cambridge, UK, 2021; pp. 1–232. [Google Scholar]
- Zhou, X.; Kim, S.; Wang, Y.; Zhang, K. Beyond sparsity: An empirical study of structured collaboration in modular AI. Neurocomputing 2025, 657, 131616. [Google Scholar] [CrossRef]
- Zuboff, S. The Age of Surveillance Capitalism: The Fight for a Human Future at the New Frontier of Power; PublicAffairs: New York, NY, USA, 2019; pp. 1–704. [Google Scholar]
- Yao, Y.; González-Vélez, H. AI-Powered System to Facilitate Personalized Adaptive Learning in Digital Transformation. Appl. Sci. 2025, 15, 4989. [Google Scholar] [CrossRef]
- Elston, D.M. The novelty effect. J. Am. Acad. Dermatol. 2021, 85, 565–566. [Google Scholar] [CrossRef]
- Lipku. LiveTalking. Available online: https://github.com/lipku/LiveTalking (accessed on 19 December 2023).
- Luccioni, A.S.; Akerman, G.; Vishnyakova, O.; Akimov, A.; Astakhov, D.; Obukhov, V. Open LLM leaderboard. Hugging Face. Available online: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard (accessed on 1 July 2024).
- Yang, A.; Li, A.; Yang, B.; Zhang, B.; Hui, B.; Zheng, B.; Yu, B.; Gao, C.; Huang, C.; Lv, C.; et al. Qwen3 technical report. arXiv 2025, arXiv:2505.09388. [Google Scholar] [CrossRef]
- Grattafiori, A.; Dubey, A.; Jauhri, A.; Pandey, A.; Kadian, A.; Al-Dahle, A.; Letman, A.; Mathur, A.; Schelten, A.; Vaughan, A.; et al. The llama 3 herd of models. arXiv 2024, arXiv:2407.21783. [Google Scholar] [CrossRef]
- Yang, M.; Diao, M.; Luo, J.; Shen, W.; Zhang, C. GLM-4 Based Method for Automatic Construction of Content Graph. IEEE Access 2025, 13, 197300–197311. [Google Scholar] [CrossRef]
- Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar] [CrossRef]
- Guan, K.; Cao, Q.; Sun, Y.; Wang, X.; Song, R. BSharedRAG: Backbone shared retrieval-augmented generation for the E-commerce domain. arXiv 2024, arXiv:2409.20075. [Google Scholar]
- Tang, S.; Liu, A.T.; Xie, J.; Zhou, Z.; Dong, L.; Smith, N.A.; Wei, F. E5: A Self-supervised Framework for Aligning Embeddings across Languages. arXiv 2024, arXiv:2212.03533. [Google Scholar]
- GanymedeNil. Text2vec-Base-Chinese. Available online: https://huggingface.co/GanymedeNil/text2vec-base-chinese (accessed on 25 June 2024).
- Baek, J.; Hussain, A.; Liu, D.; Vincent, N.; Kim, L.H. Open WebUI: An Open, Extensible, and Usable Interface for AI Interaction. arXiv 2025, arXiv:2510.02546. [Google Scholar] [CrossRef]
- Li, J.; Zhang, J.; Bai, X.; Zheng, J.; Zhou, J.; Gu, L. Er-nerf++: Efficient region-aware neural radiance fields for high-fidelity talking portrait synthesis. Inf. Fusion 2024, 110, 102456. [Google Scholar] [CrossRef]
- Liang, C.; Wang, Q.; Chen, Y.; Tang, M. Wav2Lip-HR: Synthesising clear high-resolution talking head in the wild. Comput. Animat. Virtual Worlds 2024, 35, e2226. [Google Scholar] [CrossRef]
- Dong, Z.; Liu, X.; Chen, B.; Polak, P.; Zhang, P. Musechat: A conversational music recommendation system for videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 12775–12785. [Google Scholar]
- An, L. Ultralight-Digital-Human. Available online: https://github.com/anliyuan/Ultralight-Digital-Human (accessed on 10 October 2024).
- Mandal, S.; Ghosh, B.; Chakraborty, S.; Naskar, R. Can Deepfakes Mimic Human Emotions? A Perspective on Synthesia Videos. In Proceedings of the TENCON 2024-2024 IEEE Region 10 Conference (TENCON), Singapore, 1–4 December 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 306–309. [Google Scholar]
- Contreras, I.; Hossfeld, S.; de Boer, K.; Wiedler, J.T.; Ghidinelli, M. Revolutionising faculty development and continuing medical education through AI-generated videos. J. CME 2024, 13, 2434322. [Google Scholar] [CrossRef]
- Popov, V.; Vovk, I.; Gogoryan, V.; Sadekova, T.; Kudinov, M. Grad-tts: A diffusion probabilistic model for text-to-speech. In International Conference on Machine Learning; PMLR: New York, NY, USA, 2021; pp. 8599–8608. [Google Scholar]
- Ferguson, C.; van den Broek, E.L.; van Oostendorp, H. AI-induced guidance: Preserving the optimal zone of proximal development. Comput. Educ. Artif. Intell. 2022, 3, 100089. [Google Scholar] [CrossRef]
- Waller, R.; Dahle-Huff, K. More Knowledgeable Others: Exploring Professional Development of Rural Reading Specialists. Educ. Res. Theory Pract. 2023, 34, 68–74. [Google Scholar]
- Mayer, R.E. The past, present, and future of the cognitive theory of multimedia learning. Educ. Psychol. Rev. 2024, 36, 8. [Google Scholar] [CrossRef]
- Sweller, J.; van Merriënboer, J.J.; Paas, F. Cognitive architecture and instructional design: 20 years later. Educ. Psychol. Rev. 2019, 31, 261–292. [Google Scholar] [CrossRef]
- Vygotsky, L.S. Mind in Society: The Development of Higher Psychological Processes; Harvard University Press: Cambridge, MA, USA, 1978; pp. 1–159. [Google Scholar]
- Pelikan, H.; Hofstetter, E. Managing Delays in Human-Robot Interaction. ACM Trans. Comput.-Hum. Interact. 2023, 30, 1–42. [Google Scholar] [CrossRef]
- Radford, A.; Kim, J.W.; Xu, C.; McNeil, G.; Sridhar, A.; Sutskever, I.; Tieleman, T. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 8748–8763. [Google Scholar]







| Model | MMLU | IFEVAL | Resource Usage |
|---|---|---|---|
| Qwen3 | 81.38 | 82.3 | Medium |
| Llama-3 | 68.4 | 76.8 | Medium |
| GLM-4 | 86.5 | 87.6 | High |
| Model | Embedding Dimensions | Average Recall Rate | Model Size |
|---|---|---|---|
| BGE-Large-Zh | 1024 | High (80%+) | Large (1.3 GB) |
| E5-Large-v2-CN | 1024 | High (78%+) | Large (1.3 GB) |
| Ganymede Base | 768 | Medium (70%+) | Medium (500 MB) |
| Platform | Realism | Customization Level | Interactivity | Integration | Cost |
|---|---|---|---|---|---|
| LiveTalking | High (model-dependent) | Very High | Medium | High (RTMP/WebRTC) | Low (open-source) |
| Synthesia | Medium | Medium | Low | Medium (API) | High (subscription) |
| Hour One | Medium | Medium | Low | Medium (API) | High (subscription) |
| Characteristic | Experimental (n = 75) | Control (n = 75) | Test Statistic | p-Value |
|---|---|---|---|---|
| Age (Mean ± SD) | 21.4 ± 1.2 | 21.2 ± 1.3 | 0.328 | |
| Gender (Female %) | 48 (64.0%) | 45 (60.0%) | 0.612 | |
| Prior GPA (0–4.0) | 3.42 ± 0.35 | 3.39 ± 0.38 | 0.614 | |
| Tech Familiarity (1–5) | 3.8 ± 0.7 | 3.7 ± 0.8 | 0.413 |
| Group | Variable | n | Shapiro–Wilk Statistic (W) | p-Value |
|---|---|---|---|---|
| Exp | Pre-test Score | 75 | 0.978 | 0.245 |
| Ctrl | Pre-test Score | 75 | 0.982 | 0.355 |
| Exp | Post-test Score | 75 | 0.975 | 0.158 |
| Ctrl | Post-test Score | 75 | 0.969 | 0.062 |
| Exp | Classroom Engagement | 75 | 0.970 | 0.633 |
| Ctrl | Classroom Engagement | 75 | 0.961 | 0.430 |
| Indicator | Group | n | Mean (M) | Standard Deviation (SD) |
|---|---|---|---|---|
| Pre-test Score | Exp | 75 | 70.15 | 8.42 |
| Ctrl | 75 | 69.88 | 7.95 | |
| Post-test Score | Exp | 75 | 86.42 | 6.85 |
| Ctrl | 75 | 78.10 | 7.64 | |
| classroom Engagement | Exp | 75 | 15.84 | 3.41 |
| Ctrl | 75 | 11.20 | 4.02 | |
| Mental Effort Rating | Exp | 75 | 4.32 | 1.65 |
| Ctrl | 75 | 7.15 | 1.78 |
| Scores Indicator | t | df | p | Mean Difference | Standard Error |
|---|---|---|---|---|---|
| Pre-test Score | 0.201 | 148 | 0.841 | 0.27 | 2.353 |
| Measure | t | df | p | Mean Diff. | 95% CI of Diff. | Standard Error | Cohen’s d |
|---|---|---|---|---|---|---|---|
| Post-test Score | 7.01 | 148 | <0.001 | 8.32 | [6.01, 10.63] | 1.18 | 1.14 |
| Measure | t | df | p | Mean Diff. | 95% CI of Diff. | Standard Error | Cohen’s d |
|---|---|---|---|---|---|---|---|
| Classroom Engagement | 8.54 | 148 | <0.001 | 4.64 | [3.58, 5.70] | 0.54 | 1.39 |
| Measure | Post-Test Score |
|---|---|
| Classroom Engagement | 0.587 ** |
| Model | Variables | B | SE | β | t | p |
|---|---|---|---|---|---|---|
| 1 | (Constant) | 22.885 | 5.89 | 4.21 | <0.001 | |
| Group | 6.073 | 0.77 | 0.465 | 7.82 | <0.001 | |
| Pre-test Score | 0.546 | 0.054 | 0.582 | 9.94 | <0.001 |
| Primary Theme | Sub-Themes/Codes | Frequency | Representative Quote |
|---|---|---|---|
| 1. Information Accessibility (Cognitive Load) |
| 27 (90%) | “It acted like a smart index. I didn’t have to manage a dozen tabs; the relevant knowledge was just delivered to me.” |
| 2. Engagement & Motivation (Social Presence) |
| 25 (83%) | “Having a ‘face’ to talk to changed the vibe. It felt like studying with a partner rather than alone.” |
| 3. Personalized Support (Psychological Safety) |
| 23 (76%) | “I could fix my knowledge gaps privately without disrupting the whole class.” |
| 4. System Limitations (Technical/Pedagogical) |
| 18 (60%) | “The answers were technically correct but lacked the ‘soul’ or unique insight a real professor gives.” |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Zhou, X.; Zhao, S.; Wu, P.; Chen, Y. System Design and Evaluation of RAG-Enhanced Digital Humans in Design Education: Analyzing Cognitive Load and Instructional Efficiency. Appl. Sci. 2026, 16, 1068. https://doi.org/10.3390/app16021068
Zhou X, Zhao S, Wu P, Chen Y. System Design and Evaluation of RAG-Enhanced Digital Humans in Design Education: Analyzing Cognitive Load and Instructional Efficiency. Applied Sciences. 2026; 16(2):1068. https://doi.org/10.3390/app16021068
Chicago/Turabian StyleZhou, Xiaofei, Shiru Zhao, Pengjun Wu, and Yan Chen. 2026. "System Design and Evaluation of RAG-Enhanced Digital Humans in Design Education: Analyzing Cognitive Load and Instructional Efficiency" Applied Sciences 16, no. 2: 1068. https://doi.org/10.3390/app16021068
APA StyleZhou, X., Zhao, S., Wu, P., & Chen, Y. (2026). System Design and Evaluation of RAG-Enhanced Digital Humans in Design Education: Analyzing Cognitive Load and Instructional Efficiency. Applied Sciences, 16(2), 1068. https://doi.org/10.3390/app16021068

