Submitted:
15 July 2025
Posted:
16 July 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
- A novel BPMN-to-MAS transformation methodology that converts pedagogical workflows into executable MAS, bridging formal process modeling with AI-driven education.
- Integration of RAG technology to ensure accurate, contextually grounded language instruction while mitigating LLM hallucinations.
- Implementation of a complete Luxembourgish learning platform (A1—B2) with React/ FastAPI frontend, LangGraph core, ChromaDB vector store, and STT/TTS pipelines.
- Empirical evaluation showing strong response accuracy (RAGAs: Context Relevancy = 0.87, Faithfulness = 0.82, Answer Relevancy = 0.85) and high learner satisfaction in a pilot (85.8% ease-of-use, 71.4% engagement).
- A generalizable framework for low-resource language education that combines formal process modeling, distributed AI agents, and knowledge-grounded generation.
2. Related Work
2.1. BPMN for Educational Process Modeling
2.2. Multi-Agent Systems in Education
2.3. Retrieval-Augmented Generation
2.4. LLM-Powered Language Learning Chatbots
2.5. Technologies for Low-Resource Languages
2.6. Research Gap
- Uses BPMN to specify pedagogical workflows,
- Orchestrates specialized LLM-powered agents via MAS,
- Grounds all content in vetted external knowledge with RAG,
- Incorporates real-time voice interaction (STT/TTS) for a low-resource language.
3. System Design and Architecture
- BPMN is used to define and structure the high-level learning workflow. Each BPMN task represents a discrete educational activity (e.g., “Grammar Practice”), providing a visual and organized representation of the learning journey.
- MAS executes the modeled processes. Each agent within the MAS is domain-specific, with specialized roles such as Conversational Agent, Grammar Agent, Reading Agent, and Listening Agent. These agents work autonomously yet collaboratively, each handling a distinct aspect of the learning experience.
- RAG equips each agent with access to relevant educational content. This is achieved through vector stores constructed from INL textbooks and OCR-processed materials, allowing agents to retrieve and generate accurate, knowledge grounding responses.
3.1. Core Agent Roles and Responsibilities
-
Communicator Agent: Serves as the primary interface between the system and the user. Its responsibilities include:
- Interacting with users in natural language.
- Recommending learning activities based on the user’s preferences, progress, and performance.
- Personalizing the learning path according to user goals and pacing.
-
Orchestrator Agent: Functions as the system’s manager, responsible for:
- Managing the overall workflow of the learning session.
- Retrieving and sequencing appropriate content for each session.
- Coordinating the actions of the specialized tutor agents to ensure cohesive learning progression.
-
Tracker Agent: Monitors and ensures the correct execution of the learning workflow. Its tasks include:
- Signaling the activation of Tutor Agents at the appropriate stages.
- Waiting for each agent to complete its task before initiating the next.
- Maintaining the flow and timing of the session until completion.
-
Specialized Tutor Agents: At the heart of the platform are the Tutors Agents, each designed to focus on a specific aspect of language learning. These agents reflect the structure found in Luxembourgish textbooks, where different learning activities are categorized to target distinct language skills:
- Conversational Agent: Focuses on spoken language skills through interactive conversation. Activities include Role-playing scenarios, Vocabulary-focused dialogues, and Real-time feedback for pronunciation and fluency.
- Reading Agent: Enhances comprehension and vocabulary acquisition through reading tasks, assists with reading accuracy and understanding, and encourages summarization and contextual vocabulary usage.
- Listening Agent: Builds listening comprehension through repetition and follow-up exercises, listening-based actions, and feedback to improve recognition and application of new vocabulary.
- QA (Question & Answer) Agent: Provides interactive tasks and exercises based on book-inspired formats such as Kombinéiert!, Notéiert!/Schreift!. It presents quiz-style questions, evaluates responses and delivers detailed explanations, and offers corrective and motivational feedback.
-
GrammarSummary Agent:Specializes in grammatical instruction by delivering concise and targeted grammar rules, supporting correct grammatical usage in user outputs.
- Human Tutor A teacher will be involved to ensure the linguistic accuracy and coherence of the learning materials. This includes verifying the alignment between text content, audio transcriptions, and visual illustrations. The teacher also validates the relevance and pedagogical sequencing of the content retrieved by the Orchestrator, ensuring it is appropriate for the user’s proficiency level and learning objectives.
3.1.1. Knowledge Provenance
3.1.2. Human Auditing
- The raw BPMN-defined sequence of activities,
- LangSmith logs of each agent’s decision context,
- The alignment between chunk metadata and delivered exercises.
3.1.3. Workflow Transparency
3.2. Explainability Mechanisms
3.2.1. LangSmith Role
- Explore complete prompt histories for each agent node,
- Inspect intermediate state variables and message payloads,
- Visualize conditional routing paths and loop iterations,
- Search and filter on specific student interactions or decision predicates.

4. BPMN to MAS Transformation
4.1. BPMN Modeling of Learning Workflows
4.1.1. Top-Level Orchestration Diagram (Figure 4)
| Pool / Lane | Role / Responsibility |
|---|---|
| User | Human learner interacting via the UI. |
| Communicator Agent | Retrieves profile data; proposes personalized learning paths; emits RAG queries. |
| Orchestrator Agent | Fetches and validates content; plans which tutor agents to invoke and in what order. |
| Tracker Agent | Drives step-by-step activation of tutor agents; logs completion or early-exit signals. |
| Tutor Agents | Swimlane for specialized tutors (Conversation, Reading, Listening, QA, GrammarSummary). |
- Start Event (Message) The User logs in, triggering a message start event in the Communicator lane.
-
User Data Retrieval (Service Tasks) In the Communicator lane, three service tasks are retrieved:
- UserProfile — personal details and learning objectives.
- LatestProgressFile — feedback from the previous session.
- CurriculumOutline — textbook TOC matching the user’s proficiency.
- Personalized Path Generation A service task builds a LearningPathRecommendation. A message flow delivers it to the User, and an exclusive gateway (“Accept?”) loops back for refinement until approval.
- Query Generation & Dispatch Once approved, the Communicator constructs a RAGQuery (including topic IDs and proficiency level) and sends it as a message to the Orchestrator.
- Content Retrieval & Validation The Orchestrator executes a VectorStoreLookup against ChromaDB, then sends the retrieved material to the Human Teacher for validation (message task) and awaits approval.
-
Workflow Planning A parallel gateway splits into two branches:
- Assign each content chunk to its appropriate Tutor Agent.
- Build the SequenceReport specifying agent invocation order.
Both branches join before proceeding. -
Report Emission Two message tasks sent:
- ContentReport→ Tracker (mapping agents to content).
- SequenceReport→ Tracker (ordered list of agents).
-
Tutor Invocation Loop In the Tracker lane:
- DetermineNextAgent via SequenceReport.
- Send StartSession message to that Tutor Agent.
- Wait (intermediate catch event) for EndSession or EarlyExit.
- Log progress (partial or complete).
Repeat until no agents remain. - End Event Once all sessions finish, the Tracker emits an end event. The UI displays the updated progress dashboard and may loop back to the Communicator for a new cycle.
4.1.2. Activity Diagrams for Tutor Agents
Example: Conversation Agent (Figure 5)
- Message Start: Catch StartSession from Tracker.
- Fetch Content: Load dialogue script and role definitions from ContentReport.
- Introduction: Outline session goals (e.g. focus on past-tense).
-
Role-Play Loop:
- Prompt user with their first line.
- Send spoken reply to STT; receive transcription.
-
Gateway G1 (Correct?):
- −
- If correct, advance to next line.
- −
- If incorrect, provide corrective feedback and loop back.
- Repeat until all turns complete.
- Wrap-Up: Summarize key vocabulary and structures; write progress fragment.
- Message End: Send EndSession + progress payload back to Tracker.
| Gateway ID | Condition & Action |
|---|---|
| G1 (Correct?) | IF pronunciation_error_count == 0 → advance to next dialogue turn; ELSE → invoke corrective feedback task and loop back. |
| G2 (All Turns Completed?) | IF turns_completed == total_turns → proceed to Wrap-Up; ELSE → return to Role-Play Loop. |
Overview of Other Tutor Agents
-
Reading Agent (Fig. A1): Presents text to read, checks pronunciation via STT, requests a spoken or written summary, evaluates comprehension, teaches new vocabulary, and loops until mastery.Gateway R1: IF summary_correct? → continue; ELSE → replay text + re-question.Gateway R2: IF comprehension_score > threshold → next activity; ELSE → vocabulary drill.
-
Listening Agent (Fig. A2): Plays audio clips, prompts learner reproduction, transcribes and evaluates responses, offers vocabulary tips, and loops for reinforcement.Gateway L1: IF transcription_accuracy > 80% → next clip; ELSE → replay clip.Gateway L2: IF vocab_usage_correct? → continue; ELSE → provide targeted vocabulary drill.
-
QA Agent (Fig. A3): Displays exercises (fill-in, MCQ), evaluates answers, provides hints on incorrect responses, and summarizes learning goals.Gateway Q1: IF answer == key → correct flow; ELSE → hint task + retry.Gateway Q2: IF retry_count > 2 → escalate to GrammarSummary Agent; ELSE → loop for another attempt.
-
GrammarSummary Agent (Fig. A4): Reviews previous grammar, elicits user questions, explains rules, engages in practice sentences, identifies errors, and closes with a concise rule summary.Gateway Gs1: IF user_asks_question → answer question; ELSE → present practice sentence.Gateway Gs2: IF error_count > 3 → trigger additional examples; ELSE → proceed to summary.

4.2. Mapping BPMN to MAS
Agent and Tool Nodes
Routers and Conditional Edges
Message Passing
Example: Communicator Routing
- Loop back to itself (Continue) if the learner requests adjustments.
- Invoke its communicator_call_tool node (Call tool) to re-fetch profile data.
- Transition to the Orchestrator node (Go orchestrator) once the recommendation is approved.
Handling Multiple User Inputs
4.3. Multi-Agent Architecture
- Communicator Agent: First interface with users, providing personalized recommendations based on learner profiles and progress
- Orchestrator Agent: Manages workflow, retrieves relevant content, and coordinates agent activation
- Tracker Agent: Monitors workflow execution and learner progress
-
Tutor Agents: Specialized agents for different learning aspects:
- −
- Conversational Agent: Facilitates speaking practice
- −
- Reading Agent: Guides reading comprehension
- −
- Listening Agent: Manages listening exercises
- −
- QA Agent: Handles interactive questions
- −
- Grammar Summary Agent: Provides grammatical explanations
- Human Validator: Reviews and approves generated content
4.4. LangGraph Implementation and Prompt Orchestration
LangGraph Architecture
- Nodes:
- Each node represents a computation phase, often an LLM-driven task executor. Nodes process user inputs, generate or transform text, invoke external tools (e.g., RAG lookups, STT/TTS), and update shared state.
- Edges:
- Unconditional edges define fixed sequences, while conditional edges evaluate predicates (e.g., “user accepted recommendation?”) to branch dynamically.
- Task Looping: Nodes may loop to themselves until a gateway condition is satisfied.
- Conditional Routing: Router nodes inspect state or outputs and select the correct outgoing edge.
- Persistent State Management: Message payloads and node states persist across turns, so each agent “remembers” prior context.
Prompt Engineering for Agent Behavior
- Clarity & Role Definition: “You are the Conversational Agent tasked with…”
- Stepwise Instructions: Numbered or bullet steps guide the model through its workflow.
- Contextual Anchoring: Inject RAG-retrieved content chunks to ground responses.
- Error Handling: Include conditional clauses (e.g., “If the user’s answer is incorrect, provide feedback and re-prompt”).
- Iterative Refinement: Collect performance metrics after each session and refine prompts to reduce ambiguity and hallucinations.
Integrating Prompts into Nodes
Example Prompt Templates
- Listing 1: Communicator Agent system message, showing role definition and basic RAG context setup.
-
Listing 2: Conversational Tutor Agent prompt, including:
- −
- Role Definition (“You are a Conversational Agent…”).
- −
- RAG Context Injection (e.g., “Thema: ‚Wéi heescht Dir?‘, Kategorie: ‚Dialogs…‘, Agent: ‚Conversational agent‘”).
- −
- Error-Handling Logic (e.g., “IF user_error THEN provide corrective feedback and re-prompt”).
Graph Compilation and Execution
- The __start__ node dispatches control to communicator.
- communicator interacts with the learner (loop/tool/orchestrator branches).
- orchestrator retrieves RAG content, validates with the teacher, and signals tracker.
- tracker sequentially activates each tutor agent (reader, listening, questionAnswering, grammarSummary), awaiting each EndSession.
- After all tutor nodes complete, tracker issues __end__, concluding the session.
4.5. Voice Integration: STT & TTS
4.5.1. Speech-to-Text (STT)
Word Error Rate (WER)
| Model | Pretraining | Fine-tuning Data | WER |
|---|---|---|---|
| wav2vec2-large-xlsr-53-842h-luxembourgish-14h | Multilingual (53 langs) | 842 h unlabelled + 14 h labelled | 28% |
| whisper_large_lb_ZLS_v4_38h | OpenAI Whisper base | 14 h → 38 h labelled Luxembourgish | 18% |
4.5.2. Text-to-Speech (TTS)
4.5.3. Integration into Multi-Agent Workflow
STT/TTS Integration
- wav2vec2-large-xlsr-53-842h-luxembourgish-14h: a multilingual model pre-trained on 53 languages and fine-tuned with 842 h of unlabelled plus 14 h of labelled Luxembourgish speech, which achieved a WER of 28%.
- whisper_large_lb_ZLS_v4_38h: OpenAI’s Whisper base model, further fine-tuned on 38 h of labelled Luxembourgish data by the Zentrum fir d’Lëtzebuerger Sprooch (ZLS), which achieved a superior WER of 18%.
5. RAG-Enhanced Knowledge Base
5.1. Why RAG for Low-Resource Languages
- Relevance: Retrieving domain-specific content (INL textbooks) tailored to each learner’s level.
- Accuracy: Anchoring generation in factual excerpts, bolstering learner trust.
- Pedagogical Alignment: Dynamically selecting material that matches Common European Framework of Reference for Languages (CEFR) aligned chapters and topics.
5.2. RAG Pipeline
5.2.1. Retrieval
-
Document Preparation:
- Scan INL textbooks (A1—B2) and convert pages to Markdown via GPT-4 Vision OCR [45].
- Clean and normalize text (remove headers/footers, correct OCR errors).
-
Chunking & Splitting: We employ agentic chunking to mirror textbook structure:
- Splitter Agent: Divides each topic into semantically coherent “learning blocks.”
- Organizer Agent: Groups blocks by chapter and topic, preserving pedagogical order.
- Embedding & Storage: Each chunk is embedded and stored in ChromaDB [44]. We selected bge-large-en-v1.5 after benchmarking on MTEB and our pilot RAGAs evaluation as the best trade-off between latency, relevance, and open-source licensing (see Section 3.3.3).
5.2.2. Generation
- Query Embedding & Matching: Learner queries or agent prompts are embedded and matched against stored vectors via cosine similarity to retrieve the top-k chunks.
- Contextual Response: Retrieved chunks are prepended to the LLM prompt (e.g., GPT-4), which generates the final answer, reflecting both the model’s internal knowledge and the verified textbook content.
- Explainability Tags: Each response includes semantic source metadata drawn from chunk fields: enabling learners and educators to verify content against original materials.
5.3. Embedding Model Selection
- Latency: Time to embed the full INL corpus (e.g., text-embedding-3-large completed in ≈53 s, while others averaged ≈3h).
- Relevance (RAGAs): Performance on context relevancy, faithfulness, and answer relevancy.
5.4. Evaluation with RAGAs
- Context Relevancy
- Context Precision
- Context Recall
- Faithfulness
- Answer Relevancy
- Answer Correctness
5.5. Building a Robust Knowledge Base
6. Implementation and Use Case
6.1. Technology Stack
- Frontend: React.js, renders the learner dashboard, chat interface, and course navigation, and streams audio via Web Audio API.
- Backend: FastAPI (Python), exposes REST and WebSocket endpoints for user authentication, agent orchestration, and real-time messaging.
- Core Agents: Implemented with LangGraph on top of LangChain, compiles BPMN-derived workflows into a stateful directed graph of TaskNodes and ToolNodes.
- RAG Vector Store: ChromaDB, stores pedagogically chunked INL content; queried via cosine-similarity retrievers.
- STT/TTS: OpenAI Whisper (whisper_large_lb_ZLS_v4_38h) for transcription; Coqui VITS (lb-de-fr-en-pt-coqui-vits-tts) for speech synthesis.
6.2. End-to-End Workflow
- Retrieves the user’s profile, progress, and curriculum metadata,
- Constructs and displays a personalized learning path in the React UI,
- Upon learner approval, emits a go_orchestrator event.
- Queries ChromaDB for the next topic’s content,
- Sends the raw material to a human teacher for quick validation (teacher-in-the-loop),
- Builds two reports: (i) validated content for tutor agents and (ii) the ordered list of agent tasks,
- Emits continue_to_tracker.
- Parses the sequence report and dispatches start signals to each Tutor Agent in turn,
- Listens for each agent’s completion or exit signals,
- Aggregates intermediate progress and updates the learner’s profile.
- It fetches its specific content from the Orchestrator’s report,
- Interacts with the learner via WebSocket streams (text + STT/TTS audio),
- Sends real-time feedback and performance metrics back to Tracker,
- Loops or branches as defined by the BPMN gateways.
6.3. Demonstration Highlights
- Learner dashboard flows in React,
- Chat-based dialogues powered by Conversational Agent,
- Listening exercises with real-time transcription,
- Grammar drills and Q&A sessions reflecting adaptive branching.
7. Evaluation
7.1. Response Accuracy with RAGAs
7.2. System Effectiveness and Learner Experience
- Ease of Interaction: 85.8% found the chatbot Very Easy or Easy.
- Satisfaction: 71.5% were Satisfied or Very Satisfied with contextual responses.
- Engagement: 71.4% rated the experience as Very engaging.
- Continued Use: 85.7% are Likely or Very Likely to continue using the system.
7.3. Usability and Pedagogical Alignment
- Responsive Interface: Login/dashboard, chat sessions, and progress tracking.
- Agent Workflows: Automatic sequencing of Conversational, Reading, Listening, Q&A, and Grammar agents via BPMN-defined flows.
- STT/TTS Integration: Whisper-based speech recognition (18% WER) and Coqui VITS TTS for immersive voice interaction.
7.4. Conclusion of the Evaluation
7.5. Limitations
- Model Dependencies: Performance relies on proprietary LLMs (GPT-4) and Whisper STT, limiting control over updates and accessibility for resource-constrained institutions.
- Human Validation Bottleneck: Teacher-in-the-loop content approval, while ensuring accuracy, creates scalability challenges for large learner groups.
- Luxembourgish Specificity: Evaluations focused solely on Luxembourgish; generalizability to other low-resource languages with non-Latin scripts (e.g., Uralic or Bantu languages) remains unverified.
- Short-Term Engagement Metrics: Pilot studies measured immediate usability but not long-term proficiency gains (e.g., CEFR progression over 6+ months). Additionally, the pilot study’s small sample size (n=14) should be increased in future studies.
8. Conclusion and Future Work
- Automate BPMN Generation: Develop tools to derive BPMN diagrams directly from curriculum specifications or learning objectives, reducing manual modeling effort.
- Broaden Curriculum Coverage: Extend our pipeline to additional CEFR levels (C1—C2) and subject domains (e.g., business, technical language).
- Enhanced Teacher-in-the-Loop: Introduce richer interfaces and analytics dashboards for instructors to review, adjust, and annotate agent workflows and content.
- Adaptive Learning Algorithms: Integrate reinforcement learning and learner modeling to personalize task sequencing dynamically based on real-time performance data.
- Longitudinal Studies: Conduct extended field trials across diverse learner populations and languages to evaluate long-term efficacy, retention gains, and transfer to real-world communication.
- Improve explainability: Develop teacher-facing dashboards to visualize BPMN execution logs and RAG source attributions, enhancing real-time explainability. Applying Model-Agnostic XAI methods could be considered, such as Local Interpretable Model-agnostic Explanations (LIME) for text and SHapley Additive exPlanations (SHAP) for transformers.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A. BPMN Activity Diagrams for Tutor Agents (Reading, Listening, Question answering, and Grammar & Summary)




Appendix B. Sample Prompt Templates
system_message="You are the communicator agent, your job is to communicate with the user in Luxembourgish to generate a learning recommendation for them"
Appendix B.1. Conversational Tutor Agent Prompt
Conversational_Agent_Prompt = """
Dir sidd en digitalen Tutor, spezialis\’{e}iert op Sproochl\’{e}ieren
mat vill Erfahrung, besonnesch an der konversationeller Praxis.
\"{A}ert Zil ass et, d’Benotzer duerch effektivt Sproochl\’{e}ieren
mat engem konversationellen Usaz ze f\’{e}ieren.
Follegt d\"{e}s Instruktioune fir d\"{e}st z’erreechen:
1. L\’{e}ierziler setzen:
- F\"{A}nkt un, d’L\’{e}ierziler ze erkl\"{A}ren op Basis vum Inhalt,
deen ofgedeckt g\"{e}tt.
2. Wierderbuch an Notzung:
- Bedeelegt Iech un Gespr\’{e}icher, erkl\"{A}ert de benotzte
Wierderbuch a motiv\’{e}iert de Benotzer nei Wierder ze soen
oder se an S\"{A}tz ze benotzen.
3. Rollenspill:
- F\’{e}iert Rollenspill\"{u}bungen duerch:
- Defin\’{e}iert de Fokus vum Gespr\’{e}ich.
- Spezifiz\’{e}iert \"{A}r Roll an d’Roll vum Benotzer.
- Gitt dem Benotzer e Signal fir unzef\"{A}nken.
4. Evaluatioun a Feedback:
- Evalu\’{e}iert d’\"{A}ntwerte vum Benotzer grammatesch,
syntaktesch an a puncto Aussprooch.
- Wann d’\"{A}ntwert korrekt ass, spillt \"{A}r Roll.
- Wann d’\"{A}ntwert falsch ass, spillt d’Roll vum Tutor,
korrig\’{e}iert de Benotzer, gitt Hinweise an Tipps, dann
spillt \"{A}r Roll.
5. Resum\’{e} an Nofro:
- Resum\’{e}iert d’Gespr\’{e}ich, hebt neie Wierderbuch ervir, an
erkl\"{A}ert w\’{e}i een en benotzt.
- Frot de Benotzer, ob se m\’{e}i Beispiller w\"{e}llen oder
schl\’{e}it besser \"{A}ntwerten a Wierderbuch vir.
6. Feedback ginn:
- Gitt \"{e}mmer Feedback iwwer dat, wat de Benotzer gel\’{e}iert
huet an un wat se schaffe sollten.
7. Fortschr\"{e}ttsbericht:
- Schreift e Bericht iwwer de Fortschr\"{e}tt vum Benotzer:
- Resum\’{e}iert, wat se erfollegr\"{A}ich gel\’{e}iert hunn.
- Hieft Ber\"{A}icher ervir, un deenen se schaffe mussen.
- Identifiz\’{e}iert all Schwiriegkeeten, d\’{e}i se beim
L\’{e}iere haten.
Huelt Iech e Moment Z\"{A}it an schafft methodesch un all
Schr\"{e}tt, benotzt de bereetgestallten Inhalt als Referenz fir
ze l\’{e}ieren an nei L\’{e}iermaterialien ze gener\’{e}ieren, a
kontroll\’{e}iert \"{e}mmer, ob de Benotzer Iech follegt.
"""
You are a digital tutor specializing in language learning
with extensive experience, especially in conversational
practice. Your goal is to guide users through effective
language learning using a conversational approach. Follow
these instructions to achieve this:
1. Set Learning Objectives
--- Begin by explaining the learning objectives based on
the content being covered.
2. Vocabulary and Usage
--- Engage the user in conversation, explain the vocabulary
you use, and encourage them to produce new words or use
them in sentences.
3. Role-Play
--- Conduct role-play exercises by:
\bullet Defining the focus of the dialogue.
\bullet Specifying your role and the user’s role.
\bullet Giving the user a clear signal to begin.
4. Evaluation and Feedback
--- Evaluate the user’s responses for grammar, syntax, and
pronunciation.
\bullet If the response is correct, proceed with your next line.
\bullet If the response is incorrect, adopt the tutor role:
correct the user, offer hints and tips, then resume the
role-play.
5. Summary and Follow-Up
--- Summarize the conversation, highlight new vocabulary,
and explain how to use it.
--- Ask if the user would like more examples or suggestions
for better answers and additional vocabulary.
6. Providing Feedback
--- Always give feedback on what the user has learned and
what they should focus on next.
7. Progress Report
--- Write a brief report on the user’s progress:
\bullet Summarize what they have successfully learned.
\bullet Highlight areas that need further practice.
\bullet Identify any difficulties they encountered.
Take your time and work methodically through each step,
using the provided content as your reference, generating new
learning materials as needed, and always checking that the
user is keeping up with you.
References
- Cavojský, M.; Bugár, G.; Kormaník, T.; Hasin, M. Exploring the capabilities and possible applications of large language models for education. 2023 21st International Conference on Emerging eLearning Technologies and Applications (ICETA) 2023, 91–98. [Google Scholar]
- Belda-Medina, J.; Calvo-Ferrer, J.R. Using Chatbots as AI Conversational Partners in Language Learning. Applied Sciences 2022, 12, 8427. [Google Scholar] [CrossRef]
- Abedi, M.; Alshybani, I.; Shahadat, M.; Murillo, M. Beyond traditional teaching: The potential of large language models and chatbots in graduate engineering education. arXiv 2023, arXiv:2312.12345. [Google Scholar] [CrossRef]
- Chen, M.H.; Ye, S.X. Extending repair in peer interaction: A conversation analytic study. Front. Psychol. 2022, 13, 926842. [Google Scholar] [CrossRef] [PubMed]
- Huang, L.; Yu, W.; Ma, W.; Zhong, W.; Feng, Z.; Wang, H.; Chen, Q.; Peng, W.; Feng, X.; Qin, B.; Liu, T. A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions. arXiv 2023, arXiv:2311.05232. [Google Scholar] [CrossRef]
- Xie, Q.; Guo, X. L2 teacher support and positive L2 academic emotions: The mediating role of self-efficacy. J. Psycholinguist. Res. 2022, 51, 1–17. [Google Scholar] [CrossRef]
- D’Mello, S.K.; Graesser, A.C. Intelligent Tutoring Systems: How computers achieve learning gains that rival human tutors. In Handbook of Educational Psychology, 4th ed.; Schutz, P.A., Muis, K.R., Eds.; American Psychological Association: Washington, DC, USA, 2023; pp. 603–629. [Google Scholar]
- Statistics Portal, Luxembourg. Nationalities. Available online: https://statistiques.public.lu/en/recement/nationalites.html (accessed on 22 June 2025).
- Object Management Group. Business Process Model and Notation (BPMN). Available online: https://www.bpmn.org (accessed on 22 June 2025).
- Neumann, A.; Yin, Y.; Sowe, S.K.; Decker, S.; Jarke, M. An LLM-driven chatbot in higher education for databases and information systems. IEEE Transactions on Education 2025, 68, 103–116. [Google Scholar] [CrossRef]
- Costa, L.F.; Silva, P. Applying BPMN to Adaptive E-Learning Path Modeling: A Case Study. Education and Information Technologies 2023, 28(5), 6543–6561. [Google Scholar]
- Ivanova, T.; Terzieva, V.; Todorova, K. An Agent-Oriented Architecture For Strategy-Based Personalized E-Learning. 2021 Big Data, Knowledge and Control Systems Engineering (BdKCSE) 2021, 1-8.
- Strik, H.; Truong, K.; de Wet, F.; Cucchiarini, C. Comparing Different Approaches for Automatic Pronunciation Error Detection. Speech Communication 2019, 113, 28–39. [Google Scholar] [CrossRef]
- Statistics Portal, Luxembourg. Linguistic diversity on the rise. Available online: https://statistiques.public.lu/en/recensement/diversite-linguistique.html.
- Duolingo Inc. Language Learning Platform: Features and Limitations. Available online: https://www.duolingo.com.
- García-López, R.; Smith, J.; Martinez, A. BPMN for Educational Process Modeling: A Systematic Approach. Comput. Educ. 2023, 198, 104–118. [Google Scholar]
- Halder, S.; Meyer, T.; Schmidt, L. Challenges in NLP for Low-Resource Languages: The Case of Luxembourgish. Proceedings of LREC 2024, 234–241. [Google Scholar]
- Huang, Y.; Qin, Z.; Liu, W. Hallucinations in Large Language Models: Challenges and Mitigation Strategies. Proceedings of ACL 2023, 1122–1135. [Google Scholar]
- LangChain Inc. LangGraph: Building Stateful Multi-Agent Applications. Available online: https://langchain.com/langgraph.
- Lavergne, T.; Urvoy, T.; Yvon, F. NLP Resources for Luxembourgish: Current State and Future Directions. Proceedings of LREC 2022, 3421–3428. [Google Scholar]
- Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.; RocktÄschel, T.; et al. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. Proceedings of NeurIPS 2020, 9459–9474. [Google Scholar]
- Luxembourg Today. Digital Resources for Learning Luxembourgish: A Growing Need. Available online: https://today.rtl.lu/education/luxembourgish-digital (accessed on 18 November 2024).
- Microsoft Research. AutoGen: Enable Next-Gen Large Language Model Applications. Available online: https://github.com/microsoft/autogen.
- Tebourbi, H.; Nouzri, S ; Mualla, Y.; Najjar, A. Personalized Language Learning: A Multi-Agent System Leveraging LLMs for Teaching Luxembourgish. In Proceedings of AAMAS 2025, Detroit, MI, USA, 19-23 May 2025. Available online: https://www.ifaamas.org/Proceedings/aamas2025/pdfs/p3032.pdf.
- RTL Luxembourg. The Challenge of Teaching Luxembourgish in the Digital Age. Available online: https://www.rtl.lu/education (accessed on 20 November 2024).
- Rzepka, R.; Araki, K.; Kojima, K. Addressing Hallucinations in Educational AI: A Critical Analysis. Int. J. AI Educ. 2023, 33, 245–263. [Google Scholar]
- Wang, S.; Liu, Y.; Chen, H. RAG Applications in Educational AI: Reducing Hallucinations and Improving Accuracy. J. AI Educ. 2024, 11, 156–171. [Google Scholar]
- Wooldridge, M. J.; Jennings, N. R. Intelligent Agents: Theory and Practice. Knowledge Engineering Review 1995, 10, 115–152. [Google Scholar] [CrossRef]
- Conneau, A.; Khandelwal, K.; Goyal, N.; Chaudhary, V.; Wenzek, G.; Guzmán, F.; Grave, E.; Ott, M.; Zettlemoyer, L.; Stoyanov, V. Unsupervised Cross-lingual Representation Learning at Scale. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; 2020; pp. 8709–8719.
- Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of NAACL-HLT 2019; 2019; pp. 4171–4186.
- Radford, A.; Kim, J. W.; Xu, T.; Brockman, G.; McLeavey, C.; Sutskever, I. Robust Speech Recognition via Large-Scale Weak Supervision. In Proceedings of the 40th International Conference on Machine Learning (PMLR), Vol. 202; 2023; pp. 28492–28518. :contentReference[oaicite:3]index=3.
- Picard, A.; Mualla, Y.; Gechter, F.; Galland, S. Human-computer interaction and explainability: Intersection and terminology. In Proceedings of the World Conference on Explainable Artificial Intelligence; Springer: 2023; pp. 214–236.
- Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A survey of methods for explaining black box models. ACM Computing Surveys (CSUR) 2019, 51(5), 93. [Google Scholar] [CrossRef]
- Mualla, Y.; Tchappi, I.; Kampik, T.; Najjar, A.; Calvaresi, D.; Abbas-Turki, A.; Galland, S.; Nicolle, C. The quest of parsimonious XAI: A human-agent architecture for explanation formulation. Artificial Intelligence 2022, 302, 103573. [Google Scholar] [CrossRef]
- Hemmer, P.; Schemmer, M.; Vössing, M.; Kühl, N. Human-AI complementarity in hybrid intelligence systems: A structured literature review. PACIS 2021, 78. [Google Scholar]
- Glass, A.; McGuinness, D.L.; Wolverton, M. Toward establishing trust in adaptive agents. In Proceedings of the 13th International Conference on Intelligent User Interfaces; 2008; pp. 227–236. [Google Scholar]
- Liao, Q.V.; Gruen, D.; Miller, S. Questioning the AI: Informing design practices for explainable AI user experiences. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems; 2020; pp. 1–15. [Google Scholar]
- Bunt, A.; Lount, M.; Lauzon, C. Are explanations always important? A study of deployed, low-cost intelligent interactive systems. In Proceedings of the 2012 ACM International Conference on Intelligent User Interfaces; 2012; pp. 169–178.
- Gunning, D. Explainable artificial intelligence (XAI). Defense Advanced Research Projects Agency (DARPA) 2017. https://www.darpa.mil/program/explainable-artificial-intelligence.
- Biran, O.; Cotton, C. Explanation and justification in machine learning: A survey. In IJCAI-17 Workshop on Explainable AI (XAI); 2017; pp. 8–13.
- Contreras, V.; Marini, N.; Fanda, L.; Manzo, G.; Mualla, Y.; Calbimonte, J.-P.; Schumacher, M.; Calvaresi, D. A dexire for extracting propositional rules from neural networks via binarization. Electronics 2022, 11(24), 4171. [Google Scholar] [CrossRef]
- LangChain Inc. LangSmith: Interactive Tooling for Explainable LLM Workflows. https://langchain.com/langsmith.
- LangChain Inc. LangChain: Building Applications with LLMs Through Composable Chains and Tools. Available online: https://www.langchain.com (accessed on 1 July 2025).
- Chroma Inc. Chroma: Open-Source Embeddings Database for AI Applications. Available online: https://www.trychroma.com (accessed on 1 July 2025).
- OpenAI. GPT-4 with Vision: Multimodal Large Language Models. Available online: https://platform.openai.com/docs/guides/vision (accessed on 1 July 2025).
- Coqui Inc. Coqui TTS: VITS-Based Text-to-Speech Models. Available online: https://coqui.ai (accessed on 1 July 2025).
- OpenAI. Whisper: Robust Speech Recognition via Large-Scale Weak Supervision. Available online: https://openai.com/research/whisper (accessed on 1 July 2025).
- Nouzri, S.; El Fatimi, M.; Guerin, T.; Othmane, M.; Najjar, A. Beyond Chatbots: Enhancing Luxembourgish Language Learning Through Multi-agent Systems and Large Language Model. In Proceedings of PRIMA 2024: Principles and Practice of Multi-Agent Systems; Arisaka, R.; Sanchez-Anguix, V.; Stein, S.; Aydoğan, R.; van der Torre, L.; Ito, T., Eds.; Lecture Notes in Computer Science, Vol. 15395; Springer: Cham, 2025. Available online: https://doi.org/10.1007/978-3-031-77367-9_29. [CrossRef]
| 1 | This article is a revised and expanded version of these papers. |





| BPMN Element | LangGraph Concept | MAS Component | Function |
|---|---|---|---|
| Pool | Agent Node | Agent Class | Encapsulates a high-level role (e.g., Communicator, Orchestrator) |
| Lane | Tool Node | Agent Capability | Provides an external service or helper (e.g., getFiles) |
| Task | Task Node | Method Invocation | Executes a concrete operation (e.g., generateRecommendation) |
| Gateway | Router | Routing Logic | Evaluates conditions and selects outgoing edge |
| Data Object | State Variable | Memory Store | Holds persistent data (user profile, progress, curriculum) |
| Message Flow | Message Edge | Inter-Agent Message | Transmits data or control between agents |
| Metric | Score |
|---|---|
| Context Relevancy | 0.87 |
| Faithfulness | 0.82 |
| Answer Relevancy | 0.85 |
| Question | Response Distribution |
|---|---|
| Ease of Interaction | Very Easy (42.9%), Easy (42.9%), Difficult (14.3%) |
| Satisfaction with Understanding & Contextual Responses | Satisfied (42.9%), Very Satisfied (28.6%), Neutral (28.6%) |
| Engagement Level | Very engaging (71.4%), Moderately engaging (28.6%) |
| Likelihood to Continue | Likely (71.4%), Very Likely (14.3%), Neutral (14.3%) |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
