MarkTechPost@AI 01月31日
From Deep Knowledge Tracing to DKT2: A Leap Forward in Educational AI
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

DKT2是一种新型的基于深度学习的知识追踪框架,它利用xLSTM架构克服了传统方法的局限性。与早期模型不同,DKT2使用Rasch模型改进输入表示,并结合项目反应理论(IRT)增强可解释性。通过识别熟悉和不熟悉的知识,DKT2更好地表示了学生的学习状态。xLSTM的使用通过可修改的存储决策、增加的内存容量和完全并行化,解决了传统LSTM的局限性,从而提高了可扩展性和效率。该创新使模型在保持强大的适用性的同时,提供了比同类模型更好的预测精度。

🚀 DKT2 引入了 Rasch 模型进行输入表示,并结合 IRT 增强模型的可解释性,能够更清晰地识别学生熟悉和不熟悉的知识点。

🧠 模型核心采用 xLSTM 架构,通过 sLSTM 和 mLSTM 模块,实现了更好的内存保持、并行化优化和动态知识更新,有效提升了模型的可扩展性和效率。

🎯 DKT2 通过系统的学习流程,结合历史知识状态和预测问题,创建了学生学习进度的综合概览,并使用二元交叉熵损失进行训练,确保模型在实际应用中的稳健性。

📊 大量实验结果表明,DKT2 在多个预测任务中始终优于17个基线模型,包括单步、多步和不同历史长度的预测,证明了其在实际智能辅导系统中的有效性和优越性。

Knowledge Tracing (KT) plays a crucial role in Intelligent Tutoring Systems (ITS) by modeling students’ knowledge states and predicting their future performance. Traditional KT models, such as Bayesian Knowledge Tracing (BKT) and early deep learning-based approaches like Deep Knowledge Tracing (DKT), have demonstrated effectiveness in learning student interactions. However, recent advancements in deep sequential KT models, such as Attentive Knowledge Tracing (AKT), have increasingly prioritized predictive performance over practical applicability and comprehensive knowledge modeling. These models often face fundamental challenges, including limited parallel computing efficiency, difficulties modifying stored knowledge, and restricted storage capacity. Additionally, many deep KT models rely on future interactions, which are typically unavailable in real-world applications, limiting their usability. Addressing these challenges is critical to enhancing KT models’ scalability, interpretability, and effectiveness in large-scale educational systems.

Existing KT models utilize deep learning-based architectures to forecast student performance, with models like DKT utilizing Long Short-Term Memory (LSTM) networks to learn the dynamics of learning. Although attention-based models such as AKB use attention mechanisms to improve long-range dependencies, they take future responses as input, rendering them not applicable in real-world scenarios. Deep sequential models also suffer from parallelization and memory issues, which lowers their efficiency in working with large-scale datasets. Other methods, including graph-based and memory-augmented models, are usually not interpretable, meaning they cannot provide useful insights into the student’s learning process. These shortfalls result in a gap between theoretical breakthroughs and practical applications, where an even more scalable, interpretable, and efficient KT model is needed.

Researchers from Zhejiang University propose DKT2, a novel deep learning-based KT framework that leverages the xLSTM architecture to overcome the limitations of previous methods DKT2 is different from earlier models as it uses the Rasch model to improve input representation and incorporates Item Response Theory (IRT) for enhanced interpretability. By identifying familiar and unfamiliar knowledge, DKT2 offers a better representation of the state of learning in students. The use of xLSTM resolves the limitations of classical LSTMs through revisable storage decisions, increased memory capacity, and full parallelization, resulting in greater scalability and efficiency. The innovation allows the model to maintain robust applicability while providing better predictive accuracy than its counterparts.

DKT2 adopts a systematic learning pipeline with Rasch embedding to record student-question interactions and include difficulty levels for better input representation. The xLSTM blocks use sLSTM and mLSTM to facilitate better memory retention, parallelization optimization, and dynamic knowledge updating. The IRT prediction and knowledge decomposition module separates familiar and unfamiliar knowledge to enable more interpretable knowledge tracing. Integrated knowledge fusion fuses historical knowledge states and predicted questions to create a comprehensive overview of student learning progress. The model is trained with binary cross-entropy loss and evaluated on three large-scale datasets—Assist17, EdNet, and Comp—to ensure robustness in real-world ITS applications.

Extensive experiments on three large-scale datasets demonstrate that DKT2 consistently outperforms 17 baseline models across multiple prediction tasks, including one-step, multi-step, and varying-history-length predictions. It achieves higher accuracy, AUC, and lower RMSE compared to deep sequential models like DKT and attention-based models like AKT. The integration of xLSTM enhances parallelization and memory capacity, mitigating error accumulation in multi-step predictions, while the Rasch model and IRT improve interpretability by effectively distinguishing familiar and unfamiliar knowledge. An ablation study confirms that each component of DKT2 contributes significantly to its superior performance, particularly mLSTM, which is crucial for scalability in large-scale datasets. These results establish DKT2 as a robust and applicable solution for real-world Intelligent Tutoring Systems.

DKT2 is a breakthrough in knowledge tracing by combining xLSTM, the Rasch model, and IRT to achieve a perfect balance between prediction accuracy and real-world usability. Through interpretable knowledge state generation and parallelized, memory-saving learning, the method guarantees scalability and improved personalization in ITS applications. Areas of future work include extending DKT2’s applicability to ultra-large datasets and improving its multi-concept prediction ability to better support adaptive learning systems.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 70k+ ML SubReddit.

Meet IntellAgent: An Open-Source Multi-Agent Framework to Evaluate Complex Conversational AI System (Promoted)

The post From Deep Knowledge Tracing to DKT2: A Leap Forward in Educational AI appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

知识追踪 DKT2 xLSTM 教育AI 智能辅导系统
相关文章