计算机科学
过度拟合
启发式
机器学习
稳健性(进化)
人工智能
追踪
依赖关系(UML)
一般化
人工神经网络
数学
生物化学
基因
操作系统
数学分析
化学
作者
Shuyan Huang,Zitao Liu,Xiangyu Zhao,Weiqi Luo,Jian Weng
标识
DOI:10.1145/3539618.3592073
摘要
Knowledge tracing (KT) is the problem of predicting students' future performance based on their historical interaction sequences. With the advanced capability of capturing contextual long-term dependency, attention mechanism becomes one of the essential components in many deep learning based KT (DLKT) models. In spite of the impressive performance achieved by these attentional DLKT models, many of them are often vulnerable to run the risk of overfitting, especially on small-scale educational datasets. Therefore, in this paper, we propose sparseKT, a simple yet effective framework to improve the robustness and generalization of the attention based DLKT approaches. Specifically, we incorporate a k-selection module to only pick items with the highest attention scores. We propose two sparsification heuristics: (1) soft-thresholding sparse attention and (2) top-K sparse attention. We show that our sparseKT is able to help attentional KT models get rid of irrelevant student interactions and improve the predictive performance when compared to 11 state-of-the-art KT models on three publicly available real-world educational datasets. To encourage reproducible research, we make our data and code publicly available at https://github.com/pykt-team/pykt-toolkit1..
科研通智能强力驱动
Strongly Powered by AbleSci AI