EKT: Exercise-Aware Knowledge Tracing for Student ...

16
EKT: Exercise-Aware Knowledge Tracing for Student Performance Prediction Qi Liu , Member, IEEE, Zhenya Huang , Yu Yin, Enhong Chen , Senior Member, IEEE, Hui Xiong, Senior Member, IEEE, Yu Su, and Guoping Hu Abstract—For offering proactive services (e.g., personalized exercise recommendation) to the students in computer supported intelligent education, one of the fundamental tasks is predicting student performance (e.g., scores) on future exercises, where it is necessary to track the change of each student’s knowledge acquisition during her exercising activities. Unfortunately, to the best of our knowledge, existing approaches can only exploit the exercising records of students, and the problem of extracting rich information existed in the materials (e.g., knowledge concepts, exercise content) of exercises to achieve both more precise prediction of student performance and more interpretable analysis of knowledge acquisition remains underexplored. To this end, in this paper, we present a holistic study of student performance prediction. To directly achieve the primary goal of performance prediction, we first propose a general Exercise-Enhanced Recurrent Neural Network (EERNN) framework by exploring both student’s exercising records and the text content of corresponding exercises. In EERNN, we simply summarize each student’s state into an integrated vector and trace it with a recurrent neural network, where we design a bidirectional LSTM to learn the encoding of each exercise from its content. For making final predictions, we design two implementations on the basis of EERNN with different prediction strategies, i.e., EERNNM with Markov property and EERNNA with Attention mechanism. Then, to explicitly track student’s knowledge acquisition on multiple knowledge concepts, we extend EERNN to an explainable Exercise-aware Knowledge Tracing (EKT) framework by incorporating the knowledge concept information, where the student’s integrated state vector is now extended to a knowledge state matrix. In EKT, we further develop a memory network for quantifying how much each exercise can affect the mastery of students on multiple knowledge concepts during the exercising process. Finally, we conduct extensive experiments and evaluate both EERNN and EKT frameworks on a large- scale real-world data. The results in both general and cold-start scenarios clearly demonstrate the effectiveness of two frameworks in student performance prediction as well as the superior interpretability of EKT. Index Terms—Intelligent education, knowledge tracing, exercise content, knowledge concept Ç 1 INTRODUCTION I NTELLIGENT education systems, such as Massive Online Open Course, Knewton.com and KhanAcedemy.org, can help the personalized learning of students with computer- assisted technology by providing open access to millions of online courses or exercises. Due to their prevalence and con- venience, these systems have attracted great attentions from both educators and general publics [1], [2], [3]. Specifically, students in these systems can choose exer- cises individually according to their needs and acquire necessary knowledge during exercising. Fig. 1 shows a toy example of such exercising process of a typical student. Generally, when an exercise (e.g., e 1 ) is posted, the student reads its content (“If function...”) and applies the corre- sponding knowledge on “Function” concept to answer it. From the figure, student s 1 has done four exercises, where she only answers exercise e 2 wrong, which may demon- strate that she has well mastered knowledge concepts “Function” and “Inequality” except the “Probability” con- cept. We can see that a fundamental task in such education systems is to predict student performance (e.g., score), i.e., forecasting whether or not a student can answer an exer- cise (e.g., e 5 ) correctly in the future [4]. Meanwhile, it also requires us to track the change of students’ knowledge acquisition in their exercising process [5], [6]. In practice, the success of precise prediction could benefit both student users and system creators: (1) Students can realize their weak knowledge concepts in time and thus prepare targeted exercising [7]; (2) System creators can provide better proactive services to different students, such as learning remedy suggestion and personalized exercise recommendation [8]. In the literature, there are many efforts in predicting stu- dent performance from both educational psychology and data mining areas, such as cognitive diagnosis [9], knowl- edge tracing [5], matrix factorization [10], topic modeling [3], sparse factor analysis [2] and deep learning [11]. Specifi- cally, existing work mainly focuses on exploiting the exercising process of students, where each exercise is Q. Liu, Z. Huang, Y. Yin, and E. Chen are with the Anhui Province Key Laboratory of Big Data Analysis and Application, School of Com- puter Science and Techonology, University of Science and Technology of China, Hefei, Anhui 230026, China. E-mail: {huangzhy, yxonic} @mail.ustc.edu.cn, {qiliuql, cheneh}@ustc.edu.cn. H. Xiong is with the Management Science and Information Systems Department, Rutgers Business School, Rutgers, the State University of New Jersey, Newark, NJ 07102 USA. E-mail: [email protected]. Y. Su and G. Hu are with iFLYTEK Research, IFLYTEK Co., Ltd, Hefei, Anhui 230088, China. E-mail: {yusu, gphu}@iflytek.com. Manuscript received 19 Jan. 2019; revised 3 May 2019; accepted 6 June 2019. Date of publication 24 June 2019; date of current version 7 Dec. 2020. (Corresponding author: Enhong Chen.) Recommended for acceptance by P. Cui. Digital Object Identifier no. 10.1109/TKDE.2019.2924374 100 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 33, NO. 1, JANUARY 2021 1041-4347 ß 2019 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See ht_tps://www.ieee.org/publications/rights/index.html for more information. Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Transcript of EKT: Exercise-Aware Knowledge Tracing for Student ...

Page 1: EKT: Exercise-Aware Knowledge Tracing for Student ...

EKT: Exercise-Aware Knowledge Tracingfor Student Performance Prediction

Qi Liu ,Member, IEEE, Zhenya Huang , Yu Yin, Enhong Chen , Senior Member, IEEE,

Hui Xiong, Senior Member, IEEE, Yu Su, and Guoping Hu

Abstract—For offering proactive services (e.g., personalized exercise recommendation) to the students in computer supported

intelligent education, one of the fundamental tasks is predicting student performance (e.g., scores) on future exercises, where it is

necessary to track the change of each student’s knowledge acquisition during her exercising activities. Unfortunately, to the best of our

knowledge, existing approaches can only exploit the exercising records of students, and the problem of extracting rich information

existed in the materials (e.g., knowledge concepts, exercise content) of exercises to achieve both more precise prediction of student

performance and more interpretable analysis of knowledge acquisition remains underexplored. To this end, in this paper, we present a

holistic study of student performance prediction. To directly achieve the primary goal of performance prediction, we first propose a

general Exercise-Enhanced Recurrent Neural Network (EERNN) framework by exploring both student’s exercising records and the text

content of corresponding exercises. In EERNN, we simply summarize each student’s state into an integrated vector and trace it with a

recurrent neural network, where we design a bidirectional LSTM to learn the encoding of each exercise from its content. For making

final predictions, we design two implementations on the basis of EERNN with different prediction strategies, i.e., EERNNM with Markov

property and EERNNA with Attention mechanism. Then, to explicitly track student’s knowledge acquisition on multiple knowledge

concepts, we extend EERNN to an explainable Exercise-aware Knowledge Tracing (EKT) framework by incorporating the knowledge

concept information, where the student’s integrated state vector is now extended to a knowledge state matrix. In EKT, we further

develop a memory network for quantifying how much each exercise can affect the mastery of students on multiple knowledge concepts

during the exercising process. Finally, we conduct extensive experiments and evaluate both EERNN and EKT frameworks on a large-

scale real-world data. The results in both general and cold-start scenarios clearly demonstrate the effectiveness of two frameworks in

student performance prediction as well as the superior interpretability of EKT.

Index Terms—Intelligent education, knowledge tracing, exercise content, knowledge concept

Ç

1 INTRODUCTION

INTELLIGENT education systems, such as Massive OnlineOpen Course, Knewton.com and KhanAcedemy.org, can

help the personalized learning of students with computer-assisted technology by providing open access to millions ofonline courses or exercises. Due to their prevalence and con-venience, these systems have attracted great attentions fromboth educators and general publics [1], [2], [3].

Specifically, students in these systems can choose exer-cises individually according to their needs and acquirenecessary knowledge during exercising. Fig. 1 shows a toyexample of such exercising process of a typical student.Generally, when an exercise (e.g., e1) is posted, the student

reads its content (“If function...”) and applies the corre-sponding knowledge on “Function” concept to answer it.From the figure, student s1 has done four exercises, whereshe only answers exercise e2 wrong, which may demon-strate that she has well mastered knowledge concepts“Function” and “Inequality” except the “Probability” con-cept. We can see that a fundamental task in such educationsystems is to predict student performance (e.g., score), i.e.,forecasting whether or not a student can answer an exer-cise (e.g., e5) correctly in the future [4]. Meanwhile, it alsorequires us to track the change of students’ knowledgeacquisition in their exercising process [5], [6]. In practice,the success of precise prediction could benefit both studentusers and system creators: (1) Students can realize theirweak knowledge concepts in time and thus preparetargeted exercising [7]; (2) System creators can providebetter proactive services to different students, such aslearning remedy suggestion and personalized exerciserecommendation [8].

In the literature, there are many efforts in predicting stu-dent performance from both educational psychology anddata mining areas, such as cognitive diagnosis [9], knowl-edge tracing [5], matrix factorization [10], topic modeling [3],sparse factor analysis [2] and deep learning [11]. Specifi-cally, existing work mainly focuses on exploiting theexercising process of students, where each exercise is

� Q. Liu, Z. Huang, Y. Yin, and E. Chen are with the Anhui ProvinceKey Laboratory of Big Data Analysis and Application, School of Com-puter Science and Techonology, University of Science and Technologyof China, Hefei, Anhui 230026, China. E-mail: {huangzhy, yxonic}@mail.ustc.edu.cn, {qiliuql, cheneh}@ustc.edu.cn.

� H. Xiong is with the Management Science and Information SystemsDepartment, Rutgers Business School, Rutgers, the State University ofNew Jersey, Newark, NJ 07102 USA. E-mail: [email protected].

� Y. Su and G. Hu are with iFLYTEK Research, IFLYTEK Co., Ltd, Hefei,Anhui 230088, China. E-mail: {yusu, gphu}@iflytek.com.

Manuscript received 19 Jan. 2019; revised 3 May 2019; accepted 6 June 2019.Date of publication 24 June 2019; date of current version 7 Dec. 2020.(Corresponding author: Enhong Chen.)Recommended for acceptance by P. Cui.Digital Object Identifier no. 10.1109/TKDE.2019.2924374

100 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 33, NO. 1, JANUARY 2021

1041-4347� 2019 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See ht_tps://www.ieee.org/publications/rights/index.html for more information.

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 2: EKT: Exercise-Aware Knowledge Tracing for Student ...

usually distinguished by the corresponding knowledge con-cepts in the modeling, e.g., exercise e1 in Fig. 1 is repre-sented as the concept “Function”. In other words, existingwork models students’ knowledge states for the predictiononly based on their performance records on each knowl-edge, where two exercises (e.g., e1 and e3) labeled with thesame knowledge concept are simply identified as the same(actually, exercise e1 and e3 are quite different according totheir contents, and e3 is more difficult than e1). Therefore,these approaches cannot distinguish the knowledge acquisi-tion of two students if one solves e1 but the other solves e3since these knowledge-specific representations underutilizethe rich information of exercise materials (e.g., text con-tents), causing severe information loss [9]. To this end, weargue that it is beneficial to combine both student’s exercis-ing records and the exercise materials for more preciselypredicting student performance.

Unfortunately, there are many technical and domainchallenges along this line. First, there are diverse expres-sions of exercises, which requires a unified way to automati-cally understand and represent the characteristics ofexercises from a semantic perspective. Second, students’performance in the future is deeply relied on their long-term historical exercising, especially on their importantknowledge states. How to track the historically focusedinformation of students is very challenging. Third, the taskof student performance prediction usually suffers from the“cold start” problem [12], [13]. That is, we have to make pre-dictions for new students and new exercises. In this sce-nario, limited information could be exploited, and thus,leading to the poor prediction results. Last but not least, stu-dents usually care about not only what they need to learnbut also wonder why they need it, i.e., it is necessary toremind them whether or not they are good at a certainknowledge concept and how much they have alreadylearned about it. However, it is a nontrivial problem toeither quantify the impacts of solving each specific exercise(e.g., e1) on improving the student’s knowledge acquisition(e.g., “Function”) or interpretably track the change ofstudent’s knowledge states during the exercising process.

To directly achieve the primary goal of predicting stu-dent performance with addressing the first three challenges,in our preliminary work [14], we proposed an Exercise-Enhanced Recurrent Neural Network (EERNN) frameworkby mainly exploring both student’s exercising records andthe corresponding exercise contents. Specifically, for theexercising process modeling, we first designed a bidirec-tional LSTM to represent the semantics of each exercise byexploiting its content. The learned encodings could capture

the individual characteristics of each exercise without anyexpertise. Then, we proposed another LSTM architecture totrace student states in the sequential exercising process withthe combination of exercise representations. For makingfinal predictions, we designed two strategies on the basis ofEERNN framework. The first one was a straightforward yeteffective strategy, i.e., EERNNM with Markov property, inwhich the students’ next performance only depended oncurrent states. Comparatively, the second was a moresophisticated one, EERNNA with Attention mechanism, whichtracked the focused student states based on similar exercisesin the history. In this way, EERNN could naturally predictstudent’s future performance given her exercising records.

In EERNN model, we summarized and tracked eachstudent’s knowledge states on all concepts in one integratedhidden vector. Thus, it could not explicitly explain howmuch a student had mastered a certain knowledge concept(e.g., “Function”), which meant that the interpretability ofEERNN was not satisfying enough. Therefore, in this paper,we extend EERNN and propose an explainable Exercise-aware Knolwedge Tracing (EKT) framework to track studentstates on multiple explicit concepts simultaneously. Specifi-cally, we extend the integrated state vector of each studentto a knowledge state matrix that updates over time, whereeach vector represents her mastery level of a certain concept.At each exercising step of a certain student, we develop amemory network to quantify the different impacts on eachknowledge state when she solves a specific exercise. We alsoimplement two EKT based prediction models following theproposed strategies in EERNN, i.e., EKTM with Markov prop-erty and EKTA with Attention mechanism. Finally, we conductextensive experiments and evaluate both EERNN and EKTframeworks on a large-scale real-world dataset. The experi-mental results in both general and cold-start scenarios clearlydemonstrate the effectiveness of two proposed frameworksin student performance prediction as well as the superiorinterpretability of EKT framework.

2 RELATED WORK

The related work can be classified into the following catego-ries from both educational psychology (i.e., cognitive diag-nosis and knowledge tracing) and data mining (i.e., matrixfactorization and deep learning methods) areas.

Cognitive Diagnosis. In the domain of educational psy-chology, cognitive diagnosis is a kind of techniques thataims to predict student performance by discovering studentstates from the exercising records [9]. Generally, traditionalcognitive diagnostic models (CDM) could be grouped into

Fig. 1. Example: Left box shows the exercising process of a student, where she has already done four exercises and is going to answer exercise e5.Right table shows the corresponding materials of exercises that contain their contents and knowledge concepts.

LIU ETAL.: EKT: EXERCISE-AWARE KNOWLEDGE TRACING FOR STUDENT PERFORMANCE PREDICTION 101

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 3: EKT: Exercise-Aware Knowledge Tracing for Student ...

two categories: continuous models and discrete ones.Among them, item response theory (IRT), as a typical con-tinuous model, characterized each student by a variable,i.e., a latent trait that describes the integrated knowledgestate, from a logistic-like function [15]. Comparatively, dis-crete models, such as Deterministic Inputs, Noisy-And gatemodel (DINA), represented each student as a binary vectorwhich denoted whether she mastered or not the knowledgeconcepts required by exercises with a given Q-matrix (exer-cise-knowledge concept matrix) prior [16]. To improve pre-diction effectiveness, many variations of CDMs wereproposed by combining learning information [17], [18], [19].For example, learning factors analysis [17] and performancefactors analysis [18] incorporated the time factor into themodeling. Liu et al. [19] proposed FuzzyCDF that consid-ered both subjective and objective exercise types to balanceprecision and interpretability of the diagnosis results.

Knowledge Tracing. Knowledge tracing is an essential taskfor tracing the knowledge states of each student separately,so that we can predict her performance on future exercisingactivities, where the basic idea is similar to the typicalsequence mining in various domains [20], [21], [22], [23]. Inthis task, Bayesian knowledge tracing (BKT) [5] was one ofthe most popular models. It was a knowledge-specificmodel which assumed each student’s knowledge states as aset of binary variables, where each variable represented shehad “mastered” or “non-mastered” on a specific concept.Generally, BKT utilized a Hidden Markov Model [24] toupdate knowledge states of each student separately fol-lowed by her performance on exercises. On the basis ofBKT, many extensions were proposed by considering otherfactors, e.g., exercise difficulty [25], multiple knowledgeconcepts [26] and student individuals [27]. One step further,to improve the prediction performance, other researchersalso suggested incorporating some cognitive factors into tra-ditional BKT model [28].

Matrix Factorization. Recently, researchers have attemptedto leverage matrix factorizations from data mining field forstudent performance prediction [10], [29]. Usually, the goalof this kind of research is to predict the unknown scores ofstudents as accurate as possible given a student-exercise per-formance matrix with some known scores. For example,Thai et al. [10] leveraged matrix factorization models to proj-ect each student into a latent vector that depicted students’implicit knowledge states, and further proposed a multi-relational adaption model for the prediction in online learn-ing systems. To capture the changes of student’s exercisingprocess, some additional factors are incorporated. For exam-ple, Thai et al. [30] proposed a tensor factorization approachby adding additional time factors. Chen et al. [31] noticed theeffects of both learning theory and Ebbinghaus forgettingcurve theory and incorporated them into a unified probabi-listic framework. Teng et al. [32] further investigated theeffects of two concept graphs.

Deep Learning Methods. Learning is a very complex process,where the mastery level of students on different knowledgeconcepts is not updated separately but related to each other.Along this line, inspired by the remarkable performanceof deep learning techniques in many applications, such asspeech recognition [33], image learning [34], [35], naturallanguage processing [36], network embedding [37], [38], and

also educational applications like question difficulty predic-tion [39], some researchers attempted to use deep models forstudent performance prediction [6], [11]. Among these work,deep knowledge tracing (DKT) was the first attempt, to thebest of our knowledge, to utilize recurrent neural networks(e.g., RNN and LSTM) to model student’s exercising processfor predicting her performance [11]. Moreover, by bridgingthe relationship between exercises and knowledge concepts,Zhang et al. [6] proposed a dynamic key-value memory net-work model for improving the interpretability of the predic-tion results, and Chen et al. [40] incorporated the knowledgestructure information for dealing with the data sparsity prob-lem in knowledge tracing. Experimental results showed thatdeepmodels had achieved a great success.

Our work differs from the previous studies as follows.First, existing approaches mainly focus on exploitingstudents’ historical exercising records for their performanceprediction, while ignoring the important effects of exercisematerials (e.g., knowledge concepts, exercise content). Tothe best of our knowledge, this work is the first comprehen-sive attempt that fully explores both student’s exercisingrecords and the exercise materials. Second, previous studiesfollow the common sense that student’s next performanceonly depends on the current states, while our work deeplycaptures the focused information of students in the historyby a novel attention mechanism for improving the predic-tion. Third, we can well handle the cold-start problem byincorporating exercise correlations without any retraining.Last but not least, our work can achieve good predictionresults with interpretability, i.e., we can explain the changeof student’s knowledge states on explicit knowledge con-cepts, which is beneficial for many real-world applications,such as explainable exercise recommendation.

3 PROBLEM AND SOLUTION OVERVIEW

In this section, we first formally define the problem of stu-dent performance prediction in intelligent education. Then,we will present the overview of our study.

Problem Definition. In an intelligent education system,suppose there are jSj students and jEj exercises, where stu-dents do exercises individually. We record the exercisingprocess of a certain student as s ¼ fðe1; r1Þ; ðe2; r2Þ; . . . ;ðeT ; rT Þg; s 2 S, where et 2 E represents the exercise prac-ticed by student s at her exercising step t and rt denotes thecorresponding score. Generally, if student s answers exer-cise et right, rt equals to 1, otherwise rt equals to 0. In addi-tion to the logs of student’s exercising process, we alsoconsider the materials of exercises (some examples areshown in Fig. 1). Formally, for a certain exercise e, wedescribe it by the text content, which is combined with aword sequence as e ¼ fw1; w2; . . . ; wMg. Also, the exercise econtains its corresponding knowledge concept k comingfrom allK concepts. Please note that each exercise may con-tain multiple concepts, e.g., e5 in Fig. 1 has two concepts“Function” and “Inequality”. Without loss of generality, inthis paper, we represent each student’s exercising record ass ¼ fðe1; r1Þ; ðe2; r2Þ; . . . ; ðeT ; rT Þg or s ¼ fðk1; e1; r1Þ; ðk2; e2;r2Þ; . . . ; ðkT ; eT ; rT Þg, where the former one does not considerthe knowledge concept information. Then the problem canbe defined as:

102 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 33, NO. 1, JANUARY 2021

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 4: EKT: Exercise-Aware Knowledge Tracing for Student ...

Definition 1 (Student Performance Prediction Prob-lem). Given the exercising logs of each student and the materi-als of each exercise from exercising step 1 to T , our goal is two-fold: (1) track the change of her knowledge states and estimatehow much she masters all K knowledge concepts from step 1 toT ; (2) predict the response score erTþ1 on the next candidateexercise eTþ1.

Solution Overview.An overview of the proposed solution isillustrated in Fig. 2. From the figure, given all students’exercising records S with the corresponding exercise materi-alsE, we propose a preliminaryExercise-EnhancedRecurrentNeuralNetwork (EERNN) framework and an improvedExer-cise-aware Knowledge Tracing (EKT) framework. Then, weconduct two applications with the trained models. Specifi-cally, EERNN directly achieves the goal of student perfor-mance prediction on future exercises given her sequentialexercising records, and EKT is further capable of explicitlytracking the knowledge acquisition of students.

4 EERNN: EXERCISE-ENHANCED RECURRENT

NEURAL NETWORK

In this section, we first describe the Exercise-EnhancedRecurrent Neural Network framework that could directlyachieve the primary goal of predicting student performance.

EERNN is a general framework where we can predict stu-dent performance based on different strategies. Specifically,as shown in Fig. 3, we propose two implementations underEERNN, i.e., EERNNM with Markov property and EERNNAwith Attention mechanism. Therefore, both models have thesame process for modeling student’s exercising records yetfollow different prediction strategies.

4.1 Modeling Process of EERNN

The goal of the modeling process in EERNN framework isto model each student’s exercising sequence (with the inputnotation s). From Fig. 3, this process contains two maincomponents, i.e., Exercise Embedding (marked orange) andStudent Embedding (marked blue).

Exercise Embedding. Given the exercising process of a cer-tain student s ¼ fðe1; r1Þ; ðe2; r2Þ; . . .; ðeT ; rT Þg, as shown inFig. 3, Exercise Embedding learns the semantic representation/encoding xi of each exercise from its text content eiautomatically.

Fig. 4 shows the detailed techniques of Exercise Embed-ding. It is an implementation of a recurrent neural network,which is inspired by the typical one called Long Short-TermMemory (LSTM) [33] with minor modifications. Specifically,given the exercise content with the M words sequenceei ¼ fw1; w2; . . .; wMg, we first take word2vec [36] pre-trainedon an exercise corpus to transform each word wi in exerciseei into a d0-dimensional word embedding vector (we willdiscuss it in detail in Section 7.2). After the initialization,Exercise Embedding updates the hidden state vm 2 Rdv ofeach word wm at the mth word step with the previous hid-den state vm�1 in a formula as:

im ¼ sðZEwiwm þ ZE

vivm�1 þ bEi Þ;

fm ¼ sðZEwfwm þ ZE

vfvm�1 þ bEf Þ;

om ¼ sðZEwowm þ ZE

vovm�1 þ bEoÞ;

cm ¼ fm � cm�1 þ im � tanhðZEwcwm þ ZE

vcvm�1 þ bEc Þ;

vm ¼ om � tanhðcmÞ;

(1)

where im; fm; om represent the three gates, i.e., input, forget,output, respectively. cm is a cell memory vector. sðxÞ is the

Fig. 2. An overview of the proposed solution.

Fig. 3. The architectures of two implementations based on EERNN framework, where the shaded and unshaded symbols denote the observed andlatent variables, respectively.

LIU ETAL.: EKT: EXERCISE-AWARE KNOWLEDGE TRACING FOR STUDENT PERFORMANCE PREDICTION 103

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 5: EKT: Exercise-Aware Knowledge Tracing for Student ...

non-linear sigmoid activation function and � denotes the ele-ment-wise product between vectors. Besides, the inputweighted matrices ZE

w� 2 Rdv�d0 , recurrent weighted matri-ces ZE

v� 2 Rdv�dv and bias weighted vectors bE� 2 Rdv are all

the network parameters in Exercise Embedding.Traditional LSTM model learns each word representa-

tion by a single direction network and can not utilize thecontextual texts from the future word token [41]. To makefull use of the contextual word information of each exercise,we build a bidirectional LSTM considering the wordsequence in both forward and backward directions. Asillustrated in Fig. 4, at each word step m, the forward layerwith hidden word state v!m is computed based on both theprevious hidden state v!m�1 and the current word wm; whilethe backward layer updates hidden word state v m withthe future hidden state v mþ1 and the current word wm. As aresult, each word’s hidden representation vm can be calcu-lated with the concatenation of the forward state and back-ward state as vm ¼ concatenateð v!m; v

mÞ.

After that, to obtain the whole semantic representation ofexercise ei, we exploit the element-wise max pooling opera-tion to merge M words’ contextual representations into aglobal embedding xi 2 R2dv as xi ¼ maxðv1; v2; . . .; vMÞ.

It is worth mentioning that Exercise Embedding directlylearns the semantic representation of each exercise from itstext without any expert encoding. It can also automaticallycapture the characteristics (e.g., difficulty) of exercises anddistinguish their individual differences.

Student Embedding. After obtaining each exerciserepresentation xi from the text content ei by ExerciseEmbedding, Student Embedding aims at modeling thewhole exercising process of students and learning thehidden representations of students, which we called stu-dent states, at different exercising steps combined withthe influence of student performance in the history. Asshown in Fig. 3, EERNN assumes that the student statesare influenced by both the exercises and the correspond-ing scores she got.

Along this line, we exploit a recurrent neural network forStudent Embedding with the input of a certain student’sexercising process s ¼ fðx1; r1Þ; ðx2; r2Þ; . . .; ðxT ; rT Þg. Specif-ically, at each exercising step t, the input to the network is acombined encoding with both exercise embedding xt andthe corresponding score rt. Since students getting rightresponse (i.e., score 1) and wrong response (i.e., score 0) tothe same exercise actually reflect their different states, weneed to find an appropriate way to distinguish these differ-ent effects for a specific student.

Methodology-wise, we first extend the score value rt to afeature vector 0 ¼ ð0; 0; . . .; 0Þ with the same 2dv dimensionsof exercise embedding xt and then learn the combined inputvector ext 2 R4dv as

ext ¼ ½xt � 0� if rt ¼ 1;½0� xt� if rt ¼ 0;

�(2)

where � is the operation that concatenates two vectors.With the combined exercising sequence of a student

s ¼ fex1; ex2; . . .; exTg, the hidden student state ht 2 Rdh at herexercising step t is updated based on the current input ext

and the previous state ht�1 in a recurrent formula as

ht ¼ RNNðext; ht�1; uhÞ: (3)

In the literature, there are many variants of the RNNforms [33], [42]. In this paper, considering the fact that thelength of student’s exercising sequence can be long, we alsoimplement Eq. (3) by the sophisticated LSTM form, i.e.,ht ¼ LSTMðext; ht�1; uhÞ, which could preserve more long-term dependency in the sequence as

it ¼ sðZSexiext þ ZShiht�1 þ bS

i Þ;ft ¼ sðZSexfext þ ZS

hfht�1 þ bSf Þ;

ot ¼ sðZSexoext þ ZShoht�1 þ bS

oÞ;ct ¼ ft � ct�1 þ it � tanhðZSexcext þ ZS

hcht�1 þ bSc Þ;

ht ¼ ot � tanhðctÞ;

(4)

where ZSex� 2 Rdh�4dv ;ZSh� 2 Rdh�dh and bS

� 2 Rdh are theparameters in Student Embedding.

Particularly, the input weight matrix ZSex� 2 Rdh�4dv in

Eq. (4) can be divided into two parts, i.e., the positive one

ZSþex� 2 Rdh�2dv and the negative one ZS�ex� 2 Rdh�2dv , which

can separately capture the influences of exercise ei withboth right and wrong responses for a specific student dur-ing her exercising process. Based on these two types ofparameters, Student Embedding can naturally model theexercising process to obtain student states by integratingboth the exercise contents and the response scores.

4.2 Prediction Output of EERNN

After modeling the exercising process of each student fromexercising step 1 to T , we now introduce the detailed strate-gies of predicting her performance on exercise eTþ1. Psycho-logical results have claimed that student-exercise performances depend on both the student states and the exercisecharacteristics [9]. Following this finding, we propose twoimplementations of prediction strategies under EERNNframework, i.e., a straightforward yet effective EERNNMwithMarkov property and a more sophisticated EERNNA withAttention mechanism, based on both the learned student statesfh1; h2; . . .; hTg and the exercise embeddings fx1; x2; . . .; xTg.

EERNNM with Markov Property. For a typical sequentialprediction task, Markov property is a well understood andwidely used theory which assumes that the next statedepends only on the current state and not on the sequencesthat precede it [24]. Given this theory, as shown in Fig. 3a,when an exercise eTþ1 at step T þ 1 is posted to a student,EERNNM (1) assumes that the student applies current state

Fig. 4. Exercise Embedding for exercise ei.

104 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 33, NO. 1, JANUARY 2021

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 6: EKT: Exercise-Aware Knowledge Tracing for Student ...

hT to solve the exercise; (2) leverages Exercise Embedding toextract the semantic representation xTþ1 from exercise texteTþ1; (3) predicts her performance erTþ1 on exercise eTþ1 asfollowing formulas:

yTþ1 ¼ ReLUðW1 � ½hT � xTþ1� þ b1Þ;erTþ1 ¼ sðW2 � yTþ1 þ b2Þ;(5)

where yTþ1 2 Rdy denotes the overall presentation for pre-diction at (T þ 1)th exercising step. {W1;W2;b1;b2} are theparameters. sðxÞ is the Sigmoid activation functionsðxÞ ¼ 1

1þexpð�xÞ and � is the concatenation operation.

EERNNM presents a straightforward yet effective wayfor student performance prediction. However, in mostcases, since the current student state hT is the last hiddenstate of the LSTM-based architecture in Student Embedding,it may discard some important information when thesequence is too long, which is called the Vanish problem [43].Thus, the student states learned by EERNNMmay be some-what unsatisfactory for future performance prediction. Toaddress this question, we further propose another sophisti-cated prediction strategy, i.e., EERNNA with Attention mech-anism, to enhance the effects of important student states inthe exercising process for prediction.

EERNNA with Attention Mechanism. In Fig. 1, studentsmay get similar scores on similar exercises, e.g., student s1answers the exercises e1 and e3 right due to the possible rea-son that the both exercises are similar because of the sameknowledge concept “Function”.

According to this observation, as the red lines illustratedin Fig. 3b, EERNNA assumes that the student state at(T þ 1)th exercising step is a weighted sum aggregation ofall historical student states based on the correlationsbetween exercise eTþ1 and the historical ones fe1; e2; . . .; eTg.Formally, at next step T þ 1, we define the attentive statevector hatt of student as

hatt ¼XTj¼1

ajhj; aj ¼ cosðxTþ1; xjÞ; (6)

where xj is the exercise embedding at jth exercising stepand hj is the corresponding student state in the history.Cosine Similarities aj are denoted as the attention scores formeasuring the importance of each exercise ej in the historyfor new exercise eTþ1.

After obtaining attentive student state at step T þ 1,EERNNA predicts the performance of this student on exer-cise eTþ1 with the similar operation in Eq. (5) by replacinghT with hatt.

Particularly, through Exercise Embedding, our attentionscores aj not only measure the similarity between exercisesfrom syntactic perspective but also capture the correlationsfrom semantic view (e.g., difficulty correlation), benefitingstudent state representation for student performance pre-diction and model explanation. We will conduct the experi-mental analysis for this attention mechanism (Section 7.4).

4.3 Model Learning

The whole parameters to be updated in both proposed mod-els mainly come from three parts, i.e., parameters in ExerciseEmbedding fZE

w�;ZEv�;b

E� g, parameters in Student Embedding

fZSex�;ZSh�;b

S� g and parameters in Prediction Output fW�;b�g.

The objective function of EERNN is the negative log likeli-hood of the observed sequence of student’s exercising pro-cess from step 1 to T . Formally, at tth step, let ert be thepredicted score on exercise et through EERNN framework,rt is the actual binary score, thus the overall loss for a certainstudent is defined as

L ¼ �XTt¼1ðrtlog ert þ ð1� rtÞlog ð1� ertÞÞ: (7)

The objective function is minimized by the Adam optimi-zation [44]. Details will be specified in the experiments.

5 EKT: EXERCISE-AWARE KNOWLEDGE TRACING

EERNN can effectively deal with the problem of predictingstudent performance on future exercises. However, duringthe modeling, we just summarize and track a student’sknowledge states on all concepts in one integrated hiddenvector (i.e., ht in Eq. (4)), and this is sometimes unsatisfiedbecause it is hard to explicitly explain how much she hasmastered a certain knowledge concept (e.g., “Function”). Infact, during the exercising process of a certain student,when an exercise is given, she usually applies her relevantknowledge to solve it. Correspondingly, her performanceon the exercise, i.e., whether or not she answers it right, canalso reflect how much she has mastered the knowledge [5],[6]. For example, we could conclude that the student inFig. 1 has well mastered the “Function” and “Inequality”concepts but needs to devote more energy to the less famil-iar one “Probability”. Thus, it is valuable if we could remindher about this finding so that she could prepare the targettraining about “Probability” herself. Based on the aboveunderstanding, in this section, we further address the prob-lem of tracking student’s knowledge acquisition on multipleexplicit concepts. We extend the current EERNN and pro-pose an explainable Exercise-aware Knowledge Tracingframework by incorporating the information of knowledgeconcepts existed in each exercise.

Specifically, we extend the knowledge states of a certainstudent from the integrated vectorial representation inEERNN, i.e., ht 2 Rdh , to a matrix with multiple vectors, i.e.,Ht 2 Rdh�K , where each vector represents how much shehas mastered an explicit knowledge concept (e.g.,“Function”). Meanwhile, in EKT, we assume the student’sknowledge state matrix Ht changes over time influenced byboth text content (i.e., et) and knowledge concept (i.e., kt) ofeach exercise. Fig. 5 illustrates the overall architecture ofEKT. Comparing it with EERNN (Fig. 3), besides the Exer-cise Embedding module, another module (marked green),which we called Knowledge Embedding, is incorporated in themodeling process. With this additional facility, we can natu-rally extend the proposed prediction strategies EERNNMand EERNNA to EKTM with Markov property and EKTAwith Attention mechanism, respectively. In the following, wefirst introduce the way to implement the Knowledge Embed-ding module, followed by the details of EKTM and EKTA.

Knowledge Embedding. Given the student’s exercising pro-cess s ¼ fðk1; e1; r1Þ; ðk2; e2; r2Þ; . . .; ðkT ; eT ; rT Þg, the goal ofKnowledge Embedding is to explore the impacts of each

LIU ETAL.: EKT: EXERCISE-AWARE KNOWLEDGE TRACING FOR STUDENT PERFORMANCE PREDICTION 105

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 7: EKT: Exercise-Aware Knowledge Tracing for Student ...

exercise on improving student states from this exercise’sknowledge concepts kt, and this impact weight isdenoted by bt. Intuitively, at step t, if this exercise isrelated to the ith concept, we can just consider theimpact of this specific concept without others’ influences,i.e., b

jt ¼ 1 if j ¼ i, otherwise b

jt ¼ 0, 1 i; j K. How-

ever, in educational psychology, some findings indicatethat the knowledge concepts in one specific domain (e.g.,mathematics) are not isolated but contain correlationswith each other [6]. Hence, in our modeling, we assumethat learning one concept, for a certain student, couldalso affect the acquisition of other concepts. Thus, it isnecessary to quantify these correlation weights amongall K concepts in the knowledge space.

Along this line, as the module (marked in green) shownin Fig. 5, we investigate and propose a static memory net-work for calculating knowledge impact bt. Specifically, it isinspired by the memory-augmented neural network [45],[46], which has been successfully adopted in many applica-tions, such as question answering [47], language model-ing [48] and one-shot learning [49]. It usually contains anexternal memory component that can store the stable infor-mation. Then, during the sequence, it can read each inputand write the storage information from the memory forinfluencing its long-term dependency. Considering thisproperty, we set up a memory module with a matrixM 2 Rdk�K to store the representations of K knowledgeconcepts by dk-dimensional features.

Mathematically, at each exercising step t, when an exer-cise et comes, we first set its knowledge concept to be a one-hot encoding kt 2 f0; 1gK with the dimension equaling tothe total number K of all concepts. Since the intuitive one-hot representation is too sparse for modeling [50], we utilizean embedding matrix Wk 2 RK�dk to transfer the initialknowledge encoding kt into a low-dimensional vectorvt 2 Rdk with continuous values as: vt ¼Wk

Tkt.After that, the impact weight bi

tð1 i KÞ on the ithconcept from exercise et’s knowledge concept kt is furthercalculated by the Softmax operation of the inner productbetween the given concept encoding vt and each knowledgememory vector in the memory moduleMi as

bit ¼ SoftmaxðvTt MiÞ ¼ expðvTt MiÞPK

i¼1ðexpðvTt MiÞÞ: (8)

Student Embedding.With the knowledge impact bt of eachexercise, an improved Student Embedding will further spec-ify each knowledge acquisition of a certain student duringher exercising process. Thus, EKT could naturally trackstudent’s knowledge states on multiple concepts simulta-neously, benefiting the interpretability.

Methodology-wise, at the exercising step t, we also updateone of a student’s specific knowledge state Hi

t 2 Rdhð1 i KÞ by the LSTM network after she answers the exercise et

Hit ¼ LSTMðexi

t;Hit�1; uHiÞ; (9)

here we replace the original input ext with a new joint one exit

which is computed in the formula as: exit ¼ bitext, where ext is

the same encoding that combines the effects of both theexercise et she practices and the score rt she gets (Eq. (2)).

After modeling student’s historical exercising process, inthe prediction part of EKT, the performance of each studentis predicted based on three types of factors, i.e., her histori-cal knowledge states fH1; H2; . . . ; HTg, the embeddings ofthe exercises she practiced fx1; x2; . . . ; xTg, and the materi-als kTþ1 and eTþ1 of the candidate exercise.

EKTM with Markov Property. Similar to EERNNM, EKTMfollows the straightforward Markov property that assumesstudent performance on further exercise only depends onher current knowledge state HT . Specifically, as shown inFig. 5a, when the exercise eTþ1 is posted, EKTM first integra-tes student’s mastery on this exercise with its knowledgeimpacts bTþ1 as

sTþ1 ¼XKi¼1

biTþ1HiT ; (10)

then predicts her performance exTþ1 by changing the similaroperation in Eq. (5) as

yTþ1 ¼ ReLUðW3 � ½sTþ1 � xTþ1� þ b3Þ;erTþ1 ¼ sðW4 � yTþ1 þ b4Þ;(11)

where {W3;W4;b3;b4} are the parameters.

Fig. 5. The architectures of two implementations based on EKT framework, where the shaded and unshaded symbols denotes the observed andlatent variables, respectively.

106 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 33, NO. 1, JANUARY 2021

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 8: EKT: Exercise-Aware Knowledge Tracing for Student ...

EKTA with Attention Mechanism. EKTA also follows thesophisticated Attention mechanism to enhance the effect ofimportant states in the history for predicting student’s futureperformance, which is shown in Fig. 5b. Here, a small modifi-cation comparedwith EERNNA is thatwe extend the attentivestate vector hatt of student (Eq. (6)) to a matrix oneHatt, whereeach knowledge state slotHi

attð1 i KÞ can be computed as

Hiatt ¼

XTj¼1

ajHij; aj ¼ cosðxTþ1; xjÞ: (12)

Then, EKTA generates the prediction on exercise eTþ1 withEqs. (10) and (11) by replacingHT withHatt.

After that, we can train EKT by minimizing the sameobjective function in Eq. (7). Please note that during ourmodeling, EKT framework could enhance the interpretabil-ity of the learned matrix Ht through the impact weight bt,which could tell us the mastery levels on each concept of acertain student at exercising step t. We will discuss thedetails in the next section.

6 APPLICATION

After discussing the training stage of both EERNN and EKT,we now present the way to apply EERNN and EKT basedmodels to achieve two motivating goals, i.e., student perfor-mance prediction and knowledge acquisition tracking.

Student Performance Prediction. As one of the primaryapplications in intelligent education, student performanceprediction helps provide better proactive services to stu-dents, such as personalized exercise recommendation [8].Both EERNN and EKT can directly achieve this goal.

Specifically, with the trained EERNN (EKT) model M,given an individual student and her exercising recordsp ¼ fðkp1; ep1; rp1Þ; ðkp2; ep2; rp2Þ; . . .; ðkpT ; epT ; rpT Þg, we could pre-dict her performance on the next exercise epTþ1 by the follow-ing steps: (1) apply modelM to fit her exercising process sp

to get the student state at step T for prediction (i.e., hpT in

EERNNM or HpT in EKTM); (2) extract exercise representa-

tion xpTþ1 and knowledge impact bTþ1 by Exercise Embedding

and Knowledge Embedding; (3) predict her performance erpTþ1with Eq. (5) (Eq. (11)). Similarly, EERNNA (EKTA) gener-ates the prediction by replacing hp

T (HpT ) with hp

att (Hpatt).

Please note that student sp can be either anyone thatexists in the training stage or a new student that has nevershowed up. Equally, exercise epi in sp can also be either alearned exercise or any new exercise. Specifically, when anew student without any historical record is coming, at step1, EERNN (EKT) can model her first state h1 (H1) and makeperformance prediction by the non-personalized prior h0 inFig. 3 (H0 in Fig. 5), i.e., the state generated from all trainedstudent records. After that, EERNN (EKT) can fit her ownexercising process and make personalized predictions onthe following exercises. Similarly, when a new exercise iscoming, Exercise Embedding (Knowledge Embedding) inEERNN (EKT) can learn its representation (impact) onlybased on its original content (concept). Last but not least, allthe prediction part of EERNN (EKT) do not require anymodel retraining. Therefore, EERNN (EKT) can naturallydeal with the cold-start problem when making predictionsfor new students and new exercises.

Knowledge Acquisition Tracking. It is of great importance toremind students about how much they have mastered eachknowledge concept (e.g., with the mastery level rangesfrom 0 to 1) as they can be motived to conduct the targettraining in time for practicing more efficiently [7]. As men-tioned earlier, the EKT framework has a good ability totrack student’s knowledge acquisition with the learnedstates fH1; H2; . . . ; HTg. Inspired by [6], we introduce theway to estimate the knowledge mastery level of students.

In the prediction part, at each step t, please note thatEq. (11) predicts student performance on a specific exercise etfrom two kinds of inputs: the student’s integratedmastery forthis exercise (i.e., st) and the individual exercise embedding(i.e., xt). Thus, if we just want to estimate her mastery of the i-th specific concept without any exercise input, we can changest by her state inHt on this concept (i.e.,Hi

t ), and meanwhile,omit the input exercise embedding xt. Fig. 6 shows thedetailed process of this mastery level estimation on knowl-edge concepts. Specifically, given a student’s exercisingrecord s ¼ fðk1; e1; r1Þ; ðk2; e2; r2Þ; . . .; ðkT ; eT ; rT Þg, we firstobtain her knowledge state Ht at step t by fitting the recordfrom 1 to twith the trainedEKT. Then, to estimate hermasteryof the ith specific concept, we construct the impact weightbt ¼ ð0; . . . ; 1; . . . ; 0Þ, where the value in ith dimension equalsto 1, and also extract her knowledge stateHi

t on ith concept byEq. (10). After that, we can change Eq. (11) and finally estimatehermastery level lit by

yit ¼ ReLUðW3 � ½Hit � 0� þ b3Þ;

lit ¼ sðW4 � yit þ b4Þ;(13)

where 0 ¼ ð0; 0; . . . ; 0Þ is a masked exercise embedding withthe same dimension as xTþ1 in Eq. (11). The given input{W3;W4;b3;b4} are the same to those in Eq. (11) withoutany retraining of EKT.

Moreover, when estimating the knowledge mastery of stu-dents by EKT, we can also endow the correspondencebetween each learned vector (i.e., inM andHt (Fig. 5)) and theknowledge concept. Recall that in the training process, we getthe impact weight bt (Eq. (8)) based on the concept memoryand each given concept kt. Thus we can infer the meaning ofthe ith slot vectorMi if b

it is calculated with the highest value

in bt given a specific concept (e.g., “Function”), i.e.,Mi storesthe hidden information of “Function”. Correspondingly, thevector Hi

t represents the student’s knowledge state on thatconcept “Function” at step t. Therefore, after training, thechange of a student’s mastery level lit (Eq. (13), computed byHi

t ) could naturally reflect her mastery level on “Function”.We will conduct the detailed analysis for this estimation inthe experiments (Section 7.5).

Fig. 6. Mastery level estimation on knowledge concepts at step t.

LIU ETAL.: EKT: EXERCISE-AWARE KNOWLEDGE TRACING FOR STUDENT PERFORMANCE PREDICTION 107

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 9: EKT: Exercise-Aware Knowledge Tracing for Student ...

7 EXPERIMENTS

In this section, we conduct extensive experiments to demon-strate the effectiveness of EERNN and EKT frameworksfrom various aspects: (1) the prediction performance ofthem against the baselines; (2) the effectiveness of attentionmechanism in EERNNA and EKTA; (3) the illustration oftracking student’s knowledge states by EKT; (4) meaningfulvisualizations for student performance prediction.

7.1 Experimental Dataset

The dataset supplied by iFLYTEK Co., Ltd. was collectedfrom Zhixue.com1 a widely-used online learning system,which provided high school students with a large number ofexercise resources for exercising. In this paper, we conductexperiments on students’ records on mathematics becausethe mathematical dataset is currently the largest and mostcomplete in the system. To make sure the reliability of exper-imental results, we filtered the students that practiced lessthan 10 exercises and the exercises that no students haddone, and totally, over 5 million exercising records of 84,909students and 15,045 exercises were remained.

It is worth mentioning that our dataset contains a 3-leveltree-based structural knowledge system labeled by experts,i.e., an explicit hierarchical structure [51]. Thus, each exer-cise may have multi-level knowledge concepts. Fig. 8 showsan example of the concept “Function”. In our dataset,“Function” is a 1st-level concept and can be divided intoseven 2nd-level sub-concepts (e.g., “Concept”) and furtherforty-six 3rd-level sub-concepts (e.g., “Domain & Range”).In the following experiments, we treated the 1st-level con-cepts as the types of knowledge states to be tracked for stu-dents in EKT framework and considered all the 2nd-leveland 3rd-level sub-concepts as the knowledge features insome baselines (we will discuss later in Section 7.2.2).

We summarized the statistics of the dataset before andafter preprocessing in Table 1, and illustrated some dataanalysis in Fig. 7. Note that most exercises contain less than2 knowledge concepts and features, and one specific knowl-edge concept is related to 406 exercises on average. How-ever, each exercise owns 27 contents on average. Theseobservations prove that only using concepts or features can-not distinguish different exercises very well, causing

information loss. Thus, it is necessary to incorporate theexercise content for tracking students’ exercising process.

7.2 Experimental Setup

In this subsection, we clarify the implementation details toset up our EERNN and EKT frameworks. Then, we intro-duce the comparison baselines and evaluation metrics.

7.2.1 Implementation Details

Word Embedding. The first step is to initialize each wordrepresentation for exercise content. Please note that theword embeddings of mathematical exercises in ExerciseEmbedding are different from traditional ones, like news,because there are some mathematical formulas in the exer-cise texts. Therefore, to preserve the mathematics semantics,we developed a formula tool [52] to transform each formulainto its TEX code features. For example, the formula“

ffiffiffiffiffiffiffiffiffiffiffix� 1p

” would be the tokens of “nsqrt, {, x, �, 1, }”. Afterthis initialization, each exercise was transformed into a con-tent sequence with both vocabulary words and TEX tokens.(Fig. 7c illustrates the distribution of content length of theexercises.) Next, to extract the exclusive word embeddingsfor mathematics, we constructed a corpus of all 1,825,767exercises as shown in Table 1 and trained each word inthese exercises into an embedding vector with 50 dimen-sions (i.e., d0 = 50) by the public word2vec tool [36].

Framework Setting.We now specify the network initializa-tions in EERNN and EKT. We set the dimension dv of hid-den states in Exercise Embedding as 100, dh of hidden states

Fig. 7. Dataset Analysis: Number distribution of observed data.

Fig. 8. An example of the 3-level tree-based structural knowledge sys-tem on “Function” concept in our dataset. The 1st-level ‘Function” totallycontains 7 2nd-level concepts and 46 3rd-level concepts. For better illus-tration, we only show parts of the knowledge system.1. https://www.zhixue.com/

108 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 33, NO. 1, JANUARY 2021

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 10: EKT: Exercise-Aware Knowledge Tracing for Student ...

in Student Embedding as 100, dk of knowledge encoding inKnowledge Embedding as 25, and dy of the vectors for overallpresentation in prediction stage as 50, respectively. More-over, we set the number K of concepts to be tracked in EKTas 37 according to the statistics in Table 1.

Training Setting. We followed [53] and randomly initial-ized all parameters in EERNN and EKT with uniform distri-bution in the range ð� ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi

6=ðniþ noÞp;

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi6=ðniþ noÞp Þ, where

ni and no denoted the neuron numbers of feature input andresult output, respectively. Besides, we set mini batches as32 for training and also used dropout (with probability 0.1)to prevent overfitting.

7.2.2 Comparison Baselines

To demonstrate the effectiveness of our proposed frame-works, we compared our two EERNN based models, i.e.,EERNNM and EERNNA, and two EKT based models, i.e.,EKTM and EKTA, with many baselines from various perspec-tives. Specifically, we chose two models from educationalpsychology, i.e., Item Response Theory (IRT), Bayesian Knowl-edge Tracing (BKT), and three data mining models, i.e., Proba-bilistic Matrix Factorization (PMF), Deep Knowledge Tracing(DKT), Dynamic Key-Value Memory Networks (DKVMN) forcomparison. Then, to highlight the effectiveness of ExerciseEmbedding in our models, i.e., validating whether or not it iseffective to incorporate exercise texts for the prediction, weintroduced two variants, which are denoted as LSTMM andLSTMA. The details of them are as follows:

� IRT: IRT is a popular cognitive diagnostic model thatmodels student’s exercising records by a logistic-likefunction [15].

� BKT: BKT is a typical knowledge tracing modelwhich assumes the knowledge states of each stu-dent as a set of binary variables and traces themseparately with a kind of hidden Markovmodel [5].

� PMF: PMF is a factorization model that projects stu-dents and exercises into latent factors [30].

� DKT: DKT is a deep learning method that leveragesrecurrent neural network (RNN and LSTM) to modelstudents’ exercising process for prediction [11]. Theinputs are the one-hot encodings of student-knowl-edge representations.

� DKVMN: DKVMN is a state-of-the-art deep learningmethod that could track student states on multiple

concepts [6]. It contains a key matrix to store conceptrepresentation and a value matrix for each student toupdate the states. However, it does not consider theeffect of exercise content in the modeling.

� LSTMM: LSTMM is a variant of EERNN framework.Here, in the modeling process, we do not embedexercises from their contents, and only representthem as the one-hot encodings with both 2nd-leveland 3rd-level knowledge features.2 Then we lever-age traditional LSTM to model students’ exercisingprocess. For prediction, LSTMM follows Markovproperty strategy similar to EERNNM.

� LSTMA: LSTMA is another variant of EERNN frame-work which contains the same modeling process asLSTMM. For prediction, LSTMA follows the strategyof Attention mechanism similar to EERNNA.

For better illustration, we list the detailed characteristicsof these models in Table 2. More specifically, in the experi-ments, we used the open source to implement the BKTmodel,3 and implemented all other models by PyTorch on aLinux server with four 2.0 GHz Intel Xeon E5-2620 CPUsand a Tesla K20m GPU. All models were tuned to have thebest performance to ensure the fairness.

7.2.3 Evaluation Metrics

A qualified model for student performance predictionshould have good results from both regression and classifi-cation perspectives. In this paper, we evaluated the predic-tion performance of all models using four widely-usedmetrics [54], [55], [56], [57], [58].

From the regression perspective, we selected Mean Abso-lute Error (MAE) and Root Mean Square Error (RMSE), toquantify the distance between predicted scores and theactual ones. The smaller the values are, the better results wehave. Besides, we treated the prediction problem as a classi-fication task, where an exercising record with score 1 (0)indicates a positive (negative) instance. Thus, we used twometrics, i.e., Prediction Accuracy (ACC), Area Under an ROCCurve (AUC), for measuring. Generally, the value 0.5 ofAUC or ACC represents the performance prediction resultby randomly guessing, and the larger, the better.

7.3 Student Performance Prediction

Prediction in General Scenario. In this subsection, we comparethe overall performance of all models on student performanceprediction. To set up the experiments, we partitioned the data-set from the student’s perspective, where the exercisingrecords of each student are divided into training set and testingset with different percentages. Specifically, for a certain stu-dent, we used her first 60, 70, 80, 90 percent exercising records(with the exercises she practiced and the scores she got) astraining sets, and the remains were for testing, respectively.We repeated all experiments 5 times and report the averageresults using allmetrics.

Fig. 9 shows the overall results on this task. There areseveral observations. First, all our proposed EKT based

TABLE 1The Statistics of Mathematics Dataset

Statistics Original Pruned

# of records 68,337,149 5,596,075# of students 1,100,726 84,909# of exercises 1,825,767 15,045# of knowledge concepts 37 37# of knowledge features 550 447Avg. exercising records per student n 65.9Avg. content length per exercise n 27.3Avg. knowledge concepts per exercise n 1.12Avg. knowledge features per exercise n 1.8Avg. exercises per knowledge concept n 406.6

2. The one-hot representation is a typical manner in many models.We use knowledge features for representation because the number ofthem is much larger than the 1st-level ones, ensuring the reliability.

3. https://github.com/IEDMS/standard-bkt

LIU ETAL.: EKT: EXERCISE-AWARE KNOWLEDGE TRACING FOR STUDENT PERFORMANCE PREDICTION 109

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 11: EKT: Exercise-Aware Knowledge Tracing for Student ...

models and EERNN based models perform better thanother baseline methods. The results clearly indicate thatboth EKT and EERNN frameworks can make full use ofboth exercising records and exercise materials, benefitingthe prediction performance. Second, among our proposedmodels, we find that EKT based models (EKTA, EKTM)generate better results than EERNN based ones (EERNNA,EERNNM), indicating the effectiveness of trackingstudent’s knowledge states on multiple concepts (Ht inFig. 5) than simply modeling them with an integratedencoding (ht in Fig. 3). Third, models with Attention mecha-nism (EKTA, EERNNA, LSTMA) outperform those withMarkov property (EKTM, EERNNM, LSTMM), which dem-onstrates that it is effective to track the focused studentembeddings based on similar exercises for the prediction.Next, as our proposed models incorporate an independentExercise Embedding module for extracting exercise encodingdirectly from the text content, they outperform their var-iants (LSTMA, LSTMM) and the state-of-the-arts (DKVMN,DKT). This observation also suggests that both EKT andEERNN alleviate the information loss caused by the fea-ture-based or knowledge-specific representations in existingmethods. Last but not least, the traditional models (IRT,PMF and BKT) do not perform as well as deep learningmodels in most cases. We guess a possible reason is thatthese RNN based deep models can effectively capture thechange of student’s exercising process, and therefore, thedeep neural network structures are suitable for studentperformance prediction.

In summary, we conclude that both EKT and EERNNhave a good ability to predict student performance by taking

full advantage of both the exercising records and exercisematerials. Moreover, EKT shows the superiority of trackingstudent’smultiple knowledge states for the prediction.

Prediction on Cold-Start (New) Exercises. The task of predict-ing student performance often suffers from the “cold start”problem. Thus, in this part, we conduct detailed experimentsto demonstrate the performance of our proposed models inthis scenario from the exercise’s perspective (Experimentalanalysis on the cold-start students will be given in the follow-ing subsection). Specifically, we selected the new exercises(that never show up in training) in our experiment. Then weonly trained each model on 60, 70, 80, 90 percent trainingsets, and tested the prediction results on these new exercisesin the corresponding testing sets. Please note that, in thisexperiment, we did not change any training process and justselected the cold-start exercises for testing, thus all the mod-els do not need any retraining.

For better illustration, we reported the experimentalresults of all deep learning based models under all metrics inFig. 10. There are also similar observations as Fig. 9, whichdemonstrate the effectiveness of both EKT and EERNNframeworks once again. Clearly, from the results, EKT basedmodels, especially EKTA, perform the best, followed byEERNN based models. Also, we find that the improvementof them for prediction on new exercises are more significant.Thus, we can reach a conclusion that both EKT and EERNNwith Exercise Embedding module for representing exercisesfrom the text content could effectively distinguish the charac-teristics of each exercise. Those models are superior to LSTMbased models of using feature representation as well as thestate-of-the-art DKVMN and DKT of considering knowledge

Fig. 9. Results of student performance prediction in general scenario under four metrics.

TABLE 2Characteristics of All Models

Model Data Source Prediction Scenario KnowledgeScore Concept Content General Cold-start Tracking?

IRT [15] @ � � @ � �BKT [5] @ @ � @ � @PMF [30] @ � � @ � �DKT [11] @ @ � @ @ �DKVMN [6] @ @ � @ @ @LSTMM @ @ � @ @ �LSTMA @ @ � @ @ �EERNNM [14] @ � @ @ @ �EERNNA [14] @ � @ @ @ �EKTM @ @ @ @ @ @EKTA @ @ @ @ @ @

110 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 33, NO. 1, JANUARY 2021

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 12: EKT: Exercise-Aware Knowledge Tracing for Student ...

representation. In summary, both EKT and EERNN can dealwith the cold-start problem when predicting student perfor-mance on new exercises.

7.4 Effectiveness of Attention

As we have clarified in EKT and EERNN, EKTA (EERNNA)with Attention mechanism has a superior ability than EKTM(EERNNM) with Markov property because the former onescan track the focused student states and enhance the effectof these states when modeling each student’s exercisingprocess. To highlight the effectiveness of attention, we com-pared the performance of our proposed four models. To setup this experiment, we first divided the students into 90/10percent partitions, using the 90 percent students for trainingand the remaining 10 percent for testing. Therefore, the test-ing students never showed up in training. Then, for eachstudent in the testing process, we fitted her exercisingsequence by the trained models with different length step tfrom 0 to 180 and predicted her scores on the last 10 percentexercises in her records. We also conducted 10-fold crossvalidation to ensure the result reliability. Here, we reportedthe average performance under ACC and AUCmetrics.

Figs. 11a and 11b show the comparison results of them.From the figures, all models perform better and better as thelength of fitting sequence increases. For EERNNA andEERNNM, we find that they generate similar results whenthe fitting sequence of students is short (less than 40), how-ever, as the fitting length increases, EERNNA performs bettergradually. When the length surpasses about 60, EERNNAoutperforms EERNNM significantly. Moreover, we alsoclearly see that both EKTA and EKTM outperform EERNNAand EERNNM on both metrics, respectively. Based on thisphenomenon, we can draw the following conclusions. First,both EKTM and EERNNM are effective at the beginning of astudent’s exercising but discard some important informationwhen the sequence is long. Comparatively, EKTA and

EERNNA enhance the effect of some students’ historicalstates with the attention mechanism, benefiting the predic-tion. Second, EKT framework has better prediction ability byincorporating the information of knowledge concepts intomodeling, which is superior to EERNN. Third, notice thatour proposed EKT (EERNN) based models obtain about 0.72(0.65) on the both metrics (much better than the randomlyguessing 0.5), by the prior student state H0 in EKT (Fig. 5)and h0 in EERNN (Fig. 3), in the case of predicting the firstperformance of new students without any record (i.e., the fit-ting length is 0). Moreover, they all get better predictionswith more fitting records even if the sequence is not verylong at the first few steps. This finding also demonstratesthat both EKT and EERNN based models can guarantee theperformance in the cold-start scenario when making predic-tion for new students.

One step further, we also show the effectiveness of EKTA(EERNNA) with Attention mechanism with detailed analysisfrom a data correlation perspective, i.e., we could get betterprediction results based on the higher attention score (i.e., ain Eqs. (12) and (6)). Specifically, for predicting the perfor-mance of a certain student at one specific testing step (e.g.,the score on eTþ1), we first computed and normalized theattention scores of her historical exercises (i.e., fe1; e2; . . . ;eTg) calculated by EKTA (EERNNA) into [0, 1]. Then, wepartitioned these exercises into the low ([0, 0.33]), middle((0.33, 0.66]) and high ((0.66, 1]) groups based on attentionscores. In each group (e.g., the low), the average responsescore of the student on these exercises were used to repre-sent the response score of this group. Then, for all testingsteps of the specific student, we computed and illustratedthe euclidean Distance between the response scores in eachgroup (i.e., the low, middle, high) and the scores for predic-tion (i.e., the scores on feTþ1; eTþ2; . . .g). Finally, Fig. 12 illus-trates the distance results of all students in both scatter andbox figures. At each time step, we also added a result

Fig. 10. Results of student performance prediction on cold-start (new) exercises under four metrics.

Fig. 11. The effectiveness of attention in fitting process for testing. Fig. 12. Performance over different attention values in proposed models.

LIU ETAL.: EKT: EXERCISE-AWARE KNOWLEDGE TRACING FOR STUDENT PERFORMANCE PREDICTION 111

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 13: EKT: Exercise-Aware Knowledge Tracing for Student ...

computed with a group of 10 randomly selected exercises(namely, Random) for better illustration. From the figure, inboth EKT and EERNN models, the response scores of theexercises in high attention groups have the smallest distan-ces (largest correlation) with the score for prediction whilethe low groups are farthest. This finding demonstrates thatthe higher the attention value is, the more contribution ofthis exercise will make when predicting the response scoreon a new exercise. In conclusion, both EKT and EERNNframeworks can improve the prediction performance byincorporating the attention mechanism.

7.5 Visualizations

Visualization of Knowledge Acquisition Tracking. The impor-tant ability of EKT, which is superior to EERNN, is that itcan track student’s knowledge states on multiple conceptsto further explain the change of knowledge mastery levelsof the student, ensuring the interpretability. To make deepanalysis for this claim, we visualize the predicted masterylevels (i.e., lit in Eq. (13)) of a certain student on explicitknowledge concepts at each step during the exercising pro-cess. For better visualization, we made some preprocessingas follows. First, we selected 6 most frequent concepts thatthe student practiced since it was hard to illustrate clearly ifwe visualize all 37 concepts in one figure. Second, we justlogged students’ performance records on the knowledgeconcepts rather than distinguishing each specific exercise.In other words, if the student correctly answered an exerciseabout “Function”, we logged that she answered “Function”right. Then, we visualized the change of her states on theseconcepts modeled by EKTA (as a representative).

Fig. 13 shows the throughout results. In the left of thisfigure, the first column means the initial mastery levels of

this student (i.e., H0 at T = 0 in Fig. 5) on 6 concepts withoutany exercising, where her states differ from each other.Then, she starts exercising with the following 30 exercises onthese concepts. Meanwhile, her states on the concepts (out-put by EKTA) change gradually during the steps. Specifi-cally, when she answers an exercise right (wrong), herknowledge state on the corresponding concept increases(decreases), e.g., she acquires knowledge on “Set” after shesolves an exercise of “Set” concept at her second exercisingstep. During her exercising process, we can see that shegradually masters the concept “Set” but is incapable ofunderstanding “Function” since she does all exercises on“Set” right but fails to solve all exercises on “Function”.However, there exists an inconsistent phenomenon that hermastery level of “Function” becomes slightly lower at thethird exercising step even she answers the exercise correctly.This is because the model may not perfectly track the stu-dent with only few exercising records at the beginning, butit could get better performance if the student’s exercisingrecords are getting longer enough in the following steps.After exercising, we explain that she has well mastered theconcepts of “Set” and “Inequation”, partially mastered“Solid Geometry”, “Sequence” and “Probability”, but failedon “Function”, as illustrated in the right radar figure.

Visualization of Student Performance Prediction. BothEERNNA and EKTA also have great powers of explainingthe prediction results by the attention mechanism (i.e., theattention score a in Eqs. (6) and (12)). As an example, Fig. 14illustrates the attention scores for a student’s exercises.Here, both EERNNA and EKTA predict that the student cananswer exercise e20 correctly, because she got right answerson a similar exercise e4 in the past. Let us take into consider-ation about the exercise materials, we can conclude: (1) e4 is

Fig. 13. An example of the knowledge mastery level tracking of a certain student on six concepts during her 30 exercising steps, which is painted inthe middle matrix. Left side shows all concepts, which are marked in different colors. Top line records her performance on the 30 exercises. Rightradar figure shows her knowledge mastery levels (in the range ð0; 1Þ) on six concepts before (T = 0) and after (T = 30) exercising.

Fig. 14. Attention visualization in EERNNA and EKTA of an example student. We predict her performance on e20 based on her past 19 exerciserecords (we only show the first five exercises for better illustration). Right bars show the attention scores of two frameworks (i.e., EERNNA (blue) andEKTA (yellow)) for all exercises based on e20.

112 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 33, NO. 1, JANUARY 2021

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 14: EKT: Exercise-Aware Knowledge Tracing for Student ...

actually much more difficult than e20; (2) both e20 and e4contain the same knowledge concept “Solid Geometry”. Inaddition, EKTA endows a larger attention weight on e4 thanEERNNA, since EKTA can incorporate the exercise conceptsinto the modeling. This visualization hints that both EKTAand EERNNA are able to provide good ways for analyzingand explaining the prediction results, which is quite mean-ingful in real-world applications.

8 CONCLUSIONS

In this paper, we comprehensively studied student perfor-mance prediction problem. We first proposed a general Exer-cise-Enhanced Recurrent Neural Network frameworkexploring both student’s exercising records and the exercisecontent. Though EERNN effectively predicted student perfor-mance on future exercises, it could not track student’s knowl-edge states on multiple explicit concepts. Therefore, weextended EERNN to an explainable Exercise-aware Knowl-edge Tracing framework by further incorporating the knowl-edge concepts of each exercise. For making final predictions,we designed two strategies under both EKT and EERNN, i.e.,straightforward EKTM (EERNNM) with Markov property andsophisticated EKTA (EERNNA) with Attention mechanism.Comparatively, EKTA (EERNNA) could track the historicallyfocused states for making prediction, which was superior toEKTM (EERNNM). Finally, we conducted extensive experi-ments on a large-scale real-world dataset, and the results dem-onstrated the effectiveness and interpretability of our models.

ACKNOWLEDGMENTS

This research was partially supported by grants from theNational Natural Science Foundation of China (Grant No.s61672483, U1605251, and 91746301), the Science Foundationof Ministry of Education of China & China Mobile (No.MCM20170507), and the Iflytek joint research program. QiLiu gratefully acknowledges the support of the Young EliteScientist Sponsorship Program of CAST and the YouthInnovation Promotion Association of CAS (No. 2014299).Zhenya Huang would like to thank the China ScholarshipCouncil for their support.

REFERENCES

[1] A. Anderson, D. Huttenlocher, J. Kleinberg, and J. Leskovec,“Engaging with massive online courses,” in Proc. 23rd Int. Conf.World Wide Web, 2014, pp. 687–698.

[2] A. S. Lan, C. Studer, and R. G. Baraniuk, “Time-varying learningand content analytics via sparse factor analysis,” in Proc. 20thACM SIGKDD Int. Conf. Knowl. Discovery Data Mining, 2014,pp. 452–461.

[3] W. X. Zhao, W. Zhang, Y. He, X. Xie, and J.-R. Wen,“Automatically learning topics and difficulty levels of problemsin online judge systems,” ACM Trans. Inf. Syst., vol. 36, no. 3,2018, Art. no. 27.

[4] R. S. Baker and K. Yacef, “The state of educational data mining in2009: A review and future visions,” JEDM-J. Educational Data Min-ing, vol. 1, no. 1, pp. 3–17, 2009.

[5] A. T. Corbett and J. R. Anderson, “Knowledge tracing: Modelingthe acquisition of procedural knowledge,” User Model. User-Adapted Interaction, vol. 4, no. 4, pp. 253–278, 1994.

[6] J. Zhang, X. Shi, I. King, and D.-Y. Yeung, “Dynamic key-valuememory networks for knowledge tracing,” in Proc. 26th Int. Conf.World Wide Web, 2017, pp. 765–774.

[7] R. Grossman and E. Salas, “The transfer of training: What reallymatters,” Int. J. Training Develop., vol. 15, no. 2, pp. 103–120, 2011.

[8] G. D. Kuh, J. Kinzie, J. A. Buckley, B. K. Bridges, and J. C. Hayek,Piecing Together the Student Success Puzzle: Research, Propositions,and Recommendations: ASHE Higher Education Report, vol. 116.Hoboken, NJ, USA: Wiley, 2011.

[9] L. V. DiBello, L. A. Roussos, and W. Stout, “31A review of cogni-tively diagnostic assessment and a summary of psychometricmodels,”Handbook Statist., vol. 26, pp. 979–1030, 2006.

[10] N. Thai-Nghe, L. Drumond, A. Krohn-Grimberghe, and L. Schmidt-Thieme, “Recommender system for predicting student perform-ance,” Procedia Comput. Sci., vol. 1, no. 2, pp. 2811–2819, 2010.

[11] C. Piech, J. Bassen, J. Huang, S. Ganguli, M. Sahami, L. J. Guibas,and J. Sohl-Dickstein, “Deep knowledge tracing,” in Proc. Int.Conf. Neural Inf. Process. Syst., 2015, pp. 505–513.

[12] Q. Liu, E. Chen, H. Xiong, Y. Ge, Z. Li, and X. Wu, “A cocktailapproach for travel package recommendation,” IEEE Trans.Knowl. Data Eng., vol. 26, no. 2, pp. 278–293, Feb. 2014.

[13] K. H. Wilson, X. Xiong, M. Khajah, R. V. Lindsey, S. Zhao, Y. Karklin,E. G. Van Inwegen, B. Han, C. Ekanadham, J. E. Beck, et al.,“Estimating student proficiency: Deep learning is not the panacea,”in Proc. Neural Inf. Process. Syst. WorkshopMach. Learn. Educ., 2016, p. 3.

[14] Y. Su, Q. Liu, Q. Liu, Z. Huang, Y. Yin, E. Chen, C. Ding, S. Wei,and G. Hu, “Exercise-enhanced sequential modeling for studentperformance prediction,” in Proc. 32nd AAAI Conf. Artif. Intell.,2018, pp. 2435–2443.

[15] S. E. Embretson and S. P. Reise, Item Response Theory. London,U.K.: Psychology Press, 2013.

[16] J. De La Torre, “Dina model and parameter estimation: A didactic,”J. Educational Behavioral Statist., vol. 34, no. 1, pp. 115–130, 2009.

[17] H. Cen, K. Koedinger, and B. Junker, “Learning factors analysis-ageneral method for cognitive model evaluation and improvement,”in Proc. Int. Conf. Intell. Tutoring Syst., 2006, vol. 4053, pp. 164–175.

[18] P. I. Pavlik Jr., H. Cen, and K. R. Koedinger, “Performance factorsanalysis–a new alternative to knowledge tracing,” in Proc. Conf.Artif. Intell. Educ.: Building Learn. Syst. Care: From Knowl. Represen-tation Affect. Modelling, 2009, pp. 531–538.

[19] Q. Liu, R. Wu, E. Chen, G. Xu, Y. Su, Z. Chen, and G. Hu, “Fuzzycognitive diagnosis for modelling examinee performance,” ACMTrans. Intell. Syst. Technol., vol. 9, no. 4, 2018, Art. no. 48.

[20] H. Zhao, Q. Liu, H. Zhu, Y. Ge, E. Chen, Y. Zhu, and J. Du, “Asequential approach to market state modeling and analysis inonline P2P lending,” IEEE Trans. Syst. Man Cybern.: Syst., vol. 48,no. 1, pp. 21–33, Jan. 2018.

[21] Q. Liu, S. Wu, and L. Wang, “Multi-behavioral sequential predic-tion with recurrent log-bilinear model,” IEEE Trans. Knowl. DataEng., vol. 29, no. 6, pp. 1254–1267, Jun. 2017.

[22] Z. Ye, K. Xiao, Y. Ge, and Y. Deng, “Applying simulated annealingand parallel computing to the mobile sequential recommendation,”IEEE Trans. Knowl. Data Eng., vol. 31, no. 2, pp. 243–256, Feb. 2019.

[23] H. Zhao, B. Jin, Q. Liu, Y. Ge, E. Chen, X. Zhang, and T. Xu, “Voiceof charity: Prospecting the donation recurrence & donor retentionin crowdfunding,” IEEE Trans. Knowl. Data Eng., to be published,doi: 10.1109/TKDE.2019.2906199.

[24] L. Rabiner and B. Juang, “An introduction to hidden Markov mod-els,” IEEEASSPMag., vol. ASSPM-3, no. 1, pp. 4–16, Jan. 1986.

[25] Z. Pardos and N. Heffernan, “KT-IDEM: Introducing item diffi-culty to the knowledge tracing model,” in Proc. User Model. Adap-tion Personalization, 2011, pp. 243–254.

[26] Y. Xu and J. Mostow, “Using logistic regression to trace multiplesub-skills in a dynamic bayes net,” in Proc. Educational Data Min-ing, 2011, pp. 241–246.

[27] M. V. Yudelson, K. R. Koedinger, and G. J. Gordon,“Individualized Bayesian knowledge tracing models,” in Proc. Int.Conf. Artif. Intell. Educ., 2013, pp. 171–180.

[28] M. Khajah, R. M. Wing, R. V. Lindsey, and M. C. Mozer, “Incor-porating latent factors into knowledge tracing to predict individ-ual differences in learning,” in Proc. 7th Int. Conf. Educational DataMining, 2014, pp. 99–106.

[29] A. T€oscher, “Collaborative filtering applied to educational datamining,” J. Mach. Learn. Res., Jan. 2010.

[30] N. Thai-Nghe, L. Drumond, T. Horv�ath, A. Krohn-Grimberghe,A. Nanopoulos, and L. Schmidt-Thieme, “Factorization techniquesfor predicting student performance,” in Educational RecommenderSystems and Technologies: Practices and Challenges, IGI Global, 2012,pp. 129–153.

[31] Y. Chen, Q. Liu, Z. Huang, L. Wu, E. Chen, R. Wu, Y. Su, andG. Hu, “Tracking knowledge proficiency of students with educa-tional priors,” in Proc. 26th ACM Int. Conf. Inf. Knowl. Manage.,2017, pp. 989–998.

LIU ETAL.: EKT: EXERCISE-AWARE KNOWLEDGE TRACING FOR STUDENT PERFORMANCE PREDICTION 113

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 15: EKT: Exercise-Aware Knowledge Tracing for Student ...

[32] S.-Y. Teng, J. Li, L. P.-Y. Ting, K.-T. Chuang, and H. Liu,“Interactive unknowns recommendation in e-learning systems,”in Proc. IEEE Int. Conf. Data Mining, 2018, pp. 497–506.

[33] A. Graves, A.-R. Mohamed, and G. Hinton, “Speech recognitionwith deep recurrent neural networks,” in Proc. IEEE Int. Conf.Acoust. Speech Signal Process., 2013, pp. 6645–6649.

[34] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classifi-cation with deep convolutional neural networks,” in Proc. Int,Conf. Neural Inf. Process. Syst., 2012, pp. 1097–1105.

[35] P. Cui, S. Liu, and W. Zhu, “General knowledge embedded imagerepresentation learning,” IEEE Trans. Multimedia, vol. 20, no. 1,pp. 198–207, Jan. 2018.

[36] T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean,“Distributed representations of words and phrases and theircompositionality,” in Proc. Int. Conf. Neural Inf. Process. Syst., 2013,pp. 3111–3119.

[37] P. Cui, X. Wang, J. Pei, and W. Zhu, “A survey on networkembedding,” IEEE Trans. Knowl. Data Eng., vol. 31, no. 5, pp. 833–852,May 2019.

[38] D. Zhu, P. Cui, D. Wang, and W. Zhu, “Deep variational networkembedding in wasserstein space,” in Proc. 24th ACM SIGKDD Int.Conf. Knowl. Discovery Data Mining, 2018, pp. 2827–2836.

[39] Z. Huang, Q. Liu, E. Chen, H. Zhao, M. Gao, S. Wei, Y. Su,and G. Hu, “Question difficulty prediction for READINGproblems in standard tests,” in Proc. 31st AAAI Conf. Artif.Intell., 2017, pp. 1352–1359.

[40] P. Chen, Y. Lu, V. W. Zheng, and Y. Bian, “Prerequisite-drivendeep knowledge tracing,” in Proc. IEEE Int. Conf. Data Mining,2018, pp. 39–48.

[41] M. Tan, C. D. Santos, B. Xiang, and B. Zhou, “LSTM-based deeplearningmodels for non-factoid answer selection,” arXiv:1511.04108,2015. [Online]. Available: https://arxiv.org/abs/1511.04108

[42] J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, “Empirical evalua-tion of gated recurrent neural networks on sequence modeling,”arXiv:1412.3555, 2014. [Online]. Available: https://arxiv.org/abs/1412.3555

[43] S. Hochreiter and J. Schmidhuber, “Long short-term memory,”Neural Comput., vol. 9, no. 8, pp. 1735–1780, 1997.

[44] D. P. Kingma and J. Ba, “Adam: A method for stochastic opti-mization,” CoRR, vol. abs/1412.6980, 2015. [Online]. Available:https://arxiv.org/abs/1412.6980

[45] A. Graves, G. Wayne, M. Reynolds, T. Harley, I. Danihelka,A. Grabska-Barwi�nska, S. G. Colmenarejo, E. Grefenstette,T. Ramalho, J. Agapiou, et al., “Hybrid computing using a neuralnetworkwith dynamic external memory,”Nature, vol. 538, no. 7626,2016, Art. no. 471.

[46] S. Sukhbaatar, J. Weston, R. Fergus, et al., “End-to-end memorynetworks,” inProc. Int. Neural Inf. Process. Syst., 2015, pp. 2440–2448.

[47] C. Xiong, S. Merity, and R. Socher, “Dynamic memory networksfor visual and textual question answering,” in Proc. Int. Conf.Mach. Learn., 2016, pp. 2397–2406.

[48] A. Kumar, O. Irsoy, P. Ondruska, M. Iyyer, J. Bradbury,I. Gulrajani, V. Zhong, R. Paulus, and R. Socher, “Ask me any-thing: Dynamic memory networks for natural language proc-essing,” in Proc. Int. Conf. Mach. Learn., 2016, pp. 1378–1387.

[49] A. Santoro, S. Bartunov, M. Botvinick, D. Wierstra, and T. Lillicrap,“Meta-learning with memory-augmented neural networks,” inProc. Int. Conf. Mach. Learn., 2016, pp. 1842–1850.

[50] Y. Goldberg and O. Levy, “word2vec explained: Derivingmikolov, et al.’s negative-sampling word-embedding method,”arXiv:1402.3722, 2014. [Online]. Available: https://arxiv.org/abs/1402.3722

[51] S. Wang, J. Tang, Y. Wang, and H. Liu, “Exploring hierarchicalstructures for recommender systems,” IEEE Trans. Knowl. DataEng., vol. 30, no. 6, pp. 1022–1035, Jun. 2018.

[52] Y. Yin, Z. Huang, E. Chen, Q. Liu, F. Zhang, X. Xie, and G. Hu,“Transcribing content from structural images with spotlightmechanism,” in Proc. 24th ACM SIGKDD Int. Conf. Knowl. Discov-ery Data Mining, 2018, pp. 2643–2652.

[53] G. B. Orr and K.-R. M€uller, Neural Networks: Tricks of the Trade.Berlin, Germany: Springer, 2003.

[54] J. Fogarty, R. S. Baker, and S. E. Hudson, “Case studies in the useof ROC curve analysis for sensor-based estimates in human com-puter interaction,” in Proc. Graph. Interface, 2005, pp. 129–136.

[55] R. Wu, G. Xu, E. Chen, Q. Liu, and W. Ng, “Knowledge or gam-ing?: Cognitive modelling based on multiple-attempt response,” inProc. 26th Int. Conf. World Wide Web Companion, 2017, pp. 321–329.

[56] T. Zhang, G. Su, C. Qing, X. Xu, B. Cai, and X. Xing, “Hierarchicallifelong learning by sharing representations and integratinghypothesis,” IEEE Trans. Syst. Man Cybern.: Syst., to be published,doi: 10.1109/TSMC.2018.2884996.

[57] K. Kuang, P. Cui, S. Athey, R. Xiong, and B. Li, “Stable predictionacross unknown environments,” in Proc. 24th ACM SIGKDD Int.Conf. Knowl. Discovery Data Mining, 2018, pp. 1617–1626.

[58] L. Zhang, K. Xiao, H. Zhu, C. Liu, J. Yang, and B. Jin, “CADEN: Acontext-aware deep embedding network for financial opinionsmining,” in Proc. IEEE Int. Conf. Data Mining, 2018, pp. 757–766.

Qi Liu received the PhD degree in computer sci-ence from USTC. He is an associate professor atthe University of Science and Technology ofChina (USTC). His general area of research isdata mining and knowledge discovery. He haspublished prolifically in refereed journals and con-ference proceedings, e.g., the IEEE Transactionson Knowledge and Data Engineering, the ACMTransactions on Information Systems, the ACMTransactions on Knowledge Discovery fromData, the ACM Transactions on Intelligent Sys-

tems and Technology, KDD, IJCAI, AAAI, ICDM, SDM, and CIKM. Hehas served regularly in the program committees of a number of confer-ences, and is a reviewer for the leading academic journals in his fields.He is a member of the ACM and IEEE. He is the recipient of the KDD2018 Best Student Paper Award (Research) and the ICDM 2011 BestResearch Paper Award. He is supported by the Young Elite ScientistSponsorship Program of CAST and the Youth Innovation PromotionAssociation of CAS.

Zhenya Huang received the BE degree in soft-ware engineering from Shandong University(SDU), China, in 2014. He is currently workingtoward the PhD degree in the School of Com-puter Science and Technology, University of Sci-ence and Technology of China (USTC). His mainresearch interests include data mining and knowl-edge discovery, recommender systems, and intel-ligent education systems. He has publishedseveral papers in referred conference proceed-ings, such as AAAI’2016, AAAI’2017, CIKM’2017,KDD’2018, AAAI’2018, and DASFAA’2018.

Yu Yin received the BE degree in computerscience from the University of Science and Technol-ogy of China (USTC), China, in 2017. He is currentlyworking toward the PhD degree in the School ofComputer Science and Technology, USTC. Hismain research interests include data mining, intelli-gent education systems, and image recognition. Hewon the first prize in the Second Student RDMAProgramming Competition in 2014. He has pub-lished papers in referred conference proceedings,such as AAAI’2018 andKDD’2018.

Enhong Chen (SM’07) received the PhD degreefrom the University of Science and Technology ofChina. He is a professor and vice dean of theSchool of Computer Science, USTC. His generalarea of research includes data mining andmachine learning, social network analysis, andrecommender systems. He has published morethan 100 papers in refereed conferences andjournals, including the IEEE Transactions onKnowledge and Data Engineering, the IEEETrans. MC, KDD, ICDM, NIPS, and CIKM. He

was on the program committees of numerous conferences includingKDD, ICDM, and SDM. He received the Best Application Paper Awardon KDD-2008, the Best Student Paper Award on KDD-2018 (Research),and the Best Research Paper Award on ICDM-2011 and Best of SDM-2015. His research is supported by the National Science Foundation forDistinguished Young Scholars of China. He is a senior member ofthe IEEE.

114 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 33, NO. 1, JANUARY 2021

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.

Page 16: EKT: Exercise-Aware Knowledge Tracing for Student ...

Hui Xiong (SM’07) is currently a full professor atRutgers, the State University of New Jersey,where he received the ICDM-2011 Best Rese-arch Paper Award, and the 2017 IEEE ICDMOutstanding Service Award. His general area ofresearch is data and knowledge engineering,with a focus on developing effective and efficientdata analysis techniques for emerging data inten-sive applications. He has published prolifically inrefereed journals and conference proceedings(four books, more than 80 journal papers, and

more than 100 conference papers). He is a co-editor-in-chief of Encyclo-pedia of GIS, an associate editor of the IEEE Transactions on Knowl-edge and Data Engineering, the IEEE Transactions on Big Data, theACM Transactions on Knowledge Discovery from Data, and the ACMTransactions on Management Information Systems. He has served reg-ularly on the organization and program committees of numerous confer-ences, including as a program co-chair of the Industrial and GovernmentTrack for KDD-2012, a program co-chair for ICDM-2013, a general co-chair for ICDM-2015, and a program co-chair of the Research Track forKDD-2018. For his outstanding contributions to data mining and mobilecomputing, he was elected an ACM Distinguished scientist, in 2014. Heis a senior member of the IEEE.

Yu Su received the PhD degree from Anhui Uni-versity. He is a researcher of IFLYTEK CO., LTD.His main area of research includes data mining,machine learning, recommender systems andintelligent education systems. He has publishedseveral papers in referred conference pro-ceedings and journals, such as IJCAI’2015,AAAI’2017, AAAI’2018, KDD’2018, CIKM’2017,DASFAA’2016, and ACM TIST.

Guoping Hu received the PhD degree from theUniversity of Science and Technology of China.He is one of the founders of iFLYTEK Co. Ltd.Currently, he is senior vice president of iFLYTEK,dean of the Research Institute of iFLYTEK, direc-tor of the National Key Laboratory of CognitiveIntelligence, vice chairman of the Strategic Alli-ance of New Generation AI Industrial TechnologyInnovation, and deputy group leader of the Over-all Group of National AI Standardization. He haspossessed 65 invention patents, and published

more than 20 papers in core journals and important international confer-ences at home and abroad.

" For more information on this or any other computing topic,please visit our Digital Library at www.computer.org/csdl.

LIU ETAL.: EKT: EXERCISE-AWARE KNOWLEDGE TRACING FOR STUDENT PERFORMANCE PREDICTION 115

Authorized licensed use limited to: University of Science & Technology of China. Downloaded on December 08,2020 at 04:06:08 UTC from IEEE Xplore. Restrictions apply.