Keywords

1 Introduction

A crucial demand in higher education is SLPP [1, 2]. Unfortunately, almost all of related researches [3,4,5,6] mainly focus on GPA prediction. They treated GPA as the only indicator for evaluating students and failed to consider one abnormal situation: there are some students whose learning patterns are clearly different from others, e.g., promising students or gifted students. Note, they are less but important, and GPA is not enough.

For identifying them, we propose an effective method with strong interpretability in educational sense, and we obtain the inspiration from regression models: students from mentioned abnormal situation are equal to outliers of training samples in regression models. Then, we aim to design a robust prediction model for the task of SLPP, and identify students in mentioned abnormal situation, i.e., ASD task. In addition, for mining underlying information, we proposed graph memory networks (GMN) model [7] as our basic model. Then, to improve the robustness of GMNs and enhance the precision of training loss, we combine self-paced learning [8] (SPL) with GMNs, and proposed SPGMN algorithm. Specially, our two tasks are combined with SPGMN for better interpretability: we train a SPGMN model for SLPP task on training samples, then, we apply the trained model to handle the ASD task.

Definition 1

Abnormal student is the student whose learning pattern is clearly different from others. And, ASD is the abbreviation of abnormal student decision. Specially, this definition is with respect to the data aspect.

Summarily, the contributions of our work are: 1) SPGMN can enhance the robustness of GMNs model in SLPP task; 2) improve the precision and interpretablity of ASD task; 3) Experimental results indicate that SPGMN can really enhance the robustness of GMNs. The ASD experiment proofs that GPA is not enough, and we propose learning pattern as an supplement.

2 Proposed Methods

Self-paced Graph Memory Network. SPL incorporates a self-paced function and a pacing parameter into the learning objective of GMN to optimize the order of samples and model parameters. Each sample is assigned to a weight to reflect the easiness of the sample. Then, these weighted samples are gradually involved into training from easy to complex, as shown as follows:

$$\begin{aligned} \min _{\mathbf {w}, \mathbf {v}} \mathbb {E}(\mathbf {w}, \mathbf {v})=\sum _{i=1}^{n} \mathbf {v}_{i} \cdot \ell _{i}(y_i, g\left( \mathbf {x}_{i}, \mathbf {w}\right) ) + f(\mathbf {v}, \lambda ^{k}) \end{aligned}$$
(1)

where \(y_i\) is the ground-truth label of the i-th graph, \(\ell (\cdot )\) is the loss function of GMN, \(\mathbf {x}_i\) is the i-th trained graph, \(\mathbf {w}\) is the parameters, \(f(\mathbf {v};k)\) is a dynamic self-paced function with respect to \( v \& k\) (e.g., \(f (v, {\lambda }^k) = {\lambda }^k(\frac{1}{2}\mathbf {v}^\mathbf {2}\ -\ \mathbf {v})\)), and \(\lambda \) is the age of the SPGMN to control the learning pace, k means k-th iteration, while \(\mathbf{v} = [v_1, v_2, ..., v_n]\) selects which samples involved into training data.

Abnormal Student Detection Based on SPGMN. Here, we introduce our ASD method based on SPGMN algorithm, i.e., SPGMN-ASD. We first train a SPGMN model for SLPP task on the training data. After enough iterations, we obtain a trained SPGMN model and the training loss of each student. Obviously, the set of losses for abnormal students (outliers) has clearly difference with the set of losses for normal students (majority of training samples). Based on this consideration, we first sort \(\alpha \) values (the loss of target student divide the sum of all losses). Then we randomly set one break point from the list of non-zero \(\alpha \) values and obtain two series of \(\alpha \). Finally, if these two series cannot satisfy t-test, the corresponding students of series with larger \(\alpha \) values are detected as abnormal students.

3 Experiments

Here, we design our experiments based on real educational dataset from X-university, GPA-data. GPA-data consists of the registration information of 600 students, exam scores for courses they had enrolled (Each student totally enrolls about 70 to 80 courses.) and the background information of all courses. Here, data is constructed in the form of graphs (one graph represents a student). The nodes are courses that students had ever enrolled, and top-5 similar nodes are connected with non-weighted edges. Specially, 600 students are classified into 4 categories by GPAs, and GPA \(\in \) [1, 2, 3, 4].

Student Learning Performance Prediction. To compare the performance of SPGMN and some compared baseline/SLPP methods, i.e., GMN [7], convolutional neural network (CNN) [4], K-Nearest Neighbor (KNN) [3], decision trees (DTs) [9], naïve Bayes (NBs) [10], we follow the experimental protocol in [11] and perform 10-fold cross-validation and report the mean accuracy over all folds.

Table 1. SLPP results. Note, GPA-k means the data of first k terms of GPA-data.

We predict the final GPAs based on the data of the first k term (\(k \in [1,7]\)), i.e., GPA-k. Table 1 shows that: (1) in all semesters, our SPGMN model consistently achieves better results w.r.t compared methods; (2) as semesters go, the advantage of SPGMN is amplified, i.e., SPGMN has strong robustness.

Abnormal Students Detection. Here, we test our ASD method on the training dataset of GPA-7. Specially, we randomly repeat SPGMN-ASD 400 times and consider the k student with highest frequency appeared as abnormal students. Note, k is the average number of results of 400 experiments.

Figure 1(a) suggests that: 1) the abnormal series and the normal series has a noticeable gap at 511-th point; 2) these two series come from two different distribution. Then, Fig. 1(b) indicates that GPA is not enough to evaluating students: 1) 100% of promising students were abnormal, i.e., they need more attention from teachers; 2) For the students whose GPA is 4, 56% of them were abnormal, i.e., those gifted students need more challenged works while the remaining 44% may not need.

Fig. 1.
figure 1

(a): Results of ASD. (b): y-axis is the percent of all students who abnormal. (c): x is the percentage of major-related courses cut of all courses. y-axis is the percentage of student with different learning pattern cut of all students. (Color figure online)

By studying the 30 outliers, we found that they are not errors. Instead, it represents a noticeable event: the strategy of course selection (i.e., learning pattern) of abnormal students is different from normal studentsFootnote 1. As shown in Fig. 1(c): 1) normal students (blue bars) are more likely to select non-major-related courses, e.g., Basic of Finance for students of materials science and technology; 2) abnormal students (orange bars) are more likely to choose major-related courses, e.g., Database for students of computer science; 3) when students enrolled more major-related courses, their learning patterns are more likely abnormal, i.e., learning patterns can be considered as a supplement of GPAs.

4 Conclusion

This paper combines SLPP task and ASD task for strong interpretability of ASD task. And we combine GMNs model and SPL for enhancing the robustness of GMNs for SLPP task and improving the precision of ASD task. Then, experiments verify the mentioned advancements. Finally, we propose the learning pattern as an supplement of GPA by analyzing the results of ASD experiment.