Introduction

This work explores the problem of students’ performance assessment under partial completion of their semester studies. During the starting of lock-downs, faced by the educational institutions in developing countries like India, students’ semester assessment emerged as a challenging and important problem [14]. Students were sent home; they were forced to study through online lectures and tutorials. During this period, many of the students did not have good Internet connections which lead to inculcation bias due to unequal participation by students. Furthermore, it was hard to conduct online examinations owing to unavailability of good infrastructures at home and challenges pertaining to curbing of unfair means during exams. This made many of the faculties adopt a binary (satisfactory/unsatisfactory) grading scheme. This situation may also happen during other uncalled for disasters or pandemics. This situation led the authors to ponder as to how modern computational intelligence (CI) technologies can be leveraged to alleviate the adverse effect of COVID-19 on education.

Course completion, organization of different exams, grades, admissions, and student psychology have been severely affected by this pandemic [3, 13]. Students worldwide were under tremendous stress due to uncertainty about their final grades. On the other hand, it has been very challenging for faculties to grade their students based on partial semester completion with little-to-no means for assigning grades. We propose to solve this problem by leveraging deep learning for automatic prediction of students’ grades. To this end, different partial course-completion durations were experimented with. Furthermore, due to unavailability of a good large-scale public dataset, a student academic performance dataset was also created and shared in the public interest.

Technology has played vital role in fighting against the COVID pandemic; however, educational plight has received little-to-no attention. This work is a humble attempt on leveraging CI methods for automatic students’ performance assessment through predictive analysis. Such an automated system has numerous applications in the education sector. It is not only useful for the present pandemic situation but can also assist in admission decisions; non-completion; dropout and retention; profiling and prediction for student’s feedback and guidance [5, 15].

Previously, operational-research based approaches have existed for solving this kind of problem. One example is the Duckworth/Lewis method for predicting adjusted target scores when the Cricket game is interrupted by rain [6]. Duckworth et al. studied thousands of Cricket match data and came out with a unique exponential function that tries to model the predicted score as a function of remaining resources. The problem of predictive performance assessment is challenging due to several reasons, and a Duckworth/Lewis approach seems naive, since, due to the complexity of the problem, finding a model analytically is not possible. Estimation of marks requires modeling different kinds of correlations. Due to the uncertainty of human behavior, it is hard to conceptualize all relations needed for assessment, for example, missing dependencies such as previous scores in a subject, previous records of a student, etc. Although absolute predictions cannot be made, and there is always room for unforeseeable events, we have worked with a basic setting that considers the available marks. The proposed model tries to intelligently capture the latent correlations in students’ performance and the complexity of subjects and other parameters.

Unprecedented pandemic circumstances and its adverse effect on education have propelled the relevance and need for a predictive assessment system based on available students’ academic data. CI algorithms are a useful tool for addressing this. So far, very little work has happened on this problem. The scope of previous approaches is limited for experimenting through classical machine learning on smaller datasets with less number of courses. Most of them have focused on predicting categorical grades. There has been little-to-no study with deep learning. Previously, researchers have used evolutionary algorithms, multi-layer perceptrons, fuzzy clustering, SVM, and random forests for the predictive analysis of students’ grades [1, 4, 7, 9, 10, 12]. A summary of the related work on performance assessment methods is presented in Table 1. For a broader literature overview, we would suggest going through [8].

Although there have been previous works with classical machine learning; however, through this paper, we present refreshing ideas into the field, leveraging deep learning. We show how deep learning can be used for students’ academic score estimation; particularly, the latent space models and their ability to model the correlation between student performance through different tests. Unlike evolutionary algorithms employed by [4], our emphasis is on gradient-based optimization algorithms, since we get better generalizations, faster training, and less computing resource requirement, and these algorithms can virtually scale to any size of the dataset. In our experiments, deep algorithms were found to perform better in comparison to machine learning approaches. Our dataset consists of 15 courses. We have shown the integration of generative and temporal deep learning approaches with machine learning algorithms, and the statistical study of the features of the dataset.

Table 1 Literature review on approaches for academic performance assessment or prediction

The proposed approach takes the partial academic performance as input and passes it through a neural network encoder to estimate the distribution of marks in a latent space. Once the training is complete, the estimated distribution has the potential to capture the correlation between the partial academic performance and the true grades. After the training of neural network encoder–decoder, the parameters estimated by the latent distribution are passed through machine learning regressors to predict the final marks. The detailed proposed approach is explained in Sect. 4. The proposed approach estimates final scores in a continuous range between (0, 100) rather than doing a grade-based categorical estimation. The main contributions of this work are:

  • We formulate a CI-based solution for the problem of academic performance estimation (APE). We have experimented with different machine or deep learning approaches and proposed a variational approach using machine learning regressors for solving this problem.

  • The APE problem requires a suitable dataset to experiment under different settings. To address the unavailability of such a dataset in the public domain, a new dataset was created and released in the public interest for benchmarking the future research.

  • We have extensively experimented and compared the proposed approach under different settings on this dataset. Including statistical assessment of the released dataset.

The rest of the paper is organized as follows. The next section describes the dataset details and attributes. The proposed approach is presented after this in Sect. 4. After this, the evaluation criteria are discussed in Sect. 5. The results and conclusion are presented under the last two sections.

Dataset

The data were collected at IIT Roorkee, India, for over 1000 students in 15 under-graduate and graduate-level courses between 2006 and 2017. It has been anonymized to protect the identities of individual students. Students’ academic performance is generally evaluated based on a set of individual parameters assessed across the comprehensive course of the evaluation. The dataset is called ‘IITR-APE dataset’, and is made available at https://www.hbachchas.github.io/data.html.

Institutions use distinct sets of exams, like end-term evaluation and multiple in-course examinations, supplemented with an estimation of the class performance of the students. Some courses also involve laboratory work with hands-on experience over real-life applications. In the wake of COVID-19, almost every educational institution is facing the problem of deferred evaluation and staggering students’ careers. This dataset is a humble effort toward the viability of recent developments in computational intelligence for the automatic assessment of students’ performance.

Let X be the set of all parameters used by the institutions for evaluation. The evaluation system customarily composites a weighted average of the collection of all parameters in X, which can be defined by the equation below

$$\begin{aligned} Y= & {} \sum _i W_i * X_i. \end{aligned}$$
(1)

In Eq. (1), \(W_i\) is the individual weight for each feature \(X_i\), where \(X_i\) is the subset of feature-set X. The total score received by a student is represented by Y. The features used in our dataset include marks obtained in two tests \(X_{\text {T1}}\) and \(X_{\text {T2}}\), assessment based on students classroom performance \(X_{\text {CW}}\), a mid-term evaluation \(X_{\text {MTE}}\), and end-term evaluation \(X_{\text {ETE}}\). Three datasets used for experimentation and evaluation consist of a subset of above-mentioned features with specific weights for a given dataset as explained below.

Dataset \({\text {D}}_1\) consists of three basic features, which are two class-test based evaluations and one assessment based on the student’s class performance. Hence, feature-set \(X_{\text {D1}}\) can be defined as, \(\langle X_{\text {T1}}, X_{\text {T2}}, X_{\text {CW}}\rangle\). To establish an experimental setup for deferred evaluation for a given academic session, the mid-term marks and end-term evaluation scores were dropped, and the remaining features were used to estimate the final score \(Y_{{\text {D}}_1}\). This analysis helps us understand how features with small weightage, in score calculations, can affect the final score.

Dataset \({\text {D}}_2\) consists of three features, viz., student’s class performance, mid-term evaluation, and end-term evaluation. Hence, feature-set \(X_{\text {D2}}\) can be defined as, \(\langle X_{\text {CW}}, X_{\text {MTE}}, X_{\text {ETE}}\rangle\). \(Y_{{\text {D}}_2}\) represents a weighted sum of features in \(X_{\text {D2}}\). Two distinct analyses were conducted over \(X_{\text {D2}}\) to predict the final scores. First, using \(X_{\text {MTE}}\) and \(X_{\text {CW}}\), and later with \(X_{\text {ETE}}\) and \(X_{\text {CW}}\). Complementing \(X_{\text {MTE}}\) with \(X_{\text {ETE}}\), helped us to analyze if deferred evaluation effects the total score estimation. On the other hand, utilizing \(X_{\text {ETE}}\) helps to understand its effect on the overall score.

Preliminaries

In this section, we discuss about popular deep learning and machine learning techniques to evaluate the datasets. For temporal evaluation, recurrent neural network (RNN) variants such as long short-term memory and gated recurrent unit (GRU) were used along with discrete machine learning classifiers, [2]. Also, we employed Variational Bayes’ encoding for extensive feature extraction with machine learning classifier to evaluate the dataset.

Long Short-Term Memory (LSTM)

Long short-term memory is a state-of-the-art neural architecture used for modeling complex long- or short-term temporal relations in sequence translation/recognition tasks. LSTM is incipiently an enhanced variant of the recurrent neural networks. LSTM structures are known for handling not only the hidden state of an RNN, but also the cell-state of each recurring block. The mathematical description of a single LSTM cell is given by Eq. (2)–(7)

$$\begin{aligned} I_j &= \sigma (\alpha _I [Z_{j-1},X_j ]+\beta _I ). \end{aligned}$$
(2)
$$\begin{aligned} F_j &= \sigma (\alpha _F [Z_{j-1},X_j ]+\beta _F ). \end{aligned}$$
(3)
$$\begin{aligned} O_j &= \sigma (\alpha _O [Z_{j-1},X_j ]+\beta _O ). \end{aligned}$$
(4)

The above equations represent the three different gates in the LSTM structure. \(\sigma ()\) represents the sigmoid function in these equations. The symbol \(\beta\) and \(\alpha\) represent bias and weight matrices, respectively, in each gate equation. \(Z_{j-1}\) represents the hidden state of the LSTM cell from the previous time step, whereas \(X_j\) represents the input to the LSTM cell at the current time step. Equations (2)–(4) represent the input gate, forget gate, and the output gate, respectively. The input gate helps to capture the new information, to be stored in the cell-state. The forget gate tells about the information; one needs to remove from the cell-state. The output gate deduces the information that we need to emit as the final output from the LSTM cell. The outcomes of these three gates are used to find the LSTM’s cell-state and hidden state, as shown in the following equations

$$\begin{aligned} {\tilde{S}}_j &= \tanh ( \alpha _S[Z_{j-1},X_j ] +\beta _S ). \end{aligned}$$
(5)
$$\begin{aligned} S_j &= F_j * S_{j-1} + I_j * {\tilde{S}}_j. \end{aligned}$$
(6)
$$\begin{aligned} Z_j &= O_j * \tanh (S_j). \end{aligned}$$
(7)

In the above equations, \(S_j\) represents the cell-state memory for a given LSTM cell at timestamp j. \(Z_j\) represents the hidden state, at timestamp j, for the LSTM cell.

LSTM has become a popular structure of choice, as it overcomes the problem of vanishing and exploding gradients in comparison to simple RNNs. In our analysis, the LSTM cells were joined with fully connected neural network layers to make predictions.

Gated Recurrent Unit (GRU)

Gated recurrent unit is a modified version of recurrent neural networks. GRU resolves the problem of vanishing gradient, which dominates in standard RNN models. It is quite similar to the LSTM structure and sometimes even gives better performance. The working of GRU cells can be understood through the equations explained below

$$\begin{aligned} H_j &= \sigma ( \alpha ^H X_j+\gamma ^H Z_{j-1} ). \end{aligned}$$
(8)

Equation (8) represents the update gate inside a GRU cell. This gate helps to learn the amount of information the GRU cell needs to pass from the previous \((j-1)\)th time step. \(\alpha ^H\) and \(\gamma ^H\) represent the weights of the new input, \(X_j\), and the information from the previous \(j-1\) time steps, \(Z_{j-1}\), respectively

$$\begin{aligned} R_j &= \sigma ( \alpha ^R X_j+\gamma ^R Z_{j-1} ). \end{aligned}$$
(9)

The above equation represents the reset gate. This gate helps the network to understand, how much of the information from the past \(j-1\) time steps, need to be forgotten

$$\begin{aligned} Z_j'= & {} \tanh ( \alpha X_j+R_j \odot \gamma Z_{j-1} ). \end{aligned}$$
(10)

Equation (10) helps in calculation of the current memory-state. This uses the input information—\(X_j\) and \(Z_{j-1}\), along with \(R_j\) from the reset gate, to store equivalent information from the past \(j-1\) and the current time step

$$\begin{aligned} Z_j= & {} H_j\odot Z_{j-1} + (1-H_t) \odot Z_j'. \end{aligned}$$
(11)

The above equation helps in the evaluation of the final hidden state information, \(Z_j\), for the current time step to pass it on to a future time step. In our experiments, the GRU units were integrated with two fully connected layers to make the grade prediction pipeline for students.

Variational Auto-encoder (VAE)

Variational auto-encoder is a common method used for feature extraction. The major difference between a simple Auto-encoder (AE) and a variational auto-encoder is that it learns the latent space variable Z in the form of a prior distribution (usually a Gaussian).

In a VAE, the distribution of latent space is mapped with a presumed distribution. This distribution is learned in the format of mean \(\mu\) and logarithmic variance \(\log \;\sigma\). To enforce this distribution to a prior distribution, we use KL divergence loss which is defined as

$$\begin{aligned} D_{\text {KL}}(N\left( \mu _i,\sigma _i\right) ,N\left( \mu _j,\sigma _j\right) ) &= \log \frac{\sigma _j}{\sigma _i} + \frac{\sigma _i^2+(\mu _i-\mu _j)^2}{2\sigma _j^2}- \frac{1}{2}. \end{aligned}$$
(12)
Fig. 1
figure 1

a Performance estimation using LSTM-based model. b Performance estimation using GRU-based model

Fig. 2
figure 2

Performance estimation using variational auto-encoder with ML regressors

The KL divergence for two mapping Normal distributions is represented by Eq. (12). Here, \(N(\mu _j,\sigma _j)\) is our prior distribution and \(N(\mu _i,\sigma _i)\) is the distribution upon which we want to enforce the prior distribution. The \(\mu _i\) and \(\sigma _i\) for the calculation are obtained from the latent space, Z, of the VAE. In our model, we have assumed the prior distribution to be N(0, 1), i.e., a Gaussian distribution with \(\mu _j\) and \(\sigma _j\) values of 0 and 1, respectively

$$\begin{aligned} D_{\text {KL}}(N\left( \mu _i,\sigma _i\right) ,N\left( 0,1\right) )= & {} \frac{1}{2}(\sigma _i^2+\mu _i^2-1-2\log \;\sigma _i). \end{aligned}$$
(13)

Equation (13) represents the KL divergence loss for, N(0, 1), prior distribution used in our model. The VAE model was trained on X and the latent space Z was extracted and further used to train the machine learning classifiers to make a final prediction of students’ grades.

Machine Learning Regressors

The modeling of input data, X, to predict the students’ grades, Y, was done using machine learning-based regressors. The regressors used during our experiments are: Multi-layer Perceptron (MLP), Linear Regression (LR), Extra Tree Regressor (ET), Random Forest Regressor (RF), XGBoost Regressor (XGB), and k-Nearest Neighbour Regressor (kNN).

The variations in results were analyzed using fivefold shuffle split cross-validation. For all our experiments, the dataset was randomly shuffled five times, and each time, a test set was drawn to estimate the performance in terms of the metric results. The mean and standard deviation of these results are reported later under the section on Results and Discussion.

Approach

The datasets mentioned in Sect. 2 include a set of different evaluation features for students’ final score estimation. In this section, we discuss about experimental approach in final grade estimation. For temporal evaluation, recurrent neural network (RNN) variants such as long short-term memory and gated recurrent unit (GRU) were used considering incoming grades a sequential data [2]. For further evaluation, we employed Variational Bayes’ encoding for extensive feature extraction with machine learning classifiers to evaluate the datasets.

Proposed Method

Initially, we perform experiments on dataset \({\text {D}}_1\) using feature set containing \(\langle X_{\text {T1}},X_{\text {T2}},X_{\text {CW}}\rangle\). We first evaluate a sequential approach using LSTM and GRU units followed by a fully connected neural network. For LSTM-based model, as shown in Fig.  1a, at every step, we take the hidden state \(Z_1, Z_2\), and \(Z_3\), and concat them to make a final feature space Z

$$\begin{aligned}&\{Z_1, Z_2, Z_3\} \longleftarrow {\text {LSTM}}(\langle X_{\text {T1}},X_{\text {T2}},X_{\text {CW}}\rangle ) \end{aligned}$$
(14)
$$\begin{aligned}&Z \longleftarrow {Z_1 \oplus Z_2 \oplus Z_3}. \end{aligned}$$
(15)

Finally, we pass these features Z through a fully connected neural network \({\text {MLP}}(Y|Z)\) that maps feature space Z to predicted performance \(\hat{Y}\)

$$\begin{aligned} \hat{Y} \longleftarrow \int _Z {\text {MLP}}(Y|Z)p(Z){\text {d}}Z. \end{aligned}$$
(16)

In the above equation, p(Z) is the probability distribution of feature space Z which is mapped to predicted performance \(\hat{Y}\). Similarly, we apply GRU model, as shown in Fig.  1b, to estimate hidden state at input time stamps \(Z_1, Z_2\) and \(Z_3\) and and concat them to make a final feature space Z

$$\begin{aligned} \{Z_1, Z_2, Z_3\} \longleftarrow {\text {GRU}}(\langle X_{\text {T1}},X_{\text {T2}},X_{\text {CW}}\rangle ). \end{aligned}$$
(17)

Finally, we send the concatenated feature space to a fully connected neural network to map it to predicted performance \(\hat{Y}\).

We also experiment a latent space-based approach using a variational auto-encoder (VAE) and then finally pass it through a machine learning regressor. To train a VAE architecture on features \(\langle X_{\text {T1}},X_{\text {T2}},X_{\text {CW}}\rangle\) to get feature vectors in the form of mean and variance as

$$\begin{aligned} F_\mu ,F_\sigma \longleftarrow {\text {VAE}}(\langle X_{\text {T1}},X_{\text {T2}},X_{\text {CW}}\rangle )_{\text {Encoder}}; \end{aligned}$$
(18)

while training the variational auto-encoder assembly, the output mean and standard deviation is used to sample from a Gaussian space as

$$\begin{aligned} F_Z = F_\mu + \epsilon * F_\sigma ,\quad {\text {for}} \ \epsilon \in N(0,1). \end{aligned}$$
(19)

In the above equation, \(F_Z\) is the sampled feature space and epsilon is sampled from a normal distribution with mean 0 and variance 1. This feature space \(F_Z\) is passed through the decoder to reconstruct the original input features.

Finally, we use feature-set \(F_\mu\) to train a machine learning regressors, namely: Multi-layer Perceptron (MLP), Linear Regression (LR), Extra Tree Regressor (ET), Random Forest Regressor (RF), XGBoost Regressor (XGB), and k-Nearest Neighbour Regressor (kNN).

We perform similar experiments on dataset \({\text {D}}_2\) using two different feature spaces including: \(\langle X_{\text {MTE}},X_{\text {CW}}\rangle\) and \(\langle X_{\text {ETE}},X_{\text {CW}}\rangle\). While using LSTM and GRU, we forward pass for two time stamps generating feature space \(Z_1\) and \(Z_2\) only. We also evaluate this dataset using variational auto-encoder and machine learning regressor assembly.

The final approach to calculate predicted performance \(\hat{{Y}}\) can be explained in a step by step manner as follows:

  • First, we utilize the dataset \(\langle X_{\text {T1}},X_{\text {T2}},X_{\text {CW}}\rangle\), \(\langle X_{\text {MTE}},X_{\text {CW}}\rangle\), or \(\langle X_{\text {ETE}},X_{\text {CW}}\rangle\) to train a VAE.

  • Then, the dataset is passed through trained encoder \(VAE(.)_{Encoder}\) to get features in term of \(F_\mu\) and \(F_\sigma\).

  • Feature-set \(F_\mu\) is selected as a set of variationally conditioned features.

  • \(F_\mu\) is then passed through the machine learning regressor to get output predicted performance \({\hat{Y}}\).

Evaluation Criteria

The models were evaluated using various evaluation criteria, which include R2-score, mean absolute error (MAE), mean squared error (MSE), and root-mean-squared error (RMSE). Calculation of all these metrics ensures a better evaluation of our regression models. The calculation involves true performance score Y, predicted performance score \({\hat{Y}}\), and the mean score \(\bar{Y}\)

$$\begin{aligned} {R}^{2} \ {\text {score}}= & {} 1-\frac{\varSigma _k(Y_k-{\hat{Y}}_k)^2}{\varSigma _k(Y_k-{\bar{Y}})^2}. \end{aligned}$$
(20)
$$\begin{aligned} {\text {MAE}}= & {} \frac{1}{\varSigma K}\varSigma _k\Vert Y_k-{\hat{Y}}_k\Vert . \end{aligned}$$
(21)
$$\begin{aligned} {\text {MSE}}= & {} \frac{1}{\varSigma K}\varSigma _k( Y_k-{\hat{Y}}_k)^2. \end{aligned}$$
(22)
$$\begin{aligned} {\text {RMSE}}= & {} \sqrt{{\text {MSE}}}. \end{aligned}$$
(23)

R2-score helps in deduction of the variation in the points along the regression line. These evaluation criteria metrics give an estimate of the effective predictions above the mean of the predicted label. MAE, MSE, and RMSE help us to infer the error in predicting students’ performance.

Fig. 3
figure 3

Distribution of data points for \({\text {D}}_1\) and \({\text {D}}_2\) datasets a test 1 (\(X_{\text {T1}}\)), b test 2 (\(X_{\text {T2}}\)), c class assessment (\(X_{\text {CW}}\)), d mid-term evaluation marks (\(X_{\text {MTE}}\)), e end-term evaluation marks (\(X_{\text {ETE}}\)), and f total performance of student (Y)

Fig. 4
figure 4

Correlation matrix for a dataset \({\text {D}}_1\); b dataset \({\text {D}}_2\)

Results and Discussion

The datasets mentioned previously, in dataset description section, can be analyzed based on two methods: one, using exploratory analysis, and other using methods discussed in the section—Approach (Fig. 2).

Exploratory Data Analysis

Exploratory analysis of the datasets can be done using gradient maps, correlation matrix, and distribution of data points. Figure 3 shows the distribution of various features of the datasets. It is clearly visible that \(X_{\text {MTE}}\) and \(X_{\text {ETE}}\) have a similar distribution to the actual students’ scores. However, the distribution of \(X_{\text {T1}}\), \(X_{\text {T2}}\), and \(X_{\text {CW}}\) are quite different from the distribution of overall performance. Visual analysis of these plots suggests some correlation between the features and final prediction score.

Moreover, the correlation matrix, in Fig. 4 shows similar results with high correlation values of 0.88 and 0.96 for \(X_{\text {MTE}}\) and \(X_{\text {ETE}}\), respectively. Unlike that, \(X_{\text {T1}}\) and \(X_{\text {T2}}\) have a lower correlation score of 0.69 and 0.64 with respect to the total performance score of students’. The correlations can further be verified with the gradient maps, as shown in Fig. 5. It clearly demonstrates that \(X_{\text {MTE}}\) vs. \(X_{\text {ETE}}\) has an evident trend of possible marks in comparison to \(X_{\text {T1}}\) vs. \(X_{\text {T2}}\).

Fig. 5
figure 5

Gradient maps of students’ performance for a \(X_{\text {T1}}\) vs. \(X_{\text {T2}}\); b \(X_{\text {ETE}}\) vs. \(X_{\text {MTE}}\)

Table 2 Results for dataset \({\text {D}}_1\) using features—\(X_{\text {T1}}\), \(X_{\text {T2}}\), and \(X_{\text {CW}}\); evaluation performed using—R2-score, MAE, MSE, RMSE

Prediction Results

The results of predictive modeling using various approaches described under the section—Approach—are given in Tables 23, and 4. For dataset \({\text {D}}_1\), the results using VAE with machine learning classifiers along with GRU and LSTM are reported. It is evident that minimum error was acquired by VAE in conjunction with Extra Tree Regressor classifier with R2-score, MAE, MSE, and RMSE of 0.720, 5.943, 77.709, and 8.781, respectively. The results show sufficient utility of \(X_{\text {T1}}\) and \(X_{\text {T2}}\) for students’ performance prediction. Also, almost similar R2-score is obtained for VAE in conjunction with Random Forest Regressor with R2-score, MAE, MSE, and RMSE of 0.720, 6.264, 78.595, and 8.815, respectively.

For dataset \({\text {D}}_2\), the results using two different experiments are shown. In Table 3, using \(X_{\text {MTE}}\) and \(X_{\text {CW}}\), it is evident that minimum error was acquired by VAE with Multi-layer Perceptron Regressor with R2-score, MAE, MSE, and RMSE values of 0.867, 5.293 43.595, and 6.590, respectively. Also, almost similar R2-score is obtained for VAE in conjunction with Linear Regressor with R2-score, MAE, MSE, and RMSE of0.866, 5.3254, 44.142, and 6.633, respectively. TThe results establish the utility of \(X_{\text {MTE}}\) and \(X_{\text {CW}}\) for predicting students’ performance in case of a deferred evaluation of an ongoing academic session.

Table 3 Results for dataset \({\text {D}}_2\) using features—\(X_{\text {MTE}}\) and \(X_{\text {CW}}\); evaluation performed using—R2-score, MAE, MSE, and RMSE
Table 4 Results for dataset \({\text {D}}_2\) using features—\(X_{\text {ETE}}\) and \(X_{\text {CW}}\); evaluation performed using—R2-score, MAE, MSE, and RMSE

In Table 4, for \(X_{\text {ETE}}\) and \(X_{\text {CW}}\), it is evident that minimum error was acquired using VAE in conjunction with Linear Regressor with with a R2-score, MAE, MSE, and RMSE of 0.947, 3.385, 17.279, and 4.154, respectively. Also, almost similar R2-score is obtained for VAE in conjunction with multi-layer perceptron with R2-score, MAE, MSE, and RMSE of 0.947, 3.385, 17.279, and 4.154, respectively. The results show that due to substantial weightage and maximum efforts given by the students, \(X_{\text {ETE}}\) has the highest impact on the evaluation of a student’s collective performance.

Analysis and Discussion

The above results show that the final performance of the student is impacted most by the \(X_{\text {ETE}}\), followed by \(X_{\text {MTE}}\), and \(X_{\text {T1}}\) and \(X_{\text {T2}}\). This is evident from the maximum weightage of \(X_{\text {ETE}}\) in the final evaluation. This is also due to maximum concern of students on end-term evaluation due to maximum weightage. Similarly, in case of a deferred evaluation of an academic performance, we can suggest from the results that using \(X_{\text {MTE}}\) along with \(X_{\text {CW}}\) is a better option for evaluation than \(X_{\text {T1}}\), \(X_{\text {T2}}\) and \(X_{\text {CW}}\).

Also, of all the models tested by us, the best results were obtained by VAE-based models methods, as shown in Table 5. This shows superiority of latent space models above sequential models for performance estimation. Our models attained fairly large R2-score, in comparison to other previous works, which shows how good deep learning techniques are for evaluation and prediction purposes. It should be noted that there is significant scope of increment in the size of our dataset, and it should only add to further improvement of the proposed deep learning approaches.

Table 5 Comparison of the proposed approach with other related works

Conclusion

This study features the role of CI in alleviation of challenges and impact of COVID pandemic on education. Application of deep learning methods for academic performance estimation is shown. State of the current arts is explained with conclusive-related work. For the purpose of evaluation and benchmarking, an anonymized students’ academic performance dataset, called IITR-APE, was created and will be released in the public domain. The promising performance of the proposed approach explains the suitability of modern-day CI methods for modeling students’ academic patterns. However, we feel that availability of larger datasets would further allow the system to be more accurate. For this, we are building a larger pubic version of the current dataset. It was observed through the performance stats and gradient maps that better prediction happens when all components of final grade have equal weightage. It is thus suggested to have a continual evaluation strategy, i.e., there should be many equally weighted tests or assignments, conducted regularly with short frequency, rather than conducting two or three high weightage exams.

This work is an attempt in the direction of enabling CI-based methods for handling adverse effects of COVID like pandemics on education. It is expected that this study may pave the path for future research in the direction of CI enabled predictive assessment of students’ marks.