A study on the efficiency and accuracy of neural network model to optimize personalized recommendation of teaching content
Publié en ligne: 21 mars 2025
Reçu: 08 nov. 2024
Accepté: 16 févr. 2025
DOI: https://doi.org/10.2478/amns-2025-0547
Mots clés
© 2025 Weihang Zhang et al., published by Sciendo
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Today’s education needs to be tailored to the needs of different learners with different teaching content and means, which requires personalized teaching. The demand for personalized education is reflected in two aspects [1-3]. On the one hand, society has a personalized demand for education. With the development of science and technology and the intensification of competition for talents, the social requirements for talents have undergone profound changes. Education has also changed rapidly from the traditional manufacturing of talents in one mold to the cultivation of diversified and personalized talents by the society. On the other hand, learners have individualized needs for education. Individual learners have different needs and different learning characteristics in all courses. Educators in traditional classroom settings face a challenge-how to adapt their teaching or educational methods to this diversity [4-7].
The need for personalized education requires personalized education, which is directly manifested in the personalization of teaching and learning. As an emerging form of education - network teaching, which is based on network technology, it is easier and more necessary to reflect the characteristics of personalization in the design. Due to the existence of personality differences in the object of education, individual learning goals, ability, interest, habits, foundation, style, personality and so on there are differences [8-10]. At present, many network teaching system, although its own resources are huge, but the learning process is fixed, learning methods and modes appear to be relatively single. This has caused a growing contradiction between the two. Many network teaching system ignores the learning itself is a personalized process, did not do according to the individual, tailored to the individual, taking into account the individual differences. Therefore, there is an urgent need for personalized recommendation function in online teaching system [11-14].
Personalized recommendation of teaching content can provide personalized service for learners based on their own preferences, information left by historical visits or relevant information of other similar learners [15-17]. Recommendations can be made in the form of recommending pages for students to browse, recommending learning content, recommending knowledge resources that are of interest to students in order to improve their learning efficiency, providing personalized information services, and distributing targeted emails. The use of neural network models to optimize personalized recommendations for teaching content can lead to more personalized web services, whereby decision makers can adjust certain information to cater to different learners [18-21].
In this paper, we constructed a student cognitive diagnostic model based on GCN, and designed a teaching resource recommendation method based on convolutional joint probability matrix factorization (CUPMF) model to optimize the efficiency and accuracy of personalized recommendation of teaching content. First, the graph convolutional neural network is combined with the traditional cognitive diagnosis model to obtain the mastery matrix for student knowledge points. Then, the CUPMF model is used to decompose the probability matrix of student cognitive diagnosis information, teaching resource information and student score performance information on teaching resources, etc., and solve the implicit feature matrix of test questions, implicit feature matrix of students, and knowledge point feature matrix by stochastic gradient descent method, so as to predict the probable performance of students on teaching resources and, according to their cognitive diagnosis results, the student-users are personalized recommendation of teaching resources. Finally, the performance of the model is experimentally verified to compare the efficiency and accuracy of traditional personalized recommendations for teaching content.
In this paper, a cognitive diagnosis model based on graph convolutional neural network (GCN) is constructed to judge the cognitive level of students, and a teaching resource recommendation method based on convolutional joint probabilistic matrix factorization (CUPMF) model is proposed to carry out personalized and intelligent recommendation of teaching resources according to the cognitive level of students, so as to improve the efficiency and accuracy of recommendation.
Suppose there are
In addition, the cognitive diagnostic model requires the input of a knowledge matrix
The task of the student cognitive diagnostic model in this paper is defined as: given an array
In order to solve the problem posed above, this paper establishes a student cognitive diagnostic model incorporating Graph Convolutional Neural Network (GCN) [22]. In this, the important mathematical notations used are as follows:
The main framework of the developed cognitive diagnostic model is shown in Figure 1. Typically, cognitive diagnostic models need to model three elements: the student

Framework of cognitive diagnosis model
In this paper, we use the vector of students’ mastery on each knowledge point
First the student vector
The sigmoid in Eq. (2) represents the nonlinear activation function, and
In this paper, the updating method of graph convolutional neural network is combined in calculating the feature vectors of the test questions, which can be used to obtain a new representation of the test question vectors by aggregating the neighbor information around each test question, and the calculation steps are as follows:
The input data of the cognitive diagnostic model is the one-hot representation vector
where each node also forms a
The initial feature matrix
Matrix
In this paper, the number of layers
Then use
The sigmoid function in Eq. (5) represents a nonlinear activation function,
From the above two-step calculation and training, the student vector
where the two vectors represented by the computational symbol ∘ are multiplied element by element, and the dimensions of vectors
In order not to lose generality, this paper assumes that each element of the parameter matrix
The loss function of the cognitive diagnostic model is the sum of the cross-entropy between each student’s predicted score
Unlike traditional models for predicting scores, the student vector
In this paper, we combine the convolutional neural network with the joint probability matrix factorization (PMF) model to propose a teaching resource recommendation method based on the convolutional joint probability matrix factorization (CUPMF) model [23]. Its algorithmic framework is shown in Fig. 2, which mainly contains three parts:
GCN-based student-user cognitive diagnostic modeling to obtain the student-knowledge point mastery matrix. Convolutional neural network module, whose main purpose is to deeply mine teaching resources in different dimensions through convolutional neural networks, while seamlessly integrating them into joint probability matrix decomposition through nonlinear transformations in the output layer. Joint probability matrix decomposition, this part performs probability matrix decomposition by combining multiple information such as students’ cognitive diagnosis information, teaching resources information, and students’ score performance information on teaching resources, and through stochastic gradient descent method, it solves the implicit feature matrix of the test questions, implicit feature matrix of the students, and the feature matrix of the knowledge points containing the parameters of the CNN, and then predicts students’ possible performances on the possible performance on teaching resources and recommend teaching resources with difficulty suitable for current student users based on their cognitive diagnosis.

Schematic diagram of CUPMF recommendation algorithm
In this paper, we choose to use convolutional neural network (CNN) to mine the teaching resource data, the CNN framework is mainly responsible for mining the potential features of the test question teaching resources, generating the implicit feature vector of the test question, constructing the implicit feature matrix representation of the test question with the CNN weight parameter, which is used in the joint probability matrix decomposition model for training and solving [24]. The convolutional network framework of the CUPMF model is shown in Fig. 3, and its consists of the following four layers:

Convolutional network diagram of CUPMF model
The word embedding layer converts the original test question information into a dense numeric matrix as input to the next convolutional layer. The test question information mainly contains three parts: the question stem, the answer and the paraphrase, and these three parts are processed through the word splitting technique, and each word is randomly initialized or converted into a word vector by the pre-trained word embedding model, and finally the test question is represented as a dense numeric matrix by connecting the word vectors in the test question information
Where
The convolutional layer is mainly used to extract the feature information of a test question. A trial context feature
where “⊗” denotes the convolution operation,
Considering the limited test question feature information captured by using only a single shared weight, this paper employs multiple sets of shared weights in the convolutional layer to obtain multiple sets of feature vectors describing test question feature information.
After the convolution operation, the test information is represented as a feature matrix with nc level dimensions, and the dimensions of the test feature vectors in the matrix are not uniform, i.e., the number of columns of the matrix is not uniform, which not only results in the vector dimensions being too high for the computational performance, but also makes it difficult to construct the subsequent layers. Therefore, this model extracts representative features from each trial feature vector by pooling layers and reduces the representation of the trial document to
The output layer is mainly responsible for making a nonlinear mapping of the output of the previous layers. Therefore, it is necessary to map
where
where
The joint probability matrix decomposition model of the CUPMF model is shown in Fig. 4, whose main idea is to decompose the student-test score information matrix

Joint probability matrix decomposition frame of CUPMF model
The prior probability of the initialization matrix
The prior probability of matrix
The initialization of matrix The weights Word vector Gaussian noise
Thus, the trial question
The probability distribution of matrix
From the implicit vector
where
Similarly, from the implicit eigenvector
where
Similarly, from the implicit eigenvector
where
Combined with the above equation for the prior probability distribution, the posterior probability distribution for matrix
The above equation can be obtained by taking logarithms on both sides of Equation (23):
where
where the value of parameter
The experiment was completed under the Python 3.11.7 environment, using the deep learning open source framework Pytorch, built on the Ubuntu 24.10 LTS operating system, remotely assisted by PyCharm software, and using an RTX4090D graphics card for high-performance computing.
In this experiment, two representative public datasets in the field of smart education were used: the ASSISTMents2015 dataset and the EdNet dataset. The ASSISTMents 2015 dataset contains the answers of some students from 2015 to 2016 collected by the ASSISTMents platform, and contains rich information such as students, exercises, and knowledge points. While EdNet is a massive dataset that collects hundreds of millions of learning behavior records from nearly a million students from multiple platforms, in this experiment, the focus is only on information about the students, the exercises, the knowledge points, and whether or not the students answered the questions correctly as a result.
In this experiment, the dataset was preprocessed as follows: first, exercises without knowledge point annotations and students with less than 20 answer records were filtered out. In addition, considering that the EdNet dataset is too large, 1200 relatively active students were randomly selected for the diagnostic study during the experiment. Meanwhile, the experiment was divided to get the training set, validation set and test set by taking the ratio of 7:1:2. After data preprocessing, the statistics of the two data sets are shown in Table 1.
The statistics of datasets
Statistical term | ASSISTMents2015 | EdNet |
---|---|---|
Number of students | 4286 | 1200 |
Number of exercises | 18026 | 13624 |
Knowledge points | 131 | 195 |
Number of answer records | 289653 | 1125341 |
Number of correctly answered exercises | 201328 | 836725 |
Number of incorrect answers to exercises | 88325 | 228616 |
Comparison of models
In this experiment, traditional diagnostic models, such as IRT, MIRT and DINA, and the latest neurocognitive diagnostic models, such as NeuralCD and RCD, are selected for comparison in order to validate the effectiveness of the cognitive diagnostic model proposed in this paper, which is based on graphical convolutional neural networks. The comparison models used are described as follows:
IRT: It is a continuous nonlinear model, which is mainly used to model the relationship between one-dimensional students’ abilities and the characteristics of the exercises (e.g., difficulty of the exercises, differentiation of the exercises, etc.). MIRT: It is an extension of the traditional unidimensional item response theory model that takes into account students’ fine-grained abilities and models student abilities and exercise characteristics from the perspective of multidimensional knowledge points. DINA: is a discrete diagnostic model that views student cognitive states as multidimensional and independent, with students either mastering or not mastering in relation to knowledge points. It also focuses on the student’s guessing factor and sliding factor. NeuralCD: Neural networks are introduced to model the complex interactions between students and exercises, and the monotonicity assumption in traditional diagnostic models is used to ensure the interpretability of student factors and exercise factors. The model uses multidimensional successive vectors to model students and exercises, where each dimension of the vectors represents a different knowledge point. For the student factor vector, each dimension corresponds to the degree to which the student has mastered the knowledge point. For the exercise factor vector, each dimension corresponds to the degree of relevance of the exercise to the knowledge point. RCD: Modeling student-exercise-knowledge point relationships based on a hierarchical graph structure, especially the dependency graph of knowledge concepts.
Evaluation Indicators
Considering that students’ ability level cannot be directly measured in cognitive diagnosis, this experiment used some common indicators to evaluate the performance of the model, such as root mean square error (RMSE), accuracy (ACC) and area under the ROC curve (AUC). From a regression perspective, RMSE is used to measure the difference between the predicted value and the actual score (0 or 1), with smaller values being better. From a categorization perspective, predicting whether a student will answer an exercise correctly can be viewed as a binary categorization problem. The predicted value is set to 1 if the predicted value is greater than or equal to 0.5, and 0 if it is not, and is measured using the ACC and AUC, with values as close to 1 as possible.
The experimental results of the GCN-based cognitive diagnostic model proposed in this paper on the ASSISTMents2015 and EdNet datasets as well as their comparison experiments are shown in Table 2.
Overall results on student performance prediction
Model | ASSISTMents2015 | EdNet | ||||
---|---|---|---|---|---|---|
ACC↑ | RMSE↓ | AUC↑ | ACC↑ | RMSE↓ | AUC↑ | |
IRT | 0.6555 | 0.5361 | 0.6092 | 0.6921 | 0.4654 | 0.7118 |
DINA | 0.6659 | 0.5337 | 0.6788 | 0.6894 | 0.4871 | 0.6826 |
MIRT | 0.7033 | 0.4695 | 0.7251 | 0.7047 | 0.4476 | 0.7273 |
NeuralCD | 0.7406 | 0.4481 | 0.7605 | 0.7166 | 0.4304 | 0.7784 |
RCD | 0.7421 | 0.4349 | 0.7938 | 0.7299 | 0.4239 | 0.7865 |
This article | 0.7826 | 0.4271 | 0.8072 | 0.7497 | 0.4043 | 0.7907 |
From the data in Table 2, it can be seen that the ACC, RMSE, and AUC of this paper’s model on the ASSISTMents2015 dataset are 0.7826, 0.4271, and 0.8072, respectively, whereas on the EdNet dataset, it achieves 0.7497, 0.4043, and 0.7907, respectively. Compared with the other comparative models, especially the state-of-the-art hierarchical graph-based RCD model, this paper’s GCN-based cognitive diagnostic model achieves the best performance on all evaluation metrics, i.e., ACC and AUC achieve the maximum value and RMSE achieves the minimum value. This phenomenon indicates the effectiveness of the graph convolutional neural network model in combination with the traditional cognitive diagnostic model. In addition, the results from the ASSISTMents2015 dataset show that the model in this paper has a more stable performance when the dataset is sparse, which suggests that the interaction function module between the student vectors and the test question vectors plays an important role in this model. In addition, observing all the comparative experiments found that the neural network-based approach has better performance than the traditional cognitive diagnostic model, which indicates that the neural network-based approach can more effectively tap into the complex relationships among students, exercises and knowledge points.
In this section of the experiment, the student score matrix

Student cognitive portrait
From Fig. 5, it can be seen that the students’ knowledge point mastery varies, for example, the cognitive vector
The student score matrix

Student learning status portrait
In this section, five recommendation methods are selected for comparison experiments, including the random recommendation selection strategy Random, the selection strategy DT based on multiple decision trees, the selection strategy IRT based on Item Response Theory and the teaching resource recommendation method based on the CUPMF model described in this paper.
Experimental environment: the experimental platform Windows11, program implementation using Python3.11.7. The data required to conduct the experiment is shown in Table 3.
Experimental data
Network node | 5 knowledge points in each chapter |
---|---|
Data | 15623 student history answer data |
KU relationship | |
Recommended number of test questions n | |
Number of question banks |
EXPERIMENTAL DESIGN: A group of target students in each of the three categories of A, B, and C were selected for test question recommendation and evaluation observation. The experimental test was conducted at the time when the students’ historical answer records reached 30 questions.
Taking the students of category A as an example, the recommendation method of this paper was applied to recommend 10 test questions for 10 students of category A respectively, totaling 100 questions, and 20 recommended test questions were obtained by removing the duplicated test questions among them. At the same time, without communicating with each other, two experienced classroom teachers were asked to recommend two groups of 10 questions, respectively, to get 19 effective test questions, and the teacher recommendation results were used to evaluate the accuracy of the algorithm recommendation. Similarly, three groups of students, A, B, and C, were tested for recommendation and response respectively, and test data were obtained. The same method is used to obtain the recommendation test results of the remaining three comparative recommendation methods (Random, DT, and IRT), and the test result data obtained from the three types of students and the five recommendation methods are analyzed to compare the performance in terms of the common indicators of the model, the recommendation accuracy, and the reasonableness of the recommendation, respectively.
In order to assess the effectiveness and accuracy of the algorithm, a variety of metrics were used to comprehensively evaluate the performance of the algorithm.
Precision, Recall and F1
Precision, Recall and F1 metrics are used to evaluate the performance of the CUPMF model and other comparative methods in terms of recommendation algorithm modeling. Precision, Recall and F1 are specifically defined as follows:
Where
MAE
MAE (Mean Absolute Error) measures the average of the absolute error between the predicted and actual values, and is used here to reflect the difference between the positive response rate of the test question recommendation results and the statistical expectation, i.e:
where
Standard deviation of ideal response rate
Unlike the “optimal” tendency of test recommendation questions and common product recommendations, a higher positive response rate is not better for students doing the exercises. Different types of students have different positive response rates within a certain range to achieve optimal learning results.
In this paper, students are categorized into three types of students, including Class A students with good foundation, Class B students with moderate foundation, and Class C students with poor foundation. The optimal range of positive answer rate varies among the three categories of students. The expected positive response rate for recommending test questions to different students depends on a number of factors, such as the difficulty of the test questions and the knowledge level of the students. Based on the analysis of the historical data of the students in the category and the answer records of the students, the historical estimated correct response rate is found, and close to this range, it is considered to have found a suitable range for the students, in which the students will learn better.
As a result, the correct response rate used in this paper, defined as CRR (Correct response rate), i.e:
Where
Based on the statistically obtained historical positive answer rate of the three types of students, in this paper, the experiment is selected as the average positive answer rate of the Nth time of the historical answer record of the same type of students, which is regarded as the ideal positive answer rate of this type of students in the Nth time of the answer. In the experiment, the 31st answer record was selected, and the ideal positive answer rate for type A students was 0.5287, for type B students was 0.4530, and for type C students was 0.4089. The standard deviation of the experiment was calculated based on the ideal positive answer rate for the corresponding type of students.
Complete the experiment, calculate the evaluation indexes, and get the comparison results of the index data of different algorithm models as shown in Table 4~Table 6. Table 4~Table 6 represent the indicator data of the test results of A, B and C students respectively.
Comparison of indicator data for test results of class A student
Algorithm/model | Precision | Recall | F1 | MAE | KU | |
---|---|---|---|---|---|---|
Random | 0.5562 | 0.7544 | 0.6403 | 0.0525 | 0.0235 | 1,2,4 |
DT | 0.6164 | 0.7613 | 0.6812 | 0.0387 | 0.0016 | 2,3,4 |
IRT | 0.6599 | 0.5836 | 0.6194 | 0.0656 | 0.0421 | 1,2,4,5 |
PMF | 0.8599 | 0.8774 | 0.8686 | 0.0308 | 0.0045 | 1,2,4 |
CUPMF | 0.8953 | 0.9105 | 0.9028 | 0.0147 | 0.0018 | 1,2,4 |
Comparison of indicator data for test results of class B student
Algorithm/model | Precision | Recall | F1 | MAE | KU | |
---|---|---|---|---|---|---|
Random | 0.6648 | 0.7534 | 0.7063 | 0.0289 | 0.0182 | 1,2,4,3 |
DT | 0.7081 | 0.7521 | 0.7294 | 0.0226 | 0.0129 | 1,2,4 |
IRT | 0.7541 | 0.8453 | 0.7971 | 0.0188 | 0.0141 | 1,2,4 |
PMF | 0.8028 | 0.9052 | 0.8509 | 0.0111 | 0.0023 | 1,2,4 |
CUPMF | 0.8517 | 0.9121 | 0.8809 | 0.0136 | 0.0019 | 1,2,4 |
Comparison of indicator data for test results of class C student
Algorithm/model | Precision | Recall | F1 | MAE | KU | |
---|---|---|---|---|---|---|
Random | 0.6062 | 0.7651 | 0.6764 | 0.0285 | 0.0051 | 1,2,4,5 |
DT | 0.7036 | 0.8247 | 0.7594 | 0.0275 | 0.0118 | 2,3,4 |
IRT | 0.6796 | 0.8247 | 0.7452 | 0.0223 | 0.0138 | 1,2,4,3 |
PMF | 0.7637 | 0.8839 | 0.8194 | 0.0182 | 0.0032 | 1,2,4 |
CUPMF | 0.8046 | 0.8921 | 0.8461 | 0.0141 | 0.0029 | 1,2,4 |
According to the experimental results, the data comparison table for each category is comprehensively observed. From the table, it can be seen that comparing the decision tree algorithm, the algorithm designed in this paper has a good performance in precision rate, recall rate and F1 indexes, and at the same time, after optimizing the sub-module and improving the performance of prediction network, the overall ability of the model has a small increase, which can reflect that the model in this paper has a higher performance than the other models.
In terms of MAE index, this paper’s model has the best performance, the mean value of error are lower, the MAE value of the test results of the three categories of students in A, B and C are 0.0147, 0.0136 and 0.0141, respectively. In terms of the standard deviation of the CRR, this paper’s model has the best recommendation, its
In this paper, we optimize the efficiency and accuracy of personalized recommendation of teaching content by constructing a student cognitive diagnosis model based on GCN and a teaching resource recommendation model based on convolutional joint probability matrix factorization (CUPMF).
First, the cognitive diagnostic model in this paper achieved the best performance against other models with ACC, RMSE and AUC of 0.7826, 0.4271 and 0.8072 on the ASSISTMents2015 dataset and 0.7497, 0.4043 and 0.7907 on the EdNet dataset, respectively. This indicates that the graph convolutional neural network model can be effectively combined with the traditional cognitive diagnosis model, which can more effectively tap into the complex relationship between students, exercises, and knowledge points.
Secondly, the cognitive portrait of student users and the learning state portrait are constructed based on the cognitive diagnosis results of students, so as to understand the cognitive state of students based on the cognitive portrait of student users, and to understand the learning state of students based on the learner state portrait, as well as to pay targeted attention to the students with high student warning coefficients.
Finally, the personalized recommendation algorithm for teaching content designed in this paper has good performance in terms of precision rate, recall rate and F1 indexes, and from the point of view of the two indexes of MAE and
Higher Education Teaching Reform Research and Practice Project of Hebei Province: Construction and Practice of Innovation and Entrepreneurship Education Curriculum System Based on “Whole Field Double”.
Higher Education Teaching Reform Research and Practice Project of Hebei Province: Construction and practice of the curriculum system of "Labor Education" in local normal universities under the background of application transformation development -- based on the perspective of "vocational maturity" (2023GJJG371).
Research and Practice Project on Teaching Reform of Innovation and Entrepreneurship Education of Hebei Provincial Department of Education: Construction of Innovation and Entrepreneurship Education Curriculum System Based on the Linkage of Three Courses: A Case Study of Primary Education (2023CXCy175).