Uneingeschränkter Zugang

A study on the efficiency and accuracy of neural network model to optimize personalized recommendation of teaching content

 und   
21. März 2025

Zitieren
COVER HERUNTERLADEN

Introduction

Today’s education needs to be tailored to the needs of different learners with different teaching content and means, which requires personalized teaching. The demand for personalized education is reflected in two aspects [1-3]. On the one hand, society has a personalized demand for education. With the development of science and technology and the intensification of competition for talents, the social requirements for talents have undergone profound changes. Education has also changed rapidly from the traditional manufacturing of talents in one mold to the cultivation of diversified and personalized talents by the society. On the other hand, learners have individualized needs for education. Individual learners have different needs and different learning characteristics in all courses. Educators in traditional classroom settings face a challenge-how to adapt their teaching or educational methods to this diversity [4-7].

The need for personalized education requires personalized education, which is directly manifested in the personalization of teaching and learning. As an emerging form of education - network teaching, which is based on network technology, it is easier and more necessary to reflect the characteristics of personalization in the design. Due to the existence of personality differences in the object of education, individual learning goals, ability, interest, habits, foundation, style, personality and so on there are differences [8-10]. At present, many network teaching system, although its own resources are huge, but the learning process is fixed, learning methods and modes appear to be relatively single. This has caused a growing contradiction between the two. Many network teaching system ignores the learning itself is a personalized process, did not do according to the individual, tailored to the individual, taking into account the individual differences. Therefore, there is an urgent need for personalized recommendation function in online teaching system [11-14].

Personalized recommendation of teaching content can provide personalized service for learners based on their own preferences, information left by historical visits or relevant information of other similar learners [15-17]. Recommendations can be made in the form of recommending pages for students to browse, recommending learning content, recommending knowledge resources that are of interest to students in order to improve their learning efficiency, providing personalized information services, and distributing targeted emails. The use of neural network models to optimize personalized recommendations for teaching content can lead to more personalized web services, whereby decision makers can adjust certain information to cater to different learners [18-21].

In this paper, we constructed a student cognitive diagnostic model based on GCN, and designed a teaching resource recommendation method based on convolutional joint probability matrix factorization (CUPMF) model to optimize the efficiency and accuracy of personalized recommendation of teaching content. First, the graph convolutional neural network is combined with the traditional cognitive diagnosis model to obtain the mastery matrix for student knowledge points. Then, the CUPMF model is used to decompose the probability matrix of student cognitive diagnosis information, teaching resource information and student score performance information on teaching resources, etc., and solve the implicit feature matrix of test questions, implicit feature matrix of students, and knowledge point feature matrix by stochastic gradient descent method, so as to predict the probable performance of students on teaching resources and, according to their cognitive diagnosis results, the student-users are personalized recommendation of teaching resources. Finally, the performance of the model is experimentally verified to compare the efficiency and accuracy of traditional personalized recommendations for teaching content.

Cognitive diagnosis and neural network-based model for teaching content recommendation

In this paper, a cognitive diagnosis model based on graph convolutional neural network (GCN) is constructed to judge the cognitive level of students, and a teaching resource recommendation method based on convolutional joint probabilistic matrix factorization (CUPMF) model is proposed to carry out personalized and intelligent recommendation of teaching resources according to the cognitive level of students, so as to improve the efficiency and accuracy of recommendation.

GCN-based diagnostic model of student cognition
Definition of cognitive diagnostic problems

Suppose there are M student, N test questions and P knowledge points. Then all the students can be denoted as S = S1,S2,⋯,SM, the test questions can be denoted as E = E1,E2,⋯,EN, and the knowledge points as K = K1,K2,⋯,KP In the usual process of doing the questions, each student will choose a part of the test questions to practice, and the obtained answer records are recorded as R, R is a three-dimensional array R = (Si,Ej,tij), where SiS, EjE, tij is the score of the students’ Si answers to the test questions Ej, and the score is tij = 1 when the answers are correct, and tij = 0 when the answers are incorrect.

In addition, the cognitive diagnostic model requires the input of a knowledge matrix Q = Qj,lN×P which is usually labeled by an expert or a teacher, and the elements Qj,l of the matrix are defined as follows: Qj,l={ 1,QuestionEjexamines this knowledge pointKl0,QuestionEjexamines this knowledge pointKl

The task of the student cognitive diagnostic model in this paper is defined as: given an array R of each student’s answer records and a matrix Q of knowledge points, the potential mastery usi of each student on each of the known knowledge points Kl and the corresponding characteristics of the multiple test questions themselves (difficulty vdiffj of the questions, differentiation vdiskj , etc.) is obtained by predicting each student’s Si scores yi,j on the test questions Ej.

Diagnostic Modeling of Student Cognition

In order to solve the problem posed above, this paper establishes a student cognitive diagnostic model incorporating Graph Convolutional Neural Network (GCN) [22]. In this, the important mathematical notations used are as follows:

xsi --one-hot vectors of students.

xej --one-hot vector of test questions.

Qj,l --Knowledge Matrix.

usi --student mastery vector.

V(r) --the characterization matrix of the test questions.

A --the adjacency matrix of the test nodes.

vej --The updated vector of test questions using graph convolutional neural network.

Qe -- the knowledge point related vector of the test question.

vdiffj --Difficulty of the test questions.

vdiskj --The differentiation of the test questions.

yij --Final predicted score.

tij --True Score.

The main framework of the developed cognitive diagnostic model is shown in Figure 1. Typically, cognitive diagnostic models need to model three elements: the student Si, the test questions Ej and the interaction function between them. The inputs to the cognitive diagnostic model are the one-hot vector representation of the student xsi , the one-hot vector representation of the test questions xej and the knowledge matrix Q = Qj,lN×P, and then the potential feature vectors of the student and the test questions are obtained through the operation with the parameter matrix and the iterative computation of the graphical convolutional neural network: usi and vej , after which a three-layer fully connected neural network is utilized to predict the interaction function between them, and finally the predicted scores of the students are obtained yi,j, and the diagnosis of each student is the student potential feature vector usi .

Figure 1.

Framework of cognitive diagnosis model

Student Vector

In this paper, we use the vector of students’ mastery on each knowledge point usi to represent the students Si, each item in the vector usi is consecutive and belongs to the range of [0, 1], and the higher its value represents the better mastery of this student on that knowledge. The length of usi is the number P of knowledge points K and the total number of vectors is the number M of students.

First the student vector usi is computed, which is obtained by multiplying the one-hot representation vector xsi of students with the trainable parameter matrix B: usi=sigmoid(xsi×B)

The sigmoid in Eq. (2) represents the nonlinear activation function, and usi(0,1)1×P is the student vector, where the parentheses (0,1) denote continuous quantities whose values are all between 0 and 1.

xsi(0,1)1×M is the one-hot representation vector of the students Si input at the beginning, where the parentheses (0,1) indicate that the vector xsi takes only two values: 0 and 1. M represents the total number of students, and B ∈ℝM×P is the parameter matrix for the computation process, which can be determined by training.

Test Question Vector

In this paper, the updating method of graph convolutional neural network is combined in calculating the feature vectors of the test questions, which can be used to obtain a new representation of the test question vectors by aggregating the neighbor information around each test question, and the calculation steps are as follows:

The input data of the cognitive diagnostic model is the one-hot representation vector xej of each test question Ej, whose dimension and number are the number of test questions N. Each test question is treated as a node in a graph convolutional neural network, and the initial feature vector of each node is the one-hot vector xej of the test questions, and all the vectors xej are formed into a N×N -dimensional matrix V(0) which is said to be the initial feature matrix of all the test questions, and each of its rows is the feature vector of a test question, i.e. Vj(0)=xej

where each node also forms a N×N -dimensional adjacency matrix A between them, and the elements of the matrix A have only 0 and 1 values, which are defined as follows: Aj,j={ 1,There is a certain connection between testEjand testEj'0,There is no relationship between questionEjand questionEj'

The initial feature matrix V(0) and the adjacency matrix Aj,j of the test problem are the two input data of the graph convolutional neural network. If we consider it as a common neural network structure, the update iteration formula between layers is as follows: V(r+1)=sigmoid(D12AD12V(r)W(r))

Matrix à = A + I in Eq. (4), where A is the adjacency matrix and I is the unit matrix. D is the degree matrix of matrix Ã, which is a diagonal matrix with diagonal elements being the sum of the elements of each row of matrix Ã. Matrix V(r) is the feature matrix corresponding to each test question at each network layer, the input layer at the beginning is the initial feature matrix of the test question V(0), and sigmoid denotes the nonlinear activation function. W(r) is the matrix of weight parameters to be trained to obtain in each layer of the graph convolutional neural network.

In this paper, the number of layers r of the graph convolutional neural network is set to 2, so it is necessary to go through three iterations of Eq. (4) to update, and finally we can get the updated feature matrix of the test questions V(3), and then each row of V(3) is split, and finally we can get the feature vector of each test question after the update of the graph convolutional neural network vej .

Then use vej to calculate the different potential test feature vectors corresponding to each test question, which are classified into three categories in this paper: the knowledge point related vector of the test question Qe, the difficulty vector of the test question vdiffj and the differentiation vector of the test question vdiskj , respectively, and the updating formula is as follows: Qej=sigmoid(vej×Q)vdiffj=sigmoid(vej×C)vdiskj=sigmoid(vej×D)

The sigmoid function in Eq. (5) represents a nonlinear activation function, vej(0,1)1×N is the updated test question vector obtained in the previous step, the knowledge point matrix Q ∈ (0,1)N×P, where P is the number of knowledge points, and the knowledge point correlation vector Qej(0,1)1×P . C ∈ ℝN×P and D ∈ ℝN×1 are parameter matrices, which can be obtained from the model training, multiplying the test question eigenquantities with the parameter matrices to obtain the test question difficulty vector vdiffj(0,1)1×P and the test question differentiation vector vdiskj(0,1) required by the model in the next step.

Interaction function

From the above two-step calculation and training, the student vector usi and the three types of feature vectors of the test questions Qej,vdiffj,vdiskj have been obtained, inspired by the multidimensional item response theory method (MIRT), this paper designs the interaction function of the first layer of the model as: yij0=Qej(usivdiffj)×vdiskj

where the two vectors represented by the computational symbol ∘ are multiplied element by element, and the dimensions of vectors usi and vdiffj are both (0,1)P. The resulting yij(0) is input into two fully connected layers and an output layer, which ultimately yields the predicted scores of the student cognitive diagnostic model for the student on the individual test questions yij, as computed as follows: yij(1)=sigmoid(W1×yij(0)T+b1)yij(2)=sigmoid(W2×yij(1)+b2)yij=sigmoid(W3×yij(2)+b3)

Wi,i = 1,2,3 in Eq. (7) is the parameter matrix to be trained to obtain in each layer, and bi,j = 1,2,3 is the offset of the network.

In order not to lose generality, this paper assumes that each element of the parameter matrix Wi in each layer is positive, easy to obtain in each layer, there is yijusi>0 constant, which ensures the monotonicity of the diagnostic model, i.e., when a student has a low level of mastery of the knowledge of the point corresponds to the prediction of the score is also low.

The loss function of the cognitive diagnostic model is the sum of the cross-entropy between each student’s predicted score yij and the student’s true score tij on each question, calculated as follows: lossfunction=i,j(tijlogyij+(1tij)log(1yij))

Unlike traditional models for predicting scores, the student vector usi , which is trained to minimize the value of the loss function, is a diagnostic of the student’s potential knowledge state, and it represents the potential mastery of student Si in each knowledge point.

Recommendation of teaching resources based on the CUPMF model
Overall framework of the CUPMF model

In this paper, we combine the convolutional neural network with the joint probability matrix factorization (PMF) model to propose a teaching resource recommendation method based on the convolutional joint probability matrix factorization (CUPMF) model [23]. Its algorithmic framework is shown in Fig. 2, which mainly contains three parts:

GCN-based student-user cognitive diagnostic modeling to obtain the student-knowledge point mastery matrix.

Convolutional neural network module, whose main purpose is to deeply mine teaching resources in different dimensions through convolutional neural networks, while seamlessly integrating them into joint probability matrix decomposition through nonlinear transformations in the output layer.

Joint probability matrix decomposition, this part performs probability matrix decomposition by combining multiple information such as students’ cognitive diagnosis information, teaching resources information, and students’ score performance information on teaching resources, and through stochastic gradient descent method, it solves the implicit feature matrix of the test questions, implicit feature matrix of the students, and the feature matrix of the knowledge points containing the parameters of the CNN, and then predicts students’ possible performances on the possible performance on teaching resources and recommend teaching resources with difficulty suitable for current student users based on their cognitive diagnosis.

Figure 2.

Schematic diagram of CUPMF recommendation algorithm

Convolutional Neural Networks

In this paper, we choose to use convolutional neural network (CNN) to mine the teaching resource data, the CNN framework is mainly responsible for mining the potential features of the test question teaching resources, generating the implicit feature vector of the test question, constructing the implicit feature matrix representation of the test question with the CNN weight parameter, which is used in the joint probability matrix decomposition model for training and solving [24]. The convolutional network framework of the CUPMF model is shown in Fig. 3, and its consists of the following four layers:

Figure 3.

Convolutional network diagram of CUPMF model

Word embedding layer

The word embedding layer converts the original test question information into a dense numeric matrix as input to the next convolutional layer. The test question information mainly contains three parts: the question stem, the answer and the paraphrase, and these three parts are processed through the word splitting technique, and each word is randomly initialized or converted into a word vector by the pre-trained word embedding model, and finally the test question is represented as a dense numeric matrix by connecting the word vectors in the test question information TjRp×l: Tj=[ wi1wiwi+1 ]

Where p is the dimension of the vector and l denotes the number of word vectors.

Convolutional layer

The convolutional layer is mainly used to extract the feature information of a test question. A trial context feature cijR is extracted by the j th shared weight WcjRp×w , whose window size w indicates the number of surrounding words, i.e: cij=fun(WcjD(:,i:(i+w1))+bcj)

where “⊗” denotes the convolution operation, bcjR is the deviation corresponding to Wcj , and fun is a nonlinear excitation function. In addition, this paper uses ReLU to avoid the gradient vanishing problem. Then, by constructing a contextual feature vector cjRl−w+1 with weights Wcj : cj=[c1j,c2j,,cij,,clw+1j]

Considering the limited test question feature information captured by using only a single shared weight, this paper employs multiple sets of shared weights in the convolutional layer to obtain multiple sets of feature vectors describing test question feature information.

Pooling layer

After the convolution operation, the test information is represented as a feature matrix with nc level dimensions, and the dimensions of the test feature vectors in the matrix are not uniform, i.e., the number of columns of the matrix is not uniform, which not only results in the vector dimensions being too high for the computational performance, but also makes it difficult to construct the subsequent layers. Therefore, this model extracts representative features from each trial feature vector by pooling layers and reduces the representation of the trial document to nc fixed-length feature vector by constructing the subsumption operation of fixed-length feature vectors, i.e.: df=[max(c1),max(c2),,max(cj),,max(cnc)]

Output Layer

The output layer is mainly responsible for making a nonlinear mapping of the output of the previous layers. Therefore, it is necessary to map df on the k -dimensional space of the joint probability matrix decomposition model to accomplish the recommendation task, i.e., to generate the trial potential matrix by using the conventional nonlinear mapping: Dj=tanh(Wf2{tanh(Wf1df+bf1)}+bf2)

where Wf1Rf×nc and Wf2Rk×f are the mapping matrices, and bf1 and bf2 are the deviation vectors of Wf1,Wf1 and DjRk. Finally, through the convolution and nonlinear transformation processing of the hidden layer, the convolution part of the CUPMF model is approximately intertwined into a nonlinear function, which takes the test word vectors as inputs, and the outputs are the implicit feature vectors corresponding to each test question: Dj=Cnn(W,Tj)

where w denotes all weight and bias variables, Tj denotes the word vector of trial j after diving through words, and Dj denotes the implicit feature vector of trial j.

Joint probability matrix decomposition

The joint probability matrix decomposition model of the CUPMF model is shown in Fig. 4, whose main idea is to decompose the student-test score information matrix R ∈ ℝm×n collected by the platform, the student-knowledge point mastery matrix A ∈ ℝm×l constructed by the cognitive diagnostic model, and the test-question-knowledge point relationship matrix Q ∈ ℝn×l recorded manually by the domain experts through the matrix decomposition technique, and obtain the student implicit feature matrix U ∈ ℝm×v, knowledge point implicit feature matrix K ∈ ℝl×v, and test question implicit feature matrix D = CNN(W,TS),D ∈ ℝn×v. Then, the original correlation information matrix is represented by the Bayesian criterion, and the real correlation information matrix data is fitted by continuous training to obtain the relevant parameters of each implicit feature matrix, and finally, the trained implicit feature matrix predicts the student’s performance on the teaching resources, and the recommendation is made in conjunction with the mastery of the student-knowledge points.

Figure 4.

Joint probability matrix decomposition frame of CUPMF model

The prior probability of the initialization matrix U follows a Gaussian distribution with mean 0 and variance σU2 : p(U|σU2)=i=1mG(Ui|0,σU2I)

The prior probability of matrix K obeys a Gaussian distribution with mean 0 and variance σk2 : p(K|σK2)=i=1lG(Ki|0,σK2I)

The initialization of matrix D differs from the traditional joint probability matrix decomposition in that it is determined by three main variables:

The weights W between neurons in the convolutional network, and the weight W probability distribution: p(W|σW2)=kG(wk|0,σW2)

Word vector Tj representing the trial j generated by the word embedding technique.

Gaussian noise εj~N(0,σD2) variable.

Thus, the trial question j undergoes the CNN network to generate the implicit feature vector Dj: Dj=CNN(W,Tj)+εj

The probability distribution of matrix D is obtained by the above equation: p(D|W,T,σD2)=j=1nG(Dj|CNN(W,Tj),σD2I)

From the implicit vector Ui of student i and the implicit vector Dj = CNN(W,Tj) of test j, it can be obtained that the probability rij of student i ‘s score on test j obeys a Gaussian distribution with mean h(UiTCNN(W,Tj)) and variance σR2 and is independent, respectively, and its conditional probability distribution has the mathematical expression: p(R|U,D,σR2)=i=1mj=1n[ G(rij|h(UiTCNN(W,Tj)),σR2I) ]lijR

where IijR is an indicator function, IijR=1 if student i has done test j and IijR=0 otherwise. h(x) is a sigmoid function, i.e., the value of UiTCnn(W,Tj) is mapped to the range (0, 1).

Similarly, from the implicit eigenvector Ui of student i and the implicit eigenvector Kj of knowledge point j, it can be obtained that: Student i ‘s mastery of knowledge point j, ∝ij, satisfies a Gaussian distribution with mean h(UiTKj) and variance σA2 and independent, with a conditional probability distribution: p(A|U,K,σA2)=i=1mj=1l[ G(αij|h(UiTKj),σA2I) ]lijA

where IijA is an indicator function, IijA=1 if student i has mastered knowledge point j and IijA=0 otherwise.

Similarly, from the implicit eigenvector Di of test question i and the implicit eigenvector Kj of knowledge point j, we can get that: the association case qij of test question i and knowledge point j satisfies the Gaussian distribution with mean h(CNN(W,Tj)TKj) and variance σQ2 and is independent, and its conditional probability distribution is: p(Q|D,K,σQ2)=i=1mj=1l[ G(αij|h(CNN(W,Tj)TKj),σQ2I) ]lijQ

where IijQ is the indicator function, IijQ=1 if test i examines knowledge point j, and IijQ=0 otherwise.

Combined with the above equation for the prior probability distribution, the posterior probability distribution for matrix U,D,W,K is obtained from the Bayesian criterion as: (U,D,W,K|R,A,Q,T,σU2,σW2,σD2,σK2,σR2,σA2,σQ2)=p(Q|D,K,σQ2)p(K|σK2)p(D|W,T,σD2)p(W|σW2)p(R|U,D,σR2)p(U|σU2)p(D|W,T,σD2)p(W|σW2)p(A|U,K,σA2)p(U|σU2)p(K|σK2)

The above equation can be obtained by taking logarithms on both sides of Equation (23): lnp(U,D,W,I&|R,A,Q,T,σU2,σW2,σD2,σK2,σR2,σA2,σQ2)=12σQ2i=1nj=1lIijQ(qijh(Cnn(W,Ti)TKj))212σR2i=1mj=1nIijR(rijh(UiTCnn(W,Tj)))212σA2i=1mj=1lIijA(αijh(UiTKj))212σU2i=1mUiTUi12σD2j=1n(DjCnn(W,Tj))212σW2i=1|Wi|WiTWi12σK2i=1lKiTKii=1mj=1nIijRlnσRi=1mj=1lIijAlnσAi=1nj=1lIijQlnσQpi=1mlnσUpi=1llnσKpi=1nlnσDpi=1|Wi|lnσW+C

where p denotes the dimension of the implicit eigenvector and C is a constant. The maximization formula (24) can be regarded as an unconstrained optimization problem, which is equivalent to the minimization formula (25): E(U,D,W,K,R,A,Q)=φQ2i=1nj=1lIijQ(qijh(Cnn(W,Ti)TKj))2+12i=1mj=1nIijR(rijh(UiTCnn(W,Tj)))2+φA2i=1mj=1lIijA(αijh(UiTKj))2+φU2i=1mUiTUi+φD2j=1n(DjCnn(W,Tj))2+φw2i=1|Wi|WiTWi+φK2i=1lKiTKi

where the value of parameter φQ,φA,φU,φD,φW,φK is σR2σQ2,σR2σA2,σR2σU2,σR2σD2,σR2σW2,σR2σK2 , respectively.

Model application and analysis
Comparative Experiments on Cognitive Diagnostic Models
Experimental environment

The experiment was completed under the Python 3.11.7 environment, using the deep learning open source framework Pytorch, built on the Ubuntu 24.10 LTS operating system, remotely assisted by PyCharm software, and using an RTX4090D graphics card for high-performance computing.

Experimental data set and preprocessing

In this experiment, two representative public datasets in the field of smart education were used: the ASSISTMents2015 dataset and the EdNet dataset. The ASSISTMents 2015 dataset contains the answers of some students from 2015 to 2016 collected by the ASSISTMents platform, and contains rich information such as students, exercises, and knowledge points. While EdNet is a massive dataset that collects hundreds of millions of learning behavior records from nearly a million students from multiple platforms, in this experiment, the focus is only on information about the students, the exercises, the knowledge points, and whether or not the students answered the questions correctly as a result.

In this experiment, the dataset was preprocessed as follows: first, exercises without knowledge point annotations and students with less than 20 answer records were filtered out. In addition, considering that the EdNet dataset is too large, 1200 relatively active students were randomly selected for the diagnostic study during the experiment. Meanwhile, the experiment was divided to get the training set, validation set and test set by taking the ratio of 7:1:2. After data preprocessing, the statistics of the two data sets are shown in Table 1.

The statistics of datasets

Statistical term ASSISTMents2015 EdNet
Number of students 4286 1200
Number of exercises 18026 13624
Knowledge points 131 195
Number of answer records 289653 1125341
Number of correctly answered exercises 201328 836725
Number of incorrect answers to exercises 88325 228616
Experimental setup

Comparison of models

In this experiment, traditional diagnostic models, such as IRT, MIRT and DINA, and the latest neurocognitive diagnostic models, such as NeuralCD and RCD, are selected for comparison in order to validate the effectiveness of the cognitive diagnostic model proposed in this paper, which is based on graphical convolutional neural networks. The comparison models used are described as follows:

IRT: It is a continuous nonlinear model, which is mainly used to model the relationship between one-dimensional students’ abilities and the characteristics of the exercises (e.g., difficulty of the exercises, differentiation of the exercises, etc.).

MIRT: It is an extension of the traditional unidimensional item response theory model that takes into account students’ fine-grained abilities and models student abilities and exercise characteristics from the perspective of multidimensional knowledge points.

DINA: is a discrete diagnostic model that views student cognitive states as multidimensional and independent, with students either mastering or not mastering in relation to knowledge points. It also focuses on the student’s guessing factor and sliding factor.

NeuralCD: Neural networks are introduced to model the complex interactions between students and exercises, and the monotonicity assumption in traditional diagnostic models is used to ensure the interpretability of student factors and exercise factors. The model uses multidimensional successive vectors to model students and exercises, where each dimension of the vectors represents a different knowledge point. For the student factor vector, each dimension corresponds to the degree to which the student has mastered the knowledge point. For the exercise factor vector, each dimension corresponds to the degree of relevance of the exercise to the knowledge point.

RCD: Modeling student-exercise-knowledge point relationships based on a hierarchical graph structure, especially the dependency graph of knowledge concepts.

Evaluation Indicators

Considering that students’ ability level cannot be directly measured in cognitive diagnosis, this experiment used some common indicators to evaluate the performance of the model, such as root mean square error (RMSE), accuracy (ACC) and area under the ROC curve (AUC). From a regression perspective, RMSE is used to measure the difference between the predicted value and the actual score (0 or 1), with smaller values being better. From a categorization perspective, predicting whether a student will answer an exercise correctly can be viewed as a binary categorization problem. The predicted value is set to 1 if the predicted value is greater than or equal to 0.5, and 0 if it is not, and is measured using the ACC and AUC, with values as close to 1 as possible.

Analysis of experimental results

The experimental results of the GCN-based cognitive diagnostic model proposed in this paper on the ASSISTMents2015 and EdNet datasets as well as their comparison experiments are shown in Table 2.

Overall results on student performance prediction

Model ASSISTMents2015 EdNet
ACC↑ RMSE↓ AUC↑ ACC↑ RMSE↓ AUC↑
IRT 0.6555 0.5361 0.6092 0.6921 0.4654 0.7118
DINA 0.6659 0.5337 0.6788 0.6894 0.4871 0.6826
MIRT 0.7033 0.4695 0.7251 0.7047 0.4476 0.7273
NeuralCD 0.7406 0.4481 0.7605 0.7166 0.4304 0.7784
RCD 0.7421 0.4349 0.7938 0.7299 0.4239 0.7865
This article 0.7826 0.4271 0.8072 0.7497 0.4043 0.7907

From the data in Table 2, it can be seen that the ACC, RMSE, and AUC of this paper’s model on the ASSISTMents2015 dataset are 0.7826, 0.4271, and 0.8072, respectively, whereas on the EdNet dataset, it achieves 0.7497, 0.4043, and 0.7907, respectively. Compared with the other comparative models, especially the state-of-the-art hierarchical graph-based RCD model, this paper’s GCN-based cognitive diagnostic model achieves the best performance on all evaluation metrics, i.e., ACC and AUC achieve the maximum value and RMSE achieves the minimum value. This phenomenon indicates the effectiveness of the graph convolutional neural network model in combination with the traditional cognitive diagnostic model. In addition, the results from the ASSISTMents2015 dataset show that the model in this paper has a more stable performance when the dataset is sparse, which suggests that the interaction function module between the student vectors and the test question vectors plays an important role in this model. In addition, observing all the comparative experiments found that the neural network-based approach has better performance than the traditional cognitive diagnostic model, which indicates that the neural network-based approach can more effectively tap into the complex relationships among students, exercises and knowledge points.

Construction of student user profile based on cognitive diagnosis

In this section of the experiment, the student score matrix X and knowledge point examination matrix Q are obtained as inputs in the GCN-based cognitive diagnostic model after data preprocessing, and the EM algorithm is continuously looped and iterated to update the parameter failure rate s and guessing rate g, and eventually the updated s and g are obtained and the knowledge point mastery matrix A is computed, and the resulting cognitive portraits of some of the students are shown in Fig. 5.

Figure 5.

Student cognitive portrait

From Fig. 5, it can be seen that the students’ knowledge point mastery varies, for example, the cognitive vector α = (1,1,0,0,0,0,0,1,1,1,0) for the student numbered S1, and the cognitive vector α = (1,1,1,1,1,1,1,1,1,1) for the student numbered S2. Teachers can understand the cognitive state of students based on the student user portraits derived from the GCN-based cognitive diagnostic model, and students can also perform targeted gap-finding based on their cognitive portraits.

The student score matrix X obtained from preprocessing is also used as the input of S-P table (Student-Problem Chart) analysis to get the student learning state portrait as shown in Fig. 6. Teachers can understand the students’ learning status from the learner state portrait obtained from the S-P Table analysis and pay targeted attention to students with high student warning coefficients.

Figure 6.

Student learning status portrait

Experiments and Analysis of Personalized Recommendation of Teaching Content
Experimental design and data set description

In this section, five recommendation methods are selected for comparison experiments, including the random recommendation selection strategy Random, the selection strategy DT based on multiple decision trees, the selection strategy IRT based on Item Response Theory and the teaching resource recommendation method based on the CUPMF model described in this paper.

Experimental environment: the experimental platform Windows11, program implementation using Python3.11.7. The data required to conduct the experiment is shown in Table 3.

Experimental data

Network node 5 knowledge points in each chapter
Data 15623 student history answer data
KU relationship
Recommended number of test questions n
Number of question banks

EXPERIMENTAL DESIGN: A group of target students in each of the three categories of A, B, and C were selected for test question recommendation and evaluation observation. The experimental test was conducted at the time when the students’ historical answer records reached 30 questions.

Taking the students of category A as an example, the recommendation method of this paper was applied to recommend 10 test questions for 10 students of category A respectively, totaling 100 questions, and 20 recommended test questions were obtained by removing the duplicated test questions among them. At the same time, without communicating with each other, two experienced classroom teachers were asked to recommend two groups of 10 questions, respectively, to get 19 effective test questions, and the teacher recommendation results were used to evaluate the accuracy of the algorithm recommendation. Similarly, three groups of students, A, B, and C, were tested for recommendation and response respectively, and test data were obtained. The same method is used to obtain the recommendation test results of the remaining three comparative recommendation methods (Random, DT, and IRT), and the test result data obtained from the three types of students and the five recommendation methods are analyzed to compare the performance in terms of the common indicators of the model, the recommendation accuracy, and the reasonableness of the recommendation, respectively.

Evaluation indicators

In order to assess the effectiveness and accuracy of the algorithm, a variety of metrics were used to comprehensively evaluate the performance of the algorithm.

Precision, Recall and F1

Precision, Recall and F1 metrics are used to evaluate the performance of the CUPMF model and other comparative methods in terms of recommendation algorithm modeling. Precision, Recall and F1 are specifically defined as follows: precision=TPTP+FP recall=TPTP+FN F1=2*precision*recallprecision+recall

Where TP in Eq. denotes the number of recommended test questions that correctly hit the teacher-recommended test questions, FP denotes the number of recommended test questions that failed to correctly hit the teacher-recommended test questions, and TN denotes the number of algorithm-recommended test questions that did not include the teacher-recommended test questions.

MAE

MAE (Mean Absolute Error) measures the average of the absolute error between the predicted and actual values, and is used here to reflect the difference between the positive response rate of the test question recommendation results and the statistical expectation, i.e: MAE=i=1n| predicatediactuali |n

where n is the number of test questions, predicatedi is the predicted positive response rate value of the test questions for that student, and actuali is the true expected positive response rate value for that category of students.

Standard deviation of ideal response rate

Unlike the “optimal” tendency of test recommendation questions and common product recommendations, a higher positive response rate is not better for students doing the exercises. Different types of students have different positive response rates within a certain range to achieve optimal learning results.

In this paper, students are categorized into three types of students, including Class A students with good foundation, Class B students with moderate foundation, and Class C students with poor foundation. The optimal range of positive answer rate varies among the three categories of students. The expected positive response rate for recommending test questions to different students depends on a number of factors, such as the difficulty of the test questions and the knowledge level of the students. Based on the analysis of the historical data of the students in the category and the answer records of the students, the historical estimated correct response rate is found, and close to this range, it is considered to have found a suitable range for the students, in which the students will learn better.

As a result, the correct response rate used in this paper, defined as CRR (Correct response rate), i.e: CRR¯=RightTotal

Where Total is the total number of answers of the student’s type of student group in the recommended test questions, Right is the number of correct answers of the students of this type in the recommended test questions, and CRR¯ can reflect the actual difficulty of the recommended questions for this type of students.

Based on the statistically obtained historical positive answer rate of the three types of students, in this paper, the experiment is selected as the average positive answer rate of the Nth time of the historical answer record of the same type of students, which is regarded as the ideal positive answer rate of this type of students in the Nth time of the answer. In the experiment, the 31st answer record was selected, and the ideal positive answer rate for type A students was 0.5287, for type B students was 0.4530, and for type C students was 0.4089. The standard deviation of the experiment was calculated based on the ideal positive answer rate for the corresponding type of students.

Analysis of experimental results

Complete the experiment, calculate the evaluation indexes, and get the comparison results of the index data of different algorithm models as shown in Table 4~Table 6. Table 4~Table 6 represent the indicator data of the test results of A, B and C students respectively.

Comparison of indicator data for test results of class A student

Algorithm/model Precision Recall F1 MAE CRR¯ std. deviation KU
Random 0.5562 0.7544 0.6403 0.0525 0.0235 1,2,4
DT 0.6164 0.7613 0.6812 0.0387 0.0016 2,3,4
IRT 0.6599 0.5836 0.6194 0.0656 0.0421 1,2,4,5
PMF 0.8599 0.8774 0.8686 0.0308 0.0045 1,2,4
CUPMF 0.8953 0.9105 0.9028 0.0147 0.0018 1,2,4

Comparison of indicator data for test results of class B student

Algorithm/model Precision Recall F1 MAE CRR¯ std. deviation KU
Random 0.6648 0.7534 0.7063 0.0289 0.0182 1,2,4,3
DT 0.7081 0.7521 0.7294 0.0226 0.0129 1,2,4
IRT 0.7541 0.8453 0.7971 0.0188 0.0141 1,2,4
PMF 0.8028 0.9052 0.8509 0.0111 0.0023 1,2,4
CUPMF 0.8517 0.9121 0.8809 0.0136 0.0019 1,2,4

Comparison of indicator data for test results of class C student

Algorithm/model Precision Recall F1 MAE CRR¯ std. deviation KU
Random 0.6062 0.7651 0.6764 0.0285 0.0051 1,2,4,5
DT 0.7036 0.8247 0.7594 0.0275 0.0118 2,3,4
IRT 0.6796 0.8247 0.7452 0.0223 0.0138 1,2,4,3
PMF 0.7637 0.8839 0.8194 0.0182 0.0032 1,2,4
CUPMF 0.8046 0.8921 0.8461 0.0141 0.0029 1,2,4

According to the experimental results, the data comparison table for each category is comprehensively observed. From the table, it can be seen that comparing the decision tree algorithm, the algorithm designed in this paper has a good performance in precision rate, recall rate and F1 indexes, and at the same time, after optimizing the sub-module and improving the performance of prediction network, the overall ability of the model has a small increase, which can reflect that the model in this paper has a higher performance than the other models.

In terms of MAE index, this paper’s model has the best performance, the mean value of error are lower, the MAE value of the test results of the three categories of students in A, B and C are 0.0147, 0.0136 and 0.0141, respectively. In terms of the standard deviation of the CRR, this paper’s model has the best recommendation, its CRR¯ standard deviation is the smallest, and the standard deviation of the test results of the three categories of students in A, B and C are CRR¯ 0.0018, 0.0019 and 0.0029, indicating that it is closest to the ideal correct response rate in the more expected range, so it is more reasonable., 0.0019 and 0.0029, indicating that it is closest to the ideal correct response rate, which is in the more expected range, and therefore more reasonable. Among the three categories of A, B and C students, the standard deviation of the C students with poorer foundation is the largest, reflecting that the algorithm needs to be improved to consider more about the reasonableness of the recommendation for the C students. In addition, from the point of view of the number of knowledge units (KUs) included in the recommended test questions, the model in this paper has good diversity in the recommended test questions while improving personalization.

Conclusion

In this paper, we optimize the efficiency and accuracy of personalized recommendation of teaching content by constructing a student cognitive diagnosis model based on GCN and a teaching resource recommendation model based on convolutional joint probability matrix factorization (CUPMF).

First, the cognitive diagnostic model in this paper achieved the best performance against other models with ACC, RMSE and AUC of 0.7826, 0.4271 and 0.8072 on the ASSISTMents2015 dataset and 0.7497, 0.4043 and 0.7907 on the EdNet dataset, respectively. This indicates that the graph convolutional neural network model can be effectively combined with the traditional cognitive diagnosis model, which can more effectively tap into the complex relationship between students, exercises, and knowledge points.

Secondly, the cognitive portrait of student users and the learning state portrait are constructed based on the cognitive diagnosis results of students, so as to understand the cognitive state of students based on the cognitive portrait of student users, and to understand the learning state of students based on the learner state portrait, as well as to pay targeted attention to the students with high student warning coefficients.

Finally, the personalized recommendation algorithm for teaching content designed in this paper has good performance in terms of precision rate, recall rate and F1 indexes, and from the point of view of the two indexes of MAE and CRR¯ standard deviation, it has the best performance, with lower error mean and CRR¯ standard deviation, indicating that it is the closest to the ideal correct response rate, which is in the range of more expected, and thus more reasonable. Meanwhile, from the point of view of the number of knowledge units (KUs) included in the recommended test questions, the model in this paper has a good diversity of recommended test questions while perfecting personalization.

Funding:

Higher Education Teaching Reform Research and Practice Project of Hebei Province: Construction and Practice of Innovation and Entrepreneurship Education Curriculum System Based on “Whole Field Double”.

Higher Education Teaching Reform Research and Practice Project of Hebei Province: Construction and practice of the curriculum system of "Labor Education" in local normal universities under the background of application transformation development -- based on the perspective of "vocational maturity" (2023GJJG371).

Research and Practice Project on Teaching Reform of Innovation and Entrepreneurship Education of Hebei Provincial Department of Education: Construction of Innovation and Entrepreneurship Education Curriculum System Based on the Linkage of Three Courses: A Case Study of Primary Education (2023CXCy175).

Sprache:
Englisch
Zeitrahmen der Veröffentlichung:
1 Hefte pro Jahr
Fachgebiete der Zeitschrift:
Biologie, Biologie, andere, Mathematik, Angewandte Mathematik, Mathematik, Allgemeines, Physik, Physik, andere