Otwarty dostęp

Construction of personalized online educational resources based on deep learning in higher education self-study examination environment

 oraz   
19 mar 2025

Zacytuj
Pobierz okładkę

Introduction

Online learning has many synonyms, such as e-learning, e-learning, online teaching, online education, online training, Internet education or distance education, distance learning, digital learning, etc. In fact, these terms are more or less the same, and they are only called differently at different times, in different fields or from different angles. For example, from the perspective of the learner is online learning, from the perspective of the pedagogue, the teaching institution or platform and the society is online education, and the research field uses more of the two conceptual terms of online learning and online education [1-4]. Online education refers to the use of network technology, multimedia technology and other modern information technology means to carry out a new form of education, is built on the basis of modern electronic information and communication technology, it is based on the learner as the main body, the main use of a variety of media and a variety of interactive means of systematic teaching and interactive contact between students and teachers, students and educational institutions. Simply put, online education is the form of education carried out in the Internet environment. At present, academic education is not the goal of online education, online education is more auxiliary education, continuing education, vocational training and lifelong learning [5-9]. After 2017, online education has been led by various technologies from gradual maturity to in-depth development, and the various vertical fields of the education industry, such as pre-school education, K12 education, higher education, etc., have ushered in the deep integration of science and technology, and the education model The education mode and learning mode have all changed greatly, and the specific content of education, the tools used, and the forms adopted have all been comprehensively innovated and transformed. The education industry, relying on the high-speed development of mobile Internet, big data and artificial intelligence, will truly realize mass learning and lifelong learning [10-13].

With the dramatic increase in the number of network online education users, in order to meet the personalized needs of different users for learning resources, the number of online education resources has also increased exponentially year by year. At present, more than two hundred of the high-quality catechism courses offered by more than one thousand Chinese universities have landed on the relevant course platforms in Europe and the United States [14-17]. The huge amount of educational resources can provide online learners with rich sources of information and learning opportunities, but at the same time, it also creates the problems of “educational overload” and “knowledge lost”, which is manifested in the fact that learners need to spend a lot of time and energy to search and check the resources, which greatly reduces the learning efficiency and reduces the learning time of the learners. The efficiency of learning is greatly reduced, and the initiative of learning is minimized [18-20]. In the environment of higher education self-study examination, how to use deep learning technology to select suitable learning resources for learners from the massive learning resources is an urgent problem to be solved for realizing intelligent education [21-22].

This paper proposes a personalized education recommendation method based on knowledge tracking.The study focuses on the impact of learners’ forgetting behavior on their knowledge mastery, and proposes a deep knowledge tracking model (LFKT) that integrates learning and forgetting. The attention layer takes the exercise questions and the set of knowledge points covered by the exercise questions as inputs, calculates the knowledge correlation weights between the exercise questions and each knowledge point, filters out the knowledge points covered by the exercise questions with a knowledge point filter, and calculates the knowledge correlation between the exercise questions and the knowledge points they cover. The weights are used to calculate the learner’s comprehensive mastery of the knowledge points covered by the exercises embedded vector to predict the learner’s performance in answering the questions. The algorithm design of the supplementary exam resource recommendation method has been completed and tested on the dataset to check its performance effects.

Personalized education shared resource platform based on knowledge tracking
Cognitive Diagnostic Theory

Cognitive diagnostic theory is a type of measurement theory in educational psychology, which plays an important role in diagnosing the cognitive-behavioral process and structure of learners.Meanwhile, cognitive diagnostic theory has been increasingly concerned by scholars at home and abroad, and its model has been expanding. However, the theoretical basis of the model is also different. The main models include item response theory, rule space model, and attribute hierarchy model.

Project Response Theory

Item Response Theory (IRT) is the classic cognitive diagnostic model constructed in the field of psychometrics. IRT is a probabilistic description of how item response outcomes to item answering are affected by the joint action of the student’s level of competence and the item characteristics in the form of a typical item characteristic curve [23].

The project contains three parameters: differentiation parameter a, which is the slope of the curve on, indicates the level of differentiation between different students on each test question. Difficulty parameter b, which is the position on the power scale of the curve, indicates the difficulty of the test questions. Parameter c, the probability of guessing the correct answer to the test question, is the asymptote.

Commonly used item characteristic function models are single-parameter model, two-parameter model and three-parameter model. The item characteristic function of each model is shown in equations (1)~(3):

Single-parameter model: Pij(θj)=11+eV(θjbi) \[{{P}_{ij}}({{\theta }_{j}})=\frac{1}{1+{{e}^{-V({{\theta }_{j}}-{{b}_{i}})}}}\]

A two-parameter model: Pij(θj)=11+eVai(θjbi) \[{{P}_{ij}}({{\theta }_{j}})=\frac{1}{1+{{e}^{-V{{a}_{i}}({{\theta }_{j}}-{{b}_{i}})}}}\]

Three-parameter modeling: Pij(θj)=ci+(1ci)11+eVai(θjbi) \[{{P}_{ij}}({{\theta }_{j}})={{c}_{i}}+(1-{{c}_{i}})\frac{1}{1+{{e}^{-V{{a}_{i}}({{\theta }_{j}}-{{b}_{i}})}}}\]

Pji(θi) denotes the specific meaning of the probability that a student will answer test question j correctly when the response situation is θi, where v is the constant 1.7, e is the base of the natural logarithm which is lnN, i is the number assigned to the student, θi represents the student’s i proficiency value, j the numbering of the test question, and aj,bj,cj denotes the discriminatory, difficulty, and the guessing parameter of the test j question, respectively.

DINA model

DINA is a multidimensional discrete cognitive diagnostic model for item response functions defined as in equation (4), which represents the probability that student i will answer correctly on test question j.

P(xij=1|αi,qj,ηij)=(1sj)ηg¯gj(1ηg¯) \[P({{x}_{ij}}=1|{{\alpha }_{i}},{{q}_{j}},{{\eta }_{ij}})={{(1-{{s}_{j}})}^{{{\eta }_{{\bar{g}}}}}}\;g_{j}^{(1-{{\eta }_{{\bar{g}}}})}\]

ηij denotes that student i completes the answer to test question j without taking into account the error rate and guessing rate, the number of knowledge points is K, and the calculation is expressed as equation (5).

ηij=k=1Kαikqik \[{{\eta }_{ij}}=\prod\limits_{k=1}^{K}{{{\alpha }_{ik}}^{{{q}_{ik}}}}\]

For student i, the value of ηni can be 1 only if he/she has mastered the relevant knowledge attributes of test j. When the value of ηi is 0, it means that student i has not mastered any of the relevant knowledge attributes of test j. In addition, equations (6) and (7) are used to express the specific meanings of error rate sj and guessing rate gi, respectively. si indicates the probability that a student has mastered all the knowledge attributes to be examined in the test question to be answered but still answers incorrectly, and gi indicates the probability that a student has answered correctly without having fully mastered all the knowledge attributes to be examined in the test question to be answered, and there are only two parameters for each item in the DINA model, i.e., the s failure parameter and the g guessing parameter.

sj=P(xij=0|αi,qj,ηij=1) \[{{s}_{j}}\;=\;P(\;{{x}_{ij}}\;=\;0\;|\;{{\alpha }_{\;i}},{{q}_{\;j}},{{\eta }_{\;ij}}\;=\;1)\] gj=P(xij=1|αi,qj,ηij=0) \[{{g}_{j}}\;=\;P(\;{{x}_{ij}}\;=\;1\;|\;{{\alpha }_{\;i}}\;,{{q}_{\;j}}\;,{{\eta }_{\;ij}}\;=\;0)\]

The association between test questions and knowledge points is portrayed in the DINA model by introducing the test question knowledge point association matrix Q, which is defined and labeled by experts for each test question.

Neurocognitive diagnostic models

In traditional cognitive diagnostic models, the interaction function between students and test items mainly relies on expert design, and the disadvantages of manual design are that it is more labor-intensive and the simplified function fails to reflect the intricate relationship between subjects and test items. However, the use of neural networks and cognitive diagnostic related technology combination provides a new idea, neurocognitive diagnostic model in the traditional cognitive diagnostic model based on the solution to this problem, it is directly from the data to learn students, knowledge and test scores of the interaction function, rather than manually designed, in the assessment of the state of knowledge and ability level of the student has a good performance.

The neurocognitive diagnostic model consists of three parts: the student’s mastery of the knowledge points, the factors in the test questions, and the model’s interaction function. The major difference between the neurocognitive diagnostic model and the initial cognitive diagnostic model is that the interaction function is composed of multiple layers of neural networks interlinked with each other. The inputs are the student’s mastery level on the test question, the R matrix of the student’s response record on the test question, and the relevance matrix Q of the test question’s knowledge points, and the output is the predicted probability of the student answering the test question correctly.

To ensure that each dimension corresponds to the student’s mastery level on the corresponding knowledge point at the end of model training, we make the following constraints.

1) The input layer needs to contain FsFkn (“∘” means multiply by elements), which is done so that each dimension in Fs corresponds to a knowledge point in the corresponding dimension in Fkn.

2) The “monotonicity assumption” constraint. If the students’ mastery of a knowledge point is increasing, the probability of correctly answering a test question also increases monotonically and not necessarily strictly monotonically. In the training process of the model, Fkn controls the relevant dimensions in Fs that need to be adjusted, and when the predicted value of test responses is smaller than the true response value, the value in Fs is increased, and vice versa, the value in Fs is decreased.

LFKT knowledge tracking algorithm

In this paper, we propose a deep knowledge tracking model LFKT that integrates learning and forgetting.The LFKT model is divided into attention layer forgetting layer prediction layer learning layer knowledge level output layer.

Attention layer

The role of the attention layer is to compute the knowledge correlation weights between exercises and knowledge points. The inputs to the attention layer are the learner’s current exercise topic et and the set of knowledge points covered by the topic kt. In order to map et to a continuous vector space, et is multiplied by the exercise embedding matrix A(dk×|E|) to generate a dk -dimensional exercise embedding vector vt. where each dk -dimensional column vector in the exercise embedding matrix A is an embedding representation of an exercise. The knowledge points covered by each exercise are labeled by the expert and stored in the set kt. In this paper, we filter out irrelevant knowledge points and retain the knowledge points covered by the exercises through the knowledge point filter (K-Filter). Matrix Rt stores the embedding vectors of the knowledge points covered by the exercises, where each dk -dimensional column vector in matrix Rt is the embedding vector of a knowledge point covered by an exercise. The inner product between the embedding vector vt of the exercise and the embedding vector Rt(i) of the covered knowledge point is calculated, and the Softmax value of the inner product is calculated and stored in vector RSt. Vector RSt represents the knowledge-related weights between exercise vt and the knowledge points covered by the exercise. See Equation (8) for details.

RSt(i)=Softmax(vtTRt(i)) \[R{{S}_{t}}(i)=Soft\max ({{v}_{t}}^{T}{{R}_{t}}(i))\]

Softmax(zi)=ezi/jezi$Softmax({{z}_{i}})={{e}^{zi}}/\sum\limits_{j}{{{e}^{{{z}_{i}}}}}$ of them. The |K| -dimensional vector wt is a vector of knowledge-related weights between the exercises and all knowledge points, and since the knowledge points covered by the exercises are filtered out by the knowledge filter, the model needs to put the knowledge-related weights of the knowledge points covered by the exercises into the corresponding positions of wi. First, the |K| -dimensional all-zero vector wt is initialized. First, the 5-dimensional all-zero vector 6 is initialized, i.e., wt ←[0,⋯⋯,0]. After that, the weights of each dimension of RSt are put into the corresponding position of wt, i.e., wi[ki[i]]–RSt[i], to obtain the knowledge-related weights of the exercises and each knowledge point.

Oblivion layer

Since the learner forgets the knowledge points that have not been reviewed in time during the learning period, and also forgets the knowledge in the interval between two learning sessions, the knowledge mastery embedding Mt.lv$M_{t.l}^{v}$ at the end of the last learning session is not the same as the knowledge mastery embedding MtFV$M_{t}^{FV}$ at the beginning of the current learning session. The forgetting layer rooted the factors affecting knowledge forgetting Ft to forget the knowledge mastery embedding matrix Mt.Iy$M_{t.I}^{y}$ at the end of the last study to get the knowledge mastery embedding matrix MtFV$M_{t}^{FV}$ at the beginning of the current study of the learner. inspired by the forgetting gate and the input gate in the LSTM, according to the factors affecting knowledge forgetting Ft to update the Mt1v$M_{t-1}^{v}$, the first step is to erase the original information in the matrix , and then to write the information. For modeling forgetting behavior, the erasure process controls the decline of the learner’s mastery of the knowledge point, and the writing process controls the update of the learner’s mastery of the knowledge point [24].

A fully connected layer with a Sigmoid activation function is utilized to convert the factors Ft(i) affecting the learner’s forgetting level of knowledge point i into the corresponding memory erasure vector fet(i) for knowledge point i: fet(i)=Sigmoid(FETFt(i)+bfe) \[f{{e}_{t}}(i)=\text{Sigmoid(}F{{E}^{T}}{{F}_{t}}(i)+{{b}_{fe}}\text{)}\]

where Sigmoid(zi) = 1/(1+ezi). The weight matrix FE of the fully connected layer is of shape (dv + dcdv and the bias vector bfe of the fully connected layer is of dimension dv. The memory erasure vector fei(i) is a column vector of dimension dv and all values in the vector are in the range (0, 1). A fully connected layer with a Tanh activation function is utilized to convert the factors Fd(i) affecting how much the learner has forgotten about a knowledge point i into a memory update vector fat(i) corresponding to the knowledge point i: fat(i)=Tanh(FATFt(i)+bfa) \[f{{a}_{t}}(i)=\operatorname{Tanh}(F{{A}^{T}}{{F}_{t}}(i)+{{b}_{fa}})\]

where Tanh(zi) = (ezje−zj)/(ezj+e−zj). The weight matrix FA of the fully connected layer is of shape (dv+dcdv, the bias vector bfa of the fully connected layer is of dimension dv, and the memory update vector fah(i) is a column vector of dimension dv. MtFV$M_{t}^{FV}$ is obtained by updating Mt1v$M_{t-1}^{v}$ according to the memory erase vector and the memory update vector: MtFV(i)=Mt1V(i)(1fat(i))(1+fat(i)) \[M_{t}^{FV}(i)=M_{t-1}^{V}(i)(1-f{{a}_{t}}(i))(1+f{{a}_{t}}(i))\]

The forgetting layer outputs the knowledge mastery embedding matrix at the beginning of this study MtFV$M_{t}^{FV}$. The forgetting layer can be used to model the learner’s forgetting process of different knowledge points separately, and to calculate the degree of forgetting of different knowledge points according to the difference in the learner’s learning history of different knowledge points.

Learning layer

After a learner takes a course, the online education system prepares questions for the learner to test the learner’s learning. The learning layer tracks the changes in knowledge acquisition during the learning process based on the learner’s answer results, and models the learning behavior by updating the knowledge acquisition embedding matrix MtFV$M_{t}^{FV}$ at the beginning of the learning process to the knowledge acquisition embedding matrix Mtv$M_{t}^{v}$ at the end of the learning process. Tuple (et,rt) represents the learner’s answer result at time t. In order to map tuple (et,rt) to a continuous vector space, tuple (et,rt) is multiplied with the answer result embedding matrix B of size dv*2|E| to obtain a dv -dimensional answer result embedding vector st. The learning layer takes the question-answering result embedding vector st and the knowledge-related weight vector wt corresponding to the exercise questions as inputs, and uses the learner’s practice results in the online education system as the indirect feedback of the learner’s learning effect, updating the learner’s knowledge mastery state in the process of learning and modeling the learning behavior through the LSTM network: Mtν(i)=LSTM(st,wt(i)MtFV(i);θ) \[M_{t}^{\nu }(i)=\text{LSTM(}{{s}_{t}},{{w}_{t}}(i)M_{t}^{FV}(i);\theta \text{)}\]

As can be seen from Eq. (12), the learning layer only updates the learner’s knowledge mastery embedded in the knowledge points covered by the exercises, and the learning layer does not process the knowledge points that have nothing to do with the exercises.

Forecasting layer

The purpose of the prediction layer is to predict the learner’s performance on the next candidate exercise et+1. Since the learner’s forgetting behavior during the two-answer interval affects his/her knowledge mastery state, the prediction layer predicts the probability of the learner answering Exercise et+1 correctly based on the knowledge mastery embedding matrix Mt+1FV$M_{t+1}^{FV}$ at the start of the answer at the current moment. By weighting and summing the knowledge-related weights wt+1 and the learner’s knowledge mastery embedding Mi+1FV$M_{i+1}^{FV}$ at the start of the current moment, the resulting vector dn+1 is the weighted mastery embedding vector of the knowledge points covered by the exercise.

dt+1=i=1Kwt+1(i)Mt+1FV(i) \[{{d}_{t+1}}=\sum\limits_{i=1}^{K}{{{w}_{t+1}}}(i)M_{t+1}^{FV}(i)\]

The success of the learner in answering an exercise is not only related to the learner’s comprehensive mastery of the knowledge points covered by the exercise, but also related to the exercise itself, so the combined vector [dt+1,vt+1] obtained by connecting vectors dt+1 and vt+1 is input to a fully connected layer using the Tanh activation function, and the output vector ht+1 contains not only the learner’s comprehensive mastery of the knowledge points covered by the exercise, but also the characteristics of the exercise itself. Matrix WI and vector bI represent the weights and biases of the fully connected layer, respectively.

ht+I=Tanh(WIT[dt+I,vt+I]+bI) \[{{h}_{t+I}}=\text{Tanh(}W_{I}^{T}[{{d}_{t+I}},{{v}_{t+I}}]+{{b}_{I}}\text{)}\]

Finally, vector ht+1 is input into a fully connected layer with a Sigmoid activation function to obtain pt+1 which represents the probability of the learner answering the exercise correctly. Where matrix W2 and vector b2 represent the weights and biases of the fully connected layer respectively.

pt+1=Sigmoid(W2Tht+1+b2) \[{{p}_{t+1}}=\text{Sigmoid}(W_{2}^{T}{{h}_{t+1}}+{{b}_{2}})\]
Knowledge level output layer

The knowledge level output layer takes the knowledge mastery embedding matrix Mtv$M_{t}^{v}$ at the end of the learner’s study as input, and outputs a |K| -dimensional vector value representing the learner’s knowledge mastery level at the end of this study, with each dimension of the vector between (0, 1), indicating the learner’s mastery level of the knowledge point.

In the prediction layer, at each time node t, the learner’s performance on a particular exercise ei is predicted based on two inputs: the learner’s comprehensive mastery embedding vector dt for the knowledge points covered by that exercise and the exercise embedding vector vt. Thus, if one just wants to estimate the learner’s mastery of the i th knowledge point without any inputs from the exercise, one can omit the exercise embedding v while letting the learner’s knowledge mastery embedding i column Miv(i)$M_{i}^{v}(i)$ of matrix Miv$M_{i}^{v}$ as the input to the equation. After the output matrix Miv$M_{i}^{v}$ of the learning layer, in order to estimate the mastery level for the i th knowledge point, the weight vector βi = (0,⋯,1,⋯,0) is constructed in which the value of the i th dimension is equal to 1, and the knowledge mastery embedding vector Miv(i)$M_{i}^{v}(i)$ for the i th knowledge point is extracted, after which the knowledge mastery level is estimated using equations (17) and (18): Mtν(i)=βiMtν \[M_{t}^{\nu }(i)={{\beta }_{i}}M_{t}^{\nu }\] yt(i)=Tanh(WIT[MtV(i),0]+bI) \[{{y}_{t}}(i)=\operatorname{Tanh}(W_{I}^{T}[M_{t}^{V}(i),0]+{{b}_{I}})\] valuet(i)=Sigmoid(W2Tyt(i)+b2) \[valu{{e}_{t}}(i)=Sigmoid(W_{2}^{T}{{y}_{t}}(i)+{{b}_{2}})\]

Vector θ = (0,0,⋯,0) has the same dimension as the exercise embedding vt and is used to complement the vector dimension. Parameter Wl,W2,bI,b2 is exactly the same as in Eq. (14) and Eq. (15). The mastery level of each knowledge in the knowledge space is calculated sequentially to get the learner knowledge mastery vector valuet.

Model Optimization Objectives

The parameters that the model needs to be trained are mainly the exercise embedding matrix A, the answer result embedding matrix B, the neural network weights and biases, and the knowledge matrix MK. In this paper, we optimize each parameter by minimizing the cross-entropy loss function between the model’s prediction of the learner’s answer result pt and the learner’s true result of answering the question rt. The loss function is shown in Equation (19). In this paper, the Adam method is used to optimize the parameters.

L=t(rtlogpt+(1rt)log(1pt)) \[L=-\sum\limits_{t}{({{r}_{t}}\log {{p}_{t}}+(1-{{r}_{t}})\log (1-{{p}_{t}}))}\]
Personalized test question recommendation method

Personalized test recommendation is a kind of intelligent educational aids, which can accurately recommend suitable test questions for learners according to their learning situation, interests and subject characteristics, etc., helping students to better master their subject knowledge and improve their learning efficiency and performance. In order to achieve the above goal, students interact with learning materials (test questions, exercises, etc.) to achieve the purpose of improving performance and cognition, recommender systems play an important role in which the learning recommender system also came into being. Knowledge tracking technology can obtain the mastery of students’ knowledge points, and these data can provide the basis for personalized test recommendation, so as to better provide students with suitable test recommendation services. At the same time, the personalized test recommendation can adjust the difficulty of the recommended test questions according to the students’ answers, helping students to better improve and consolidate their knowledge, and it can also collect the students’ answer data and feedback it to the knowledge tracking model for more accurate knowledge analysis and evaluation.

Problem description

In the problem of test question recommendation in education, assume that there are N student SS = {s1,…,sn,…,sN} and M test questions ExS = {ex1,…,exi,…,exM}, and the online learning interaction behavior of the student can be defined as the sequence: R = {(e1,r1),…,(et,rt),…,(eT,rT)}, where etExS denotes the questions that the student sn has done at the moment of t, and rt denotes the corresponding scoring performance. The time state based on the time after the t moment is represented by ksT, βn,i denotes the difficulty for the student sn test question exi, and αn,i denotes the recommendation of test question exi to the student sn.

Extensive problem training is an important part of the teaching task. The purpose of doing problems is to master the knowledge rather than doing the problems themselves, and students digest and absorb the newly learned knowledge and consolidate the knowledge and methods they have already learned through doing problems and tests. However, there are individual differences in the degree of students’knowledge mastery, and there are a large number of test materials, such as different levels of difficulty, so how to get the most suitable test questions for their own knowledge status from the huge amount of test materials, and personalized test recommendations for each student is a difficult problem.

In recent years, researchers have made many attempts and studies on test question recommendations. Collaborative filtering-based recommendation algorithm is one of the most representative recommendation methods, which recommends test questions by considering the similarity between students or exercises. However, the collaborative filtering method adopts a static view in the recommendation process, so when the recommendation environment is placed in the field of education, it ignores the dynamic characteristics of the personalized recommendation scenario, and fails to take into full consideration of the knowledge points corresponding to the test questions and the dynamically changing personalized ability differences of the students, which leads to a lack of reasonable interpretability in the recommendation results, and the students are unable to know the reasons why the test questions are recommended, and they are also unable to know their own Students cannot know the reason why the test questions are recommended, nor can they know exactly where their weak areas are, so the recommendation results may not necessarily be beneficial to students’ ability improvement. In addition, some researchers recommend test questions to students by evaluating the consistency of the content of the test questions, and some scholars recommend test questions based on students’ cognitive diagnostic ability, but these methods do not directly measure the state of students’ knowledge, but rather quantify the students’ ability as knowledge points, and the students’ cognitive status is invisible to a certain extent, and there is a certain degree of invisibility, which will lead to a relatively poor recommendation result.

Personalized Test Question Recommendation Based on GFLDKT Model

Unlike traditional positive feedback recommendation algorithms based on user and item similarity, the problem of personalized test recommendation for students in an educational setting is more specific: the ultimate goal of test recommendation is not to recommend the most interesting test questions for students, but rather to consolidate their foundations and to improve their competence in the educational domain. Similarly, test recommendation is not simply a matter of recommending the hardest or easiest questions to students. The test questions that are the most challenging may be beyond the student’s knowledge base abilities, while the easiest test questions may already contain knowledge points that the student is already proficient in. In this situation described above, students face the dilemma of not being able to practice effectively with these test questions and therefore not being able to significantly improve their knowledge points. The personalized test question recommendation method based on the GFLDKT model is shown in Fig. 1, and in this paper, based on the students’ individualized ability differences, combined with the mastery of specific knowledge points, a reasonable test question difficulty is set to provide students with personalized test question recommendations [25].

Figure 1.

Personalized test recommendation method based on GFLDKT model

Specifically, the personalized test question recommendation method in this paper is able to obtain explicit student knowledge point mastery and take the probability of a student sn correctly answering a test question exi containing a certain knowledge point kcj as the student’s test question difficulty βnji, and finally, we recommend test questions of appropriate difficulty according to the learning needs of different students. Specifically, the students’ test question difficulty βnji is defined as follows: βnji=1r^T+1 \[{{\beta }_{nji}}=1-{{\hat{r}}_{T+1}}\]

Test question difficulty is an important component of test question recommendations. Whether a student’s focus is to fill in knowledge gaps by repeating knowledge with a low level of mastery, or to consolidate and improve knowledge with a high level of mastery in order to obtain higher grades, students with different cognitive abilities have different needs for test question difficulty, and test question difficulty boundaries for that knowledge point are set ζj1,ζj2(ζj1 < ζj2) in order to better assess the mastery of students. At the same time, according to the students’ answer prediction r^T+1${{\hat{r}}_{T+1}}$, you can create a personalized student test question recommendation ensemble ExSrec to provide students with targeted practice.

ExSrecn={exi|exiExSrec,βni[ζj1,ζj2]} \[Ex{{S}_{re{{c}_{n}}}}=\{e{{x}_{i}}|e{{x}_{i}}\in Ex{{S}_{rec}}\;,\;{{\beta }_{ni}}\in [{{\zeta }_{j1}},{{\zeta }_{j2}}]\}\]

On the other hand, with the help of massive interactive data of students’ test questions, according to the probability of students’ correctly answering the test questions, the general difficulty of the test questions is obtained, and the corresponding test question recommended mass ensemble ExOrec is established, in order to solve the problem that some learners use the online education platform for the first time, which leads to the lack of analyzing data to get the accurate mastery of the knowledge points (learners will be mapped out before they use the platform for the first time, and the data of the mapping test is utilized to The data from the test will be used to initially obtain the students’ knowledge point mastery status ks1).

The higher the average correct rate of a test question, the lower the difficulty of the question, so we counted the total number of times students answered the question correctly n00 and the total number of times they answered the question incorrectly n01 from the data set, and used the correct rate of the question to calculate the difficulty of the question ϕi: ϕi=1n00n02 \[{{\phi }_{i}}=1-\frac{{{n}_{00}}}{{{n}_{02}}}\] ExOrecn={exi|exiExOrec,ϕi[ζ1,ζ2]} \[Ex{{O}_{re{{c}_{n}}}}=\{e{{x}_{i}}|e{{x}_{i}}\in Ex{{O}_{rec}}\;,\;{{\phi }_{i}}\in [{{\zeta }_{1}},{{\zeta }_{2}}]\}\]

In the process of test question recommendation, the Softmax function is first used to normalize the test question representations, and then the difficulty of the test questions is used as a weighting coefficient to calculate the real response of students with different knowledge base abilities in completing the test questions with different difficulty coefficients: MEi=softmax(Ei)=exp(Ei)j=1K(exp(Ei)) \[M{{E}_{i}}=softmax({{E}_{i}})=\frac{exp({{E}_{i}})}{\sum\limits_{j=1}^{K}{(exp({{E}_{i}}))}}\] Dj=kstjMEij×βni \[{{D}_{j}}=k{{s}_{tj}}-M{{E}_{ij}}\times {{\beta }_{ni}}\]

Finally, a threshold value of θ is set to determine whether the final test recommendation is reasonable or not, if 0 < Djθ, it indicates that the recommendation is effective; if Dj < 0 or Dj > θ, it indicates that the recommendation is unreasonable, and the recommended test questions fail to help students consolidate their knowledge base and improve their performance.

Personalized Online Education Resource Platform

Hierarchy-based platform architecture is adopted, which is divided into 4-layer architecture of physical layer, data layer, application layer and service layer, and the structure is shown in Figure 2. The physical layer completes the resource access and integration of each institution, the data layer realizes the analysis and mining of data, the application layer creates relevant applications and specifies relevant standards, and the service layer realizes the customized development of the application system.

Figure 2.

System architecture of the online education resources cloud platform

Physical layer

The physical layer completes the access integration, extraction, and cleaning of data resources from each examination base, and establishes a cross-platform fusion and shared resource cloud. This layer mainly solves the association of resources across media and languages and the problem of knowledge implicitness through multi-source heterogeneous fusion management middleware; solves the problem of ubiquitous distribution of data resource nodes; solves the problem of multimodal and dynamic change of data; completes the efficient processing and semantic understanding of heterogeneous big data; hides the heterogeneous nature of the network components, realizes the transparency of the servers, the transparency of the network and the transparency of the language, and achieves the The precise and efficient integration of cross-campus resource libraries is realized.

Data Layer

The data layer is the core layer of the platform architecture, which manages the shared resource cloud and mainly contains three aspects:

(1) Resource management: Audit the resources that join the resource cloud, allocate, access, apply, billing management and so on according to the relevant mechanism.

(2) Platform security services: ensure data storage, transmission, use of security and operational stability.

(3) Data mining: data analysis middleware conducts in-depth mining of resources, categorizes data and discovers precise users.

Application Layer

The application layer mainly builds corresponding services on the basis of the data layer, which contains basic services such as basic resource base, recommendation push service, visualization report, etc. Each institution can build a personalized resource base on this basis according to its own situation, and improve and perfect the resource base of our university.In addition, it establishes application interfaces and stipulates the use agreement, access mode, and charging mode of the application to provide services to the upper layer.

Service Layer

The function of the service layer is mainly to complete personalized service customization by each institution according to its own needs.Through the application layer interface, it selects applications and related protocols, service modes, and other system strategies from the application library, and customizes the resource management system suitable for the university.

Model performance testing experiments
Knowledge Trace Performance Testing
Data set and baseline modeling

In order to evaluate the effectiveness of the method proposed in this paper, experiments are conducted on real data collected from 3 online learning platforms, the datasets are described as follows:

ASSIST2009: This dataset is provided by the ASSISTments online teaching platform and has been used in several papers for the evaluation of knowledge tracking models. After removing duplicate records, it contains 325,673 interaction records from 4151 students on 110 knowledge points.

ASSIST2015: this dataset collected 708,631 records from 19,917 students across 100 knowledge points. Although this dataset has more records than ASSIST2009, the average number of student interaction records is less due to the larger number of students.

ASSISTchall: This dataset collected 942,816 interaction records from 686 students across 102 knowledge points. This dataset has been used in the 2017 ASSISTments Data Mining Competition and has a higher average number of records per student.

In order to verify the validity of the model in this paper, the following seven typical benchmark models are selected to compare the experimental results:

BKT: The learning process is modeled as a Hidden Markov Process, with knowledge mastery as a binary variable of mastered and un-mastered, and learning is modeled as a discrete transition from un-mastered to mastered state of a knowledge point;

DKT: a pioneering work that introduces deep learning to knowledge tracking, using recurrent neural networks to model the student’s knowledge state at each input moment;

DKVMN: modeling students’ mastery of each knowledge point using a storage matrix, which solves the problem of DKT’s inadequacy in dealing with long sequence dependencies;

GIKT: uses Graph Convolutional Networks (GCN) to model problem-knowledge point correlations, GIKT summarizes students’ mastery of knowledge points as the interaction between students’ current state, students’ history of related exercises, the target problem, and related knowledge points;

EERNN: considers information about the semantics of the exercises and uses the semantic representations of the test questions as input to the model, taking into account the bias in prediction due to the large differences brought about by the semantics of two exercises possessing the same knowledge points;

LPKT: gives equal attention to the results of incorrect and correct answers, arguing that students will learn from them even if they answer incorrectly;

ATKT: enhances the generalization of knowledge tracking using adversarial training, proposing an aggregation module for the hidden states of knowledge while emphasizing the importance of the current knowledge state.

Evaluation indicators and model parameterization

Referring to most of the research works on knowledge tracking models, this experiment adopts accuracy (ACC) and area under the ROC curve (AUC) as the evaluation metrics from the classification point of view, where AUC evaluates the probability that a predicted positive example is ranked in front of a negative example.

The experiment divides 80% of the data in each dataset into training and validation sets, and 20% into testing sets. The experiments are performed in stochastic gradient descent using a learning rate r ∈ {0.0001,0.001}, an embedding dimension k ∈ {100, 150, 200}, a training batch size B ∈ {100, 150, 200, 300}, a number of hidden layers h ∈ {32, 64, 128}, and the number of neurons in the hidden layer of the LSTM model is set to 200.The results of the experiments are taken on each dataset average of 10 calculations. This experiment is run on a 64-bit, 3.6GHz GPU (RTX 3060 12G) computer. Figure 3 shows the loss curve of the model trained in this paper.The model basically reaches convergence after 46 rounds of iterations, and the RMSE is stable around 0.88.

Figure 3.

Model training loss curve

Experimental results

The experimental results are shown in Table 1, where the black bolding is the optimal effect obtained by this dataset in the corresponding index. The experimental results show that the model proposed in this paper achieves better results on all three datasets, with ACC values of 0.823, 0.824, and 0.818, and AUC values of 0.827, 0.816, and 0.821, respectively.The two metrics are the highest among all the models, and the best results are achieved.

Experimental results and model performance contrast

ASSIST2009 ASSIST2015 ASSIST chall
ACC AUC ACC AUC ACC AUC
Model 0.562 0.6 0.551 0.567 0.507 0.51
BKT 0.677 0.712 0.671 0.679 0.67 0.685
DKT 0.657 0.666 0.656 0.66 0.653 0.668
DKVMN 0.699 0.737 0.692 0.732 0.672 0.693
GIKT 0.705 0.723 0.714 0.753 0.688 0.727
EERNN 0.71 0.737 0.727 0.765 0.7 0.762
LPKT 0.738 0.781 0.724 0.764 0.721 0.752
ATKT 0.734 0.773 0.726 0.757 0.719 0.758
This model 0.823 0.827 0.824 0.816 0.818 0.821

Although our model obtains better performance, whether the model is interpretable or not is also an important evaluation aspect of knowledge tracking, in order to justify the model tracking students’ knowledge status, we visually tracked the mastery status of five students in the same knowledge point, and the results are shown in Fig. 4. The darker colors indicate that the corresponding knowledge point has a better mastery status.It can be seen that the pace of mastery of this knowledge among different students is basically the same, which is in line with teaching practice and the law of education.The mastery status of the five students after the 30th time unit has reached more than 0.74, and the mastery status of the corresponding knowledge point has improved.Therefore, the visualization results demonstrate the rationality of the knowledge tracking process.

Figure 4.

Information status changes visualization

The model in this paper is able to capture the correlation between test questions by calculating the distance of the exercises embedded in the space, which is reflected in the knowledge and difficulty involved. The 200 exercises from the ASSIST2009 dataset were analyzed by clustering them in the knowledge space, and the results are shown in Figure 5. It can be seen that these exercises are categorized into 12 categories, and the test questions in each category may have the same knowledge points, and the closer the distance between the points represents the closer their difficulty level, and these automatically learned results can be used as a data supplement in the education domain.

Figure 5.

Information status changes visualization

In addition, this paper also analyzes the difficulty of the exercises in the three datasets. The results are shown in Fig. 6, which shows that the difficulty coefficients b of most of the exercises are distributed between 0.4 and 0.8, and the exercises of ASSISTchall are more evenly distributed in each difficulty interval relative to the other two datasets.

Figure 6.

Question analysis

Recommended Performance Tests for Test Questions

The algorithm in this paper compares the following two recommendation algorithms:

PMF: a collaborative filtering algorithm based on matrix decomposition, which utilizes the inner product of the potential feature vectors of the learner and the potential attribute vectors between knowledge points to obtain the learner’s knowledge mastery.

UserCF: Using students’ answer records to construct student scoring vectors, with correct answers as 1 and incorrect answers as 0, constructing a student-test score matrix, using the test score matrix for user-based collaborative filtering recommendation, calculating the cosine similarity between students, searching for similar students, and predicting the answer scores of target students based on the scores of similar students.

The parameter γ = 0.4 for the proportion of students’ personal factors is set in the experiments, and the results of the experiments in the ASSITments2015 dataset and the Statistics2011 dataset are shown in Tables 2 and 3, respectively.

The proposed algorithm in this paper obtains the student’s knowledge state matrix by tracking the student’s knowledge state for user-based collaborative filtering recommendations. From Table 2 and Table 3, it can be seen that the algorithm in this paper compares with the other two recommendation algorithms precision, r ecall and f1 values are the highest, with F1 values of 0.397 and 0.407, respectively, indicating that this paper’s algorithm outperforms other algorithms in terms of recommendation.

The contrast effect in the assistments2015 data set

PMF UserCF This Model
precision 0.862 0.888 0.923
recall 0.195 0.209 0.213
f1 0.354 0.374 0.397

The contrast effect in the statics2011 data set

PMF UserCF This Model
precision 0.807 0.815 0.829
recall 0.24 0.248 0.253
f1 0.395 0.404 0.407

The parameter γ is used to regulate the proportion of the students’ personal knowledge level and the group common knowledge level influencing the recommendation results. The value range of γ is [0,1], when γ is larger, it means that the score vector of the predicted exercises is closer to the personal knowledge state of the students, and the recommendation result will be more in favor of the students themselves; when γ is smaller, it means that the score vector of the predicted exercises is closer to the knowledge state of the group of students who are similar to the target students, and the recommendation result will be more in consideration of the knowledge level of the students who are similar to the target students.

Figure 7 shows the variation of parameter f1 value with respect to γ when this paper’s algorithm applies different γ parameter values for recommendation under ASSITments2015 dataset. Figure 8 shows the variation relationship when in the Statics2011 dataset.

Figure 7.

Changes in the 1 value of the assitments2015 data set

Figure 8.

Changes in the 1 value of the statics2011 data set

From the experimental results, it can be seen that the value of γ has a significant effect on the recommendation effect, and from the two datasets, it can be seen that f1 value reaches the highest when γ is from 0.3 to 0.5, which indicates that the stability of the algorithm is better when γ is from 0.3 to 0.5, and therefore γ =0.4 is selected as the algorithm parameter for the experiments.

The difficulty range of test questions is set to [λ1, λ2] (λ1 < λ2) based on students’ knowledge when recommending test questions, and in order to recommend test questions of appropriate difficulty to students, the correct response rate of the corresponding students at different levels of difficulty is conducted Experiments on the change in SR

The difficulty range of test questions is 0~1, and 0.2 is set as a difficulty interval. The probability of students’ correct answers is calculated in different difficulty intervals, and the test questions within the difficulty intervals are selected as the recommended test sets, and the probability of students’ correct answers within the difficulty ranges under different datasets is shown in Table 4.

Compare the probability of the correct answer in the data set

[λ1, λ2] 0-0.2 0.2-0.4 0.4-0.6 0.6-0.8 0.8-1.0
ASSITments2015 SR 0.938 0.749 0.586 0.389 0.164
Statics2011 SR 0.925 0.682 0.495 0.327 0.229

As can be seen in Table 4, the test questions answered by students have a higher rate of correctness as the difficulty level decreases. When [λ1, λ2] is [0.4-0.6] students’ probability of correctly answering the test questions on the ASSITments2015 dataset is 0.586 and on the Statistics2011 dataset is 0.495, while when [λ1, λ2] is [0.8-1] students’ probability of correctly answering the test questions on the ASSITments2015 dataset has a probability of correctly answering 0.164 and on the Statics2011 dataset has a probability of correctly answering 0.229, which shows that the method in this paper can recommend test questions with the appropriate range of difficulty for students.

In the experiments, the model of this paper was applied to predict the probability of scoring on test questions for all students in the ASSITments2015 and Statics2011 datasets. For each 40 students, the average score probability obtained was applied to represent the score probability of one student, and then the score probability of each question for each student was averaged and compared with the average score probability of that student in the test set, and the difference between the average score probability before and after was applied to reflect the change in the knowledge level of that student. The difficulty range was divided into five stages, where ΔB represents the amount of change in the student’s knowledge level, and the average change in knowledge level before and after the recommended questions is shown in Table 5, respectively.

Average level of knowledge changes

[λ1, λ2] 0.1-0.3 0.3-0.5 0.5-0.7 0.7-0.9 0.9-1.0
ASSITments2015 ΔB 0.0172 0.0192 0.0263 0.0168 0.0162
Statics2011 ΔB 0.0114 0.0138 0.0196 0.0093 0.0091

As can be seen from the table, after the test-question recommendation exercise, test questions of different difficulty ranges resulted in an increase in the students’ knowledge level. In the two datasets used for the experiment, the students’ knowledge growth level was the greatest for the test questions recommended to them in the [0.5,0.7] range of difficulty. Therefore, test questions with recommended difficulty ranges in the [0.5, 0.7] interval can be prioritized for the student populations in both datasets.

Conclusion

In this paper, a personalized educational resource recommendation frequency domain based on knowledge tracking is designed to improve the efficiency of teaching and examination in higher education. Its knowledge tracking performance and test question recommendation performance are examined in the dataset, and it is found that the model proposed in this paper achieves better results on the three datasets, with the ACC values of 0.823, 0.824, and 0.818, and the AUC values of 0.827, 0.816, and 0.821, respectively. Both metrics are the highest among all models to achieve the best results. The proposed algorithm in this paper obtains the knowledge state matrix of the student by tracking the student’s knowledge state for user-based collaborative filtering recommendation, which is comparable to the other two recommendation algorithms precisi on, recall and f1 have the highest values, with F1 values of 0.397 vs. 0.407, respectively, and the recommendation is significantly better than other algorithms. Together, the personalized education recommendation method based on knowledge tracking performs well and meets the design expectations. Additionally, the personalized online education resources constructed can meet the needs of self-study exams.

Funding:

1) This research was supported by the 2023 Hainan Higher Education Science Planning Research Project “Research and Practice of Online Course Quality Evaluation Method Based on Learner’s Perspective” (No.: QJY20231068).

2) Education and Teaching Reform Project of Hainan Province in 2024: “Study on Influencing Factors and Countermeasures of College Students’ Satisfaction with Online Course Resources” (No.: Hnjg2024-165).

3) Hainan Natural Science Foundation Project “Topic Emotion Classification of Online Comments Based on Topic Modeling and Deep Semantic Analysis “(620QN282).

4) Key Project of Education and Teaching Reform in Hainan Province: “Construction of Educational Big Data Fusion Analysis Model and Its Application in College Education Evaluation “(Hnjg2021ZD-48).

Język:
Angielski
Częstotliwość wydawania:
1 razy w roku
Dziedziny czasopisma:
Nauki biologiczne, Nauki biologiczne, inne, Matematyka, Matematyka stosowana, Matematyka ogólna, Fizyka, Fizyka, inne