Open Access

The Construction and Application of Rural Teachers’ Educational Research Literacy Indicator System in the Context of Deep Learning

 and   
Sep 26, 2025

Cite
Download Cover

Introduction

The key to the implementation of the rural revitalization strategy lies in the revitalization of rural education. Primary and secondary school teachers are an important group in the development and reform of rural basic education, and their roles are gradually changing to “expert” and “research-oriented” teachers. In this transformation, the educational research quality of rural teachers is especially critical [1-3]. Rural teachers are responsible for promoting the revitalization of rural education and the steady development of rural revitalization, and they are also an important force in promoting the balanced and comprehensive development of China’s basic education under the goal of building a moderately prosperous society in all aspects [4-6]. It can be seen that having good educational research literacy also becomes an important factor in promoting the professional development of rural teachers and moving towards a high-quality, specialized and innovative teaching force [7-8].

In recent years, along with the call of scientific research to “develop and strengthen schools”, most of the rural primary and secondary schools in China have increasingly changed the concept of school governance to scientific research as the first opportunity for development, and have gained considerable development, especially in the eastern part of the developed regions, some rural primary and secondary schools and their teachers have strengthened the awareness of educational research and have achieved certain results [9-10]. However, in most provinces and regions of China, there are still a large proportion of rural school teachers who have weak awareness, cognitive bias, lack of passion for action, and insufficient practical ability for educational research [11-12]. Especially in the process of specific scientific research practice, most rural teachers have a deviation of the understanding of the value of scientific research and path deviation phenomenon, do not distinguish the relationship between educational scientific research and classroom teaching, will be more focused on the subject of classroom teaching, lack of awareness of scientific research or scientific research as a task to complete the task of the superiors and perfunctory, and mechanically imitating the model and methodology of educational research in urban schools in the action. In their actions, they mechanically imitate the model and methods of scientific research in urban schools [13-16]. The old and static scientific research concepts and models have, to a certain extent, become the obstacles that prevent rural primary and secondary schools from obtaining full, effective and high-quality development in educational research for a long period of time [17-18].

In a series of strategic blueprints for development, the educational development of rural primary and secondary schools in China has met with the most suitable opportunity for development, and in this opportunity, educators in rural primary and secondary schools are also urged to meet the new mission, follow the new requirements, hold the new direction, and explore a suitable educational scientific research on the development of local rural areas [19-20]. It can be seen that having a good education research literacy is the basic prerequisite for rural teachers in China to conduct education scientific research and explore the laws of education and teaching in rural areas, and it is an important element in realizing a high-quality and high-quality teaching force [21-22].

Rural teachers may differ from urban teachers in terms of scientific research literacy due to various reasons. The construction of the evaluation index system of rural teachers’ educational research literacy, which will provide a basis for the objective and comprehensive assessment of rural teachers’ educational research literacy, provide a path for the effective enhancement of rural teachers’ educational research literacy, and also help to improve the quality of rural education and teaching and promote the development of rural basic education [23-24].

In this paper, we design the evaluation system of rural teachers’ educational research quality based on principal component analysis algorithm. Firstly, the principal component analysis step is modeled to complete the homogenization and standardization of the original data. Then for the defects that BP neural network converges slowly, is very easy to fall into local extreme points, and is extremely sensitive to parameters such as the initial weights of the network, its own learning rate and momentum, PSO algorithm is introduced to optimize the weights and thresholds of the neural network. The group intelligence generated by the cooperation and competition between particles in the population guides the optimization search and improves the performance of the BP neural network. The indicators are selected to construct the evaluation system of rural teachers’ educational research literacy and calculate the weights, experiments are designed to verify the performance of the algorithm, and finally, combined with the examples, the PSO-BP algorithm is utilized to conduct a comprehensive evaluation of rural teachers’ educational research literacy.

Algorithm for evaluating rural teachers’ educational research literacy
Principal Component Analysis
Principles of principal component analysis method

Principal component analysis utilizes the idea of dimensionality reduction, the analysis process is to transform multiple variables into a smaller number of composite variables (principal components), the transformed principal components are not related to each other, are the form of linear combinations of the original variables, so that a large amount of information can be demonstrated through the form of linear combinations, and this information will not be repeated between the variables used to describe the object of the study assumed that there are Z1, Z2, ⋯, Zp A total of P variables, Z=(Z1,Z2,,Zp)t$$Z = {\left( {{Z_1},{Z_2}, \cdots ,{Z_p}} \right)^t}$$ is a P dimensional random variable, Z is made up of hypothesized P variables. Z random vector is assumed, Σ is a covariance matrix of Z and μ is the mean of Z [25]. A linear combination of the original variables is considered and the random vector Z is treated as a linear variation: { F1=μ11Z1+μ12Z2++μ1pZp F2=μ21Z1+μ22Z2++μ2pZp Fn=μml1Z1+μm2Z2++μnpZp$$\left\{ {\begin{array}{*{20}{c}} {{F_1} = {\mu_{11}}{Z_1} + {\mu_{12}}{Z_2} + \ldots + {\mu_{1p}}{Z_p}} \\ {{F_2} = {\mu_{21}}{Z_1} + {\mu_{22}}{Z_2} + \ldots + {\mu_{2p}}{Z_p}} \\ { \cdots \cdots } \\ {{F_n} = {\mu_{ml1}}{Z_1} + {\mu_{m2}}{Z_2} + \ldots + {\mu_{np}}{Z_p}} \end{array}} \right.$$

Linear combination F1,F2,,Fm(mp)$${F_1},{F_2}, \cdots ,{F_m}\left( {m \leq p} \right)$$ is uncorrelated with the principal component, and the largest variance in linear combination Z1, Z2, ⋯, Zp is F1, in linear combinations uncorrelated with F1, F2 is the largest variance,..., and in linear combinations unrelated to FlF2, ⋯, Fm; Fm is the largest variance.

Basic steps of principal component analysis

In the use of principal component analysis, the analysis process generally has the following six steps:

Pre-processing of the original variable data for normalization, in the process of statistical analysis, because the original data itself in the nature of certain differences will affect the statistical analysis, the purpose of pre-processing is to eliminate the impact; standardization of the original variable data, after processing the original variable data in the number of levels and the differences in the quantitative outline of the data will be eliminated, and then get the analytical convenience of the Standardized matrix.

R is the covariance matrix, which is based on the standardized data matrix. The larger the value of covariance matrix R, the stronger the need for principal component analysis of the data, and the covariance matrix shows the closeness of the correlation between the data after standardization of the data. In the covariance matrix, Rij=(i,j=1,2,,p)$${R_{ij}} = \left( {i,j = 1,2, \ldots ,p} \right)$$ represents the correlation coefficient between the original data Zi and Zj. Since R is a real symmetric matrix (i.e., Rij = Rji), only the lower or upper triangular elements are used in the calculation of R. The corresponding formula is shown below: Rij=k1n(ZkjZi)(ZkjZj)k=1n(ZkjZj)2(ZkjZj)2$${R_{ij}} = \frac{{\sum\limits_{k1}^n {\left( {{Z_{kj}} - {Z_i}} \right)} \left( {{Z_{kj}} - {Z_j}} \right)}}{{\sqrt {\sum\limits_{k = 1}^n {{{\left( {{Z_{kj}} - {Z_j}} \right)}^2}} {{\left( {{Z_{kj}} - {Z_j}} \right)}^2}} }}$$

To carry out the determination of the number of principal components, the covariance matrix needs to be calculated to find its cumulative variance contribution, principal component contribution and eigenvalues. The eigenequation |λER|=0$$\left| {\lambda E - R} \right| = 0$$ is solved and the eigenvalues are λi(i=1,2,,P)$${\lambda_i}\left( {i = 1,2, \ldots ,P} \right)$$. The eigenvalues of the covariance matrix R are ranked from largest to smallest λ1λ2 ≥ … ≥ λi ≥ 0 and since R the matrix is a positive definite matrix by nature, the eigenvalues R are all positive. The variance contribution of each principal component is R the eigenvalue, and the magnitude of the influence of the principal component is related to the magnitude of the eigenvalue. The variance contribution of principal component Fi is W1=λj/j=1pλj$${W_1} = {{{\lambda_j}} \bigg/ {\sum\limits_{j = 1}^p {{\lambda_j}} }}$$ and the cumulative variance contribution is j=1nλj/j=1pλj$${\sum\limits_{j = 1}^n {{\lambda_j}} \bigg/ {\sum\limits_{j = 1}^p {{\lambda_j}} }}$$.

And then the selection of principal components, to follow the principle that in the eigenvalue is greater than 1 on the basis of its cumulative contribution rate to reach 80%, the eigenvalue λ1, λ2, ⋯, λn in the corresponding 1,2,,m(mp)$$1,2, \cdots ,m\left( {m \leq p} \right)$$, the number of the final selection of the principal components is m (m is an integer).

The scores of the principal components are calculated, and the initial factor loading matrix is constructed to interpret the principal components. The correlation coefficient Rij=(Fi,Zj)$${R_{ij}} = \left( {{F_i},{Z_j}} \right)$$ between the principal component Fi and the original index Zj is the factor loading, which can be used to better explain the economic significance of the principal components, indicating the degree of correlation between each variable and the principal component. Assuming that Uij is the eigenvector corresponding to each eigenvalue, there are: Uij=Rij/λi$${U_{ij}} = {{{R_{ij}}} \bigg/ {\sqrt {{\lambda_i}} }}$$

An expression for each principal component Fi score can then be obtained: { F1=U11Z1+U12Z2++U1pZp F2=U21Z1+U22Z2++U2pZp Fm=Um1Z1+Um2Z2++UmpZp(i=1,2,,m;j=1,2,,p)$$\left\{ {\begin{array}{*{20}{l}} {{F_1} = {U_{11}}{Z_1} + {U_{12}}{Z_2} + \ldots + {U_{1p}}{Z_p}} \\ {{F_2} = {U_{21}}{Z_1} + {U_{22}}{Z_2} + \ldots + {U_{2p}}{Z_p}} \\ { \cdots \cdots \cdots \cdots } \\ {{F_m} = {U_{m1}}{Z_1} + {U_{m2}}{Z_2} + \ldots + {U_{mp}}{Z_p}} \end{array}\left( {i = 1,2, \ldots ,m;j = 1,2, \ldots ,p} \right)} \right.$$

ZScoren is the teacher composite score function, which is applied to calculate the composite value of educational research literacy of rural teachers and the resulting results are ranked in descending order. The principal components composite model was calculated by taking the proportion of the eigenvalues of each principal component in the sum of the eigenvalues of the extracted principal components that each principal component occupies in it as weights. The resulting model is shown below: ZScoren=M1F1+M2F2++MiFi$$ZScor{e_n} = {M_1}{F_1} + {M_2}{F_2} + \cdots + {M_i}{F_i}$$

(i = 1, 2, ⋯, m; n is the number of samples, M; the principal component F; the proportion of the corresponding eigenvalues to the sum of the total eigenvalues of the principal components).

Principal component analysis of the evaluation system

Convergent processing of raw data

The normalization process is carried out in accordance with the following formula: Yj=1/|YiYi|(i=1,2,,m)$${Y'_j} = {1 \bigg/ {\left| {{Y_i} - {Y_i}} \right|}}\left( {i = 1,2, \ldots ,m} \right)$$

Where, Yj$${Y'_j}$$ is the data after the normalization process, Yi is the original formula indicator data, and Yi¯$$\overline {{Y_i}}$$ is the Yi average value.

The standardization of the original data

The index value selected in this paper is not comparable because the unit and size of each value are relatively large when the relevant data are selected, so it is difficult to compare, and the original data that are difficult to compare are standardized to achieve the elimination of the differences between different variables caused by the difference in the scale. Zi=(YiY¯i)/σYi$${Z_i} = \left( {Y_i^\prime - {{\bar Y}_i}} \right)/{\sigma _{{Y_i}}}$$ Yi¯=EYi=j=1nYij/n$$\overline {{Y_i}^\prime } = E{Y_i}^\prime = \sum\limits_{j = 1}^n {{{{{Y_{ij}^\prime}}} \bigg/ n}}$$ σYi=DYi=j=1n(YijYi¯)2/(n1)(i=1,2,,mj=1,2,,m)$${\sigma_{{Y_i}}} = \sqrt {D{Y_i}^\prime } = {{\sum\limits_{j = 1}^n {{{\left( {{Y_{ij}}^\prime - \overline {{Y_i}^\prime } } \right)}^2}} } \bigg/ {\left( {n - 1} \right)}}\left( {i = 1,2, \ldots ,mj = 1,2, \ldots ,m} \right)$$

In the above equation, Z1 indicates that the data have been standardized, Y¯1$${\bar Y_1}$$ and σYi$${\sigma_{{Y_i}}}$$ are the mean and variance of Yi$${{Y_i}^\prime}$$, respectively. In this paper, the data processing, the use of SPSS software to automatically standardize the data, according to the relevant theoretical knowledge of statistics can be known, the data after the standardized treatment to comply with the form of standard normal distribution, the data after the standardized treatment of the later statistics and analysis of the work has a great help.

BP Neural Network
Artificial Neural Networks

Artificial neural networks are, after all, a mathematical model of information processing that applies a structure similar to the synaptic connections of the brain, a neural network or neural-like network consisting of a large number of artificial neurons interconnected in a similar way to natural nerve cells. A neural network is an arithmetic model consisting of a large number of nodes (or neurons) and their interconnections, which emphasizes synergy between a large number of neurons and problem solving through learning. Each node represents a specific output function called the excitation function. The weights are the weighted values of the connection signals between two nodes, which is equivalent to the memory of an artificial neural network. The different values of weights and excitation functions determine the different outputs of the network.

A neuron is a nonlinear information processing element with multiple inputs and single outputs, which can be abstracted as a simple mathematical model based on its properties and functions [26].

Its input-output relationship can be described as: { Xj=i=1WijXiθj yj=f(Xj)$$\left\{ {\begin{array}{*{20}{l}} {{X_j} = \sum\limits_{i = 1} {{W_{ij}}} {X_i} - \theta j} \\ {{y_j} = f\left( {{X_j}} \right)} \end{array}} \right.$$

Xi(i=1,2,,n)$${X_i}\left( {i = 1,2, \cdots ,n} \right)$$ is the input signal to the neuron; each input to the neuron has a weighting coefficient Wij(i=1,2,,n)$${W_{ij}}\left( {i = 1,2, \cdots ,n} \right)$$ called the weight value, which represents the connection weights from neuron i to neuron j. Its positive and negative values simulate synaptic excitation and inhibition in biological neurons, and its magnitude represents the different connection strengths of the synapses; and yj is the output of the jth neuron; f()$$f\left( \cdot \right)$$ is the excitation function, which determines in which way the j neuron outputs when co-stimulation by the input neuron Xi(i=1,2,,n)$${X_i}\left( {i = 1,2, \cdots ,n} \right)$$ reaches a threshold value; θj denotes the threshold value of the neuron.

BP Neural Network Fundamentals

The basic principle of the BP network model to process information is: input signal Ii in the input layer, through the intermediate nodes (hidden layer points) act on the output node, after a nonlinear transformation, to produce the output signal Ok, the network training samples include the input vector I and the output expected value T, there is a deviation between the network output value Ok and the expected value T, through the adjustment of the linkage weight wji between the input node and the hidden layer node and the linkage weight wjk between the hidden layer node and the output node and the threshold value, so that the error decreases along the gradient direction. and the linkage weight 8 between the output node and the output node as well as the threshold value, so that the error decreases along the gradient direction, and after repeated learning and training, the network parameters (weights and thresholds) corresponding to the minimum error are determined, and then the training is terminated. The trained neural network can be input to similar samples, train and learn on its own and output information that minimizes the error after nonlinear transformation.

The adjustment formula for the weights thresholds is: wkj(t+1)=Δwkj+wkj(t)=αEwkj+wkj(t)=αδkHj+wkj(t)$${w_{kj}}(t + 1) = \Delta {w_{kj}} + {w_{kj}}(t) = - \alpha \frac{{\partial E}}{{{w_{kj}}}} + {w_{kj}}(t) = \alpha {\delta_k}{H_j} + {w_{kj}}(t)$$ wji(t+1)=Δwji+wji(t)=αEwji+wji(t)=ασjIi+wji(t)$${w_{ji}}(t + 1) = \Delta {w_{ji}} + {w_{ji}}(t) = - \alpha \frac{{\partial E}}{{{w_{ji}}}} + {w_{ji}}(t) = \alpha {\sigma_j}{I_i} + {w_{ji}}(t)$$ θk(t+1)=Δθk+θk(t)=βEθk+θk(t)=βδk+θk(t)$${\theta_k}(t + 1) = \Delta {\theta_k} + {\theta_k}(t) = - \beta \frac{{\partial E}}{{{\theta_k}}} + {\theta_k}(t) = \beta {\delta_k} + {\theta_k}(t)$$ θj(t+1)=Δθj+θj(t)=βEθj+θj(t)=βδj+θj(t)$${\theta_j}(t + 1) = \Delta {\theta_j} + {\theta_j}(t) = - \beta \frac{{\partial E}}{{{\theta_j}}} + {\theta_j}(t) = \beta {\delta_j} + {\theta_j}(t)$$

where Hj is the output of the hidden layer node, Ii is the input signal from the input node i, wkj(t) and wkj(t + 1) are the connection weights between the hidden layer node j and the output layer node k during the two training periods before and after, wji(t) and wji(t + 1) are the connection weights between the input node i and the hidden layer node j during the two training periods before and after, θk and θj are the thresholds at the output node k and the hidden layer node j, respectively; α and β are the learning parameters, which are generally taken as 0.1~0.9; δk and σj are the error signals of output layer node k and hidden layer node j, respectively, which are computed as Eqs: δk=αEIk=(TkOk)Ok(1Ok)$${\delta_k} = - \frac{{\alpha E}}{{\partial {I_k}}} = \left( {{T_k} - {O_k}} \right){O_k}\left( {1 - {O_k}} \right)$$ σj=αEIj=δkwkjHj(1Hj)$${\sigma_j} = - \frac{{\alpha E}}{{\partial {I_j}}} = \sum {{\delta_k}} {w_{kj}}{H_j}\left( {1 - {H_j}} \right)$$

Tk is the target output value of the sample at the output node k; Ok and Hj are the actual output values of the sample at the output node k and the hidden layer node j of the network, respectively, where the hidden layer output is computed as: Hj=f[i=1mwjiIi+θj]$${H_j} = f\left[ {\sum\limits_{i = 1}^m {{w_{ji}}} {I_i} + {\theta_j}} \right]$$

M is the number of input nodes and f is the type S activation function: f(x)=11+ex$$f(x) = \frac{1}{{1 + {e^{ - x}}}}$$

The output layer outputs are summed using linear weighting: Ok=j=1swkjHj+θk$${O_k} = \sum\limits_{j = 1}^s {{w_{kj}}} {H_j} + {\theta_k}$$

Since the desired output T and the actual output O do not coincide, an error arises, which is usually expressed in terms of variance, calculated as: E=12t=1q(TtOt)2$$E = \frac{1}{2}\sum\limits_{t = 1}^q {{{\left( {{T_t} - {O_t}} \right)}^2}}$$

Particle Swarm Algorithm (PSO)

Particle Swarm Optimization (PSO) algorithm, with the continuous application of particle swarm algorithm in recent years, has gradually become a new optimization algorithm.

PSO algorithm is a heuristic optimization calculation method, and its biggest advantages:

The algorithm is easy to understand;

No special requirements on the continuity of the definition of the optimization problem;

Fewer parameters, no need to adjust complex parameters;

The algorithm is simple and the speed of optimization is faster;

Possessing a smaller evolutionary population relative to genetic algorithms and the like;

Fast convergence;

The system is extremely robust and will not affect the solution of the whole problem because of individual differences.

PSO algorithms also have disadvantages:

It is easy to fall into the local extreme points, thus not getting the result.

Although PSO converges faster, the results are not accurate without the cooperation of other algorithms.

PSO has the ability to search globally, but there is no guarantee that it can converge on the global optimum.

Considering each bird in the search space as a potential optimal solution to the problem, we call each bird in the space a “particle”. Each particle has a fitness value determined by the optimization function, and a speed that determines the direction and distance they fly, and the particles follow the optimal particles to find the optimal particles in the search space. PSO will set the initial position and speed for a group of random particles, and then iteratively search for the optimal solution, and during each iteration, the particles will track the two extremes to update their positions. During each iteration, the particle will track two poles to update its position, the first one is the individual pole, which is the optimal solution found by itself, and the other one is the global pole, which is the optimal value of the whole population, and the other one is the local pole, which is the local pole among all the neighbors, instead of using the whole population, we can use only a part of it as the neighbors of the particles.

The particle will constantly update its velocity and new position according to the following equation: v[ ] = v[ ]+c1*rand( )*(pbest[ ]present [ ]) +c2*rand( )*(gesest[ ]present[ ])$$\begin{array}{rcl} v\left[ {\,} \right] &=& v\left[ {\,} \right] + c1^*rand\left( {\,} \right)^*\left( {pbest\left[ {\,} \right] - present\left[ {\,} \right]} \right) \\ &&+ c2^*rand\left( {\,} \right)^*\left( {gesest\left[ {\,} \right] - present\left[ {\,} \right]} \right) \\ \end{array}$$ present[ ]=present[ ]+v[ ]$$present\left[ {\,} \right] = present\left[ {\,} \right] + v\left[ {\,} \right]$$

v[ ]$$v\left[ {\,} \right]$$ is the velocity of the particle, present[ ]$$present\left[ {\,} \right]$$ is the current position of the particle, pbest[ ]$$pbest\left[ {\,} \right]$$ and gbest[ ] $$gbest\left[ {\,} \right]$$ denote the individual and global extremes, respectively, rand( )$$rand\left( {\,} \right)$$ is a random number between (0,1)$$\left( {0,1} \right)$$, and c1 and c2 are the learning factors, usually c1 = c2 = 2.

The main operation flow of the PSO algorithm is as follows:

Set the initial random position and velocity of the particle swarm;

Calculate the adaptation value of each particle in the swarm;

Compare the adaptation value calculated in the second step with the optimal adaptation value, and select it as the best current position if it is better.

Compare the computed adaptation value with the globally optimal adaptation value, and if it is better, select it as the current global best position;

Evolve the position and velocity of the particle according to the two iterative formulas given above;

If it fails to find a sufficiently superior adaptation value or the preset maximum number of iterations has been reached, then return to step (2), otherwise perform step (7);

Output gbesr[ ]$$gbesr\left[ {\,} \right]$$.

PSO-BP model
PSO-BP basic idea

The initial weights and thresholds of the BP neural network are randomly assigned, and the determination of the final weights relies heavily on the selection of the initial weights, but the BP neural network is slow to converge, is very easy to fall into the local extreme point, and is extremely sensitive to the network’s initial weights, its own learning rate, and momentum and other parameters, which need to be constantly trained in order to be gradually fixed, and over-training leads to the phenomenon of overfitting, thus affect the generalization ability of the network.

PSO, as a population theory optimization tool, combines PSO with BP neural network and uses PSO algorithm to optimize the weights and thresholds of the neural network, which can improve the performance of the BP neural network. PSO algorithm guides the optimization search through the group intelligence generated by the cooperation and competition among the particles in the population. The optimized algorithm has fast convergence speed, high robustness, strong global search ability, and the search efficiency is improved.

The process of using PSO algorithm to adjust the weights of BP neural network is: first of all, the PSO algorithm is used to replace the gradient descent method in the BP neural network to repeatedly optimize the combination of weights and parameters of the BP neural network model, until the fitness of the solution is no longer reduced, based on this, then use the BP neural network to further accurately optimize the parameters of the network obtained above until the search for the optimal parameters of the network, which can be obtained. The precise optimal parameter combination.

PSO-BP modeling

The PSO-BP model is a three-layer structure, which is the input layer, the hidden layer and the output layer. The number of neurons in the output layer is set to 1. In this paper, according to the empirical formula, it is concluded that the number of nodes in the hidden layer can be initially determined as 5-13 layers, and then the number of neurons in the hidden layer is finally determined according to the method of trial and error [27].

The specific design steps to realize the combination of particle swarm and BP neural network are as follows:

Determine the number of population particles

Based on the research experience of experts and scholars, when using particle swarm optimization algorithm to solve general function optimization problems, the number of particles is generally taken as 10, so the number of particles in the PSO-BP neural network model in this paper is 10.

Determine the inertia particles w

The inertia factor w plays a big role in the convergence of the particle swarm algorithm, and w can be taken as a random number in the interval 0-1. In this paper, we take w the maximum value of 0.90 and the minimum value of 0.30.

Determine the value of the dimension of individual particles of the particle swarm

According to the previous research experience, the dimension of the particle swarm single particle satisfies as shown in Equation (23): n=Y1*(A+1)+(Y1+Y2)$$n = {Y_1}^*\left( {A + 1} \right) + \left( {{Y_1} + {Y_2}} \right)$$

Where Y1 is the number of nodes in the single implicit layer, Y2 is the number of nodes in the output, A is the number of nodes in the input layer, according to the research design of the college students’ physical health evaluation index system, there are a total of 10 indexes, and the output is a comprehensive evaluation score, then A = 10, Y1 = 5 − 13, and Y2 = 1, which is to get the dimensions of the individual particles of each particle swarm of each particle is n = 61 − 157.

Determine the initial position of each particle of the particle swarm

Since the dimension of each particle is 61-157, the position matrix of each particle has 61-157 elements, and according to experience, the value of each element can be between [1.5,1.5]$$\left[ { - 1.5,1.5} \right]$$, that is, the lower limit of the position is -1.5 and the upper limit is 1.5.

Determine the velocity of individual particles of the particle swarm

Since the dimension of the particle is 73, it is assumed that the velocity vector of the particle is V(V1,V2,,V73)$$V\left( {{V_1},{V_2}, \cdots ,{V_{73}}} \right)$$, and the velocity value of each dimension is initialized to one-tenth of the position value, then the velocity vector formula is shown in Equation (24): Vi=(V1,V2,,V73),0<i<73,0ViVmax$${V_i} = \left( {{V_1},{V_2}, \cdots ,{V_{73}}} \right),0 < i < 73,0 \leq {V_i} \leq {V_{\max }}$$

Determine the velocity threshold of individual particles of the particle swarm Vmax

Since the number of particles selected in this paper is 10, the Vmax range is set to [10,10]$$\left[ { - 10,10} \right]$$, i.e., Vmax = 20.

Determination of acceleration constants c1 and c2

The acceleration constants are the weights that adjust the role played by their own experience and social experience in their movement. For routine problems, c1 = c2 = 2.0 is generally taken.

Selection of termination conditions

The maximum number of iterations in PSO optimized BP neural network is generally taken as 100-4000. In this paper, the optimal solution of the network is obtained after several experiments are carried out, and it is concluded that the network reaches a stable value during the solution process, at which time the maximum number of iterations is 500, so when the number of iterations reaches the maximum number of times 500, the program is terminated.

Design the fitness function

The definition of the mean square error is shown in equation (25): σ2=E=12rj(YrjDij)2$${\sigma^2} = E = \frac{1}{2}\sum\limits_r {\sum\limits_j {{{\left( {{Y_{rj}} - {D_{ij}}} \right)}^2}} }$$

Where Yrj is the desired output in the training dataset and Dij is the actual output of the node during training.

The PSO-BP modeling process is as follows:

Establish the number of neurons in the three-layer structure of the BP neural network

Initialize the particle swarm: initialize the data such as individual extreme value and global optimal value of each particle.

Determine the fitness function: the particle search performance index is represented by the minimum mean square error MSE of the neural network.

Calculate the fitness of each particle

First input a particle, you can calculate the output value of the network according to the pre-feedback method of BP neural network, and then calculate its error; use the same method to calculate the error of all samples, and then calculate the mean square error of all samples, that is, the fitness of the particle.

Return to (1) and continue to input other particles, thus calculating the fitness of all particles.

Update the individual extremes and the global optimum according to the fitness of each particle

If present < Pi, Pi = present, Pi = Xi, otherwise Pi remains unchanged; if present < Pg, Pg = present, Pg = Xi, otherwise Pg remains unchanged. Where present is the fitness of the current particle, Pi is the individual extreme value of the particle and Pg is the global optimum.

Update the velocity and position of the particles and generate a new population

Consider the speed

If Vid > Vmax, then Vid = Vmax; if Vid < −Vmax, then Vid = −Vmax, otherwise Vid remains unchanged.

Consider the position

If Xid > Xmax, then Xid = Xmax; if Xid < Xmin, then Xid = Xmin, otherwise Xid remains unchanged. Where, Vmax, Xmax, Xmin are initialized values.

Calculate the error of the algorithm: E=i=1kf(Pg(i))k$$E = \frac{{\sum\limits_{i = 1}^k f \left( {P_g^{(i)}} \right)}}{k}$$

In Equation (26), k is the current iteration number, and f(Pk(i))$$f\left( {P_k^{(i)}} \right)$$ is the fitness of the global optimum of the ird iteration.

Determine whether to meet the conditions for iteration stopping

If the error of the algorithm reaches the preset accuracy or the maximum number of iterations, the algorithm converges, and the weights and thresholds of each dimension in the global optimal value Pg of the last iteration are the optimal solution; if the number of iterations does not reach the maximum, the algorithm returns to step (4) and continues to iterate, otherwise, the algorithm is terminated.

The flowchart of PSO-BP model is shown in Fig. 1.

Figure 1.

Flow chart of PSO-BP model

Evaluation system construction and application
The construction of rural teachers’ educational research literacy index system
Questionnaire design and data collection

The questionnaire research was conducted by forwarding QR codes or links to rural primary and secondary school teachers on staff in a certain province to distribute questionnaires respectively, covering 14 cities. The questionnaire consists of three parts, the first part is the basic personal information of rural primary and secondary school teachers, including gender, age, teaching age, education, teaching subjects, title, the region where the school is located and the basic situation of scientific research, etc. The second part is about the basic situation of scientific research carried out by the school, including the opportunities for teachers to participate in scientific research training, and the form and frequency of scientific research activities carried out by the school. The third part consisted of a series of questions prepared in the form of a five-point Likert scale.

A total of 1850 valid questionnaires were returned. The average age of the research sample was 38.67 years old, with an average of 15.24 years of teaching experience, and the subjects taught covered fourteen subjects, including language and mathematics, currently carried out in primary and secondary schools, with the survey covering primary and secondary schools of all levels and types in villages, townships and counties.

Extraction of common factors to determine evaluation indicators

In this study, the Cronbach’s alpha coefficient and KMO value of the data analyzed using SPSS 26.0 software were 0.962 and 0.966 respectively, indicating that the data were well suited for factor analysis, which was then standardized on the data from the observation points and then factor analysis was done. The total variance explained is shown in Table 1.

Total variance interpretation

Rotational load squared Constituent
1 2 3
Eigenvalues 9.034 6.117 5.833
The percentage of variance 34.681 23.592 22.059
cumulative 34.681 58.273 80.332

Three factors with eigenvalues greater than 1 were extracted and the cumulative contribution of the three factors reached 80.332%.

The factor loadings of the 26 original variables on the three factors were obtained after a quadratic maximum rotation. The rotated component matrix is shown in Table 2.

The component matrix of the rotation

Index The number of component factors of the rescale
1 2 3
Literature review capability 0.871 0.238 0.163
To master the specific steps of education 0.864 0.235 0.151
Inductive sum 0.86 0.227 0.088
Ability to apply and promote results 0.861 0.262 0.191
Data analysis ability 0.864 0.255 0.207
Literature retrieval ability 0.858 0.235 0.18
To master the basic theory of education 0.851 0.24 0.111
The planning capacity of the research direction 0.844 0.283 0.239
Methodology of education research 0.837 0.272 0.252
Basic theory of pedagogy psychology 0.794 0.262 0.277
Discipline theory and knowledge 0.763 0.231 0.291
The understanding of scientific research promotion teaching 0.173 0.83 0.331
The understanding of scientific research and promotion of professional development 0.153 0.82 0.383
Understanding the significance of scientific research 0.196 0.804 0.311
Will of scientific research 0.33 0.804 0.23
Interest in research 0.331 0.771 0.232
Will be the will of the research teacher 0.392 0.767 0.138
Ability to find problems 0.473 0.656 0.268
Ability to provide new ideas 0.501 0.635 0.246
Ability to reflect on teaching 0.411 0.622 0.355
Respect research objects 0.211 0.287 0.907
Insist on 0.208 0.292 0.907
Accuracy and authenticity of data 0.218 0.289 0.883
Respect for the right of acting 0.211 0.274 0.883
Protect the privacy of the subjects 0.24 0.262 0.863
Normative reference 0.246 0.288 0.836

The results showed that Factor 1 governed 11 primitive variables such as literature review ability, Factor 2 governed 9 primitive variables such as awareness of research for teaching and learning, and Factor 3 governed 6 primitive variables such as respect for the subject of the study. Accordingly, the three factors were named, research ability factor, research awareness factor and research ethics factor.

Based on the results of the above data analysis, it was determined that the evaluation indexes of rural teachers’ scientific research literacy consisted of three dimensions and 26 indexes, and the weight assigned to each index was obtained through calculation, and the results are shown in Table 3. It can be seen that the indicator with the highest weight value is the ability to analyze information, with a weight of 0.092. This was followed by the willingness to become a research teacher with a weight of 0.052.

Comprehensive model score coefficient and index weight value

Dimension Index Linear combination coefficient Model integrated score coefficient Index weight
Scientific ability factor Scientific awareness factor Scientific moral factor
Scientific ability Literature review capability 0.288 0.089 0.068 0.174 0.035
To master the specific steps of education 0.283 0.089 0.062 0.167 0.036
Inductive sum 0.289 0.083 0.032 0.158 0.039
Ability to apply and promote results 0.286 0.107 0.078 0.175 0.040
Data analysis ability 0.288 0.096 0.082 0.174 0.092
Literature retrieval ability 0.284 0.093 0.075 0.173 0.035
To master the basic theory of education 0.283 0.093 0.042 0.161 0.028
The planning capacity of the research direction 0.281 0.108 0.098 0.185 0.038
Methodology of education research 0.274 0.107 0.103 0.183 0.036
Basic theory of pedagogy psychology 0.258 0.102 0.118 0.178 0.037
Discipline theory and knowledge 0.249 0.087 0.118 0.166 0.037
Scientific awareness The understanding of scientific research promotion teaching 0.055 0.336 0.131 0.16 0.044
The understanding of scientific research and promotion of professional development 0.051 0.333 0.155 0.156 0.035
Understanding the significance of scientific research 0.059 0.322 0.127 0.161 0.028
Will of scientific research 0.107 0.318 0.093 0.166 0.036
Interest in research 0.105 0.309 0.089 0.163 0.034
Will be the will of the research teacher 0.125 0.305 0.054 0.164 0.052
Ability to find problems 0.149 0.261 0.108 0.173 0.035
Ability to provide new ideas 0.161 0.252 0.095 0.174 0.031
Ability to reflect on teaching 0.138 0.252 0.141 0.169 0.035
Scientific ethics Respect research objects 0.07 0.112 0.375 0.167 0.037
Insist on 0.062 0.118 0.378 0.165 0.035
Accuracy and authenticity of data 0.067 0.111 0.366 0.167 0.039
Respect for the right of acting 0.065 0.107 0.361 0.163 0.035
Protect the privacy of the subjects 0.081 0.103 0.357 0.162 0.035
Normative reference 0.073 0.117 0.344 0.164 0.036
PSO-BP evaluation model training test

The PSO-BP neural network model is applied to the assessment of the development level of informationization in primary and secondary schools. To ensure the accuracy of the optimization model, the data applied in the PSO-BP neural network assessment model is consistent with the neural network model data. 80% of the 20259 valid data were utilized as the training set and 20% as the testing set. The data were normalized using Mapminmax function in MATLAB.

Firstly, the sample data were used as the population, 26 assessment indicators were used as the input data of the model, and the comprehensive development level calculated by principal component analysis and subjective-objective combined assessment was used as the output data, and experimental training was carried out respectively, which were named as Experiment 1 and Experiment 2 in the experimental process.

After determining the specific values of the parameters that need to be defined in these BP neural networks and their optimization algorithms, the model can be trained. The specific training process and code of the model is implemented in MATLAB, and the relationship between the number of iterations of the algorithm and the model training error when the algorithm optimizes the parameters of the BP neural network model in the process of model training is shown in Fig. 2 and Fig. 3.

Figure 2.

Iteration error experiment 1

Figure 3.

Iteration error experiment 2

From the figure, we can learn that with the increase in the number of iterations, the error of the model also decreases sharply, and at the same time, we can see that in experiments 1 and 2, when the number of iterations reached 21 times and 6 times or more, respectively, the error of the model also reaches the minimum and gradually tends to stabilize, which proves that in the beginning of the training model before we set the maximum number of iterations of the experiment is reasonable.

At the same time, the number of training iterations is less than the 38 and 36 iterations of BP, which proves that the excellent global optimization ability of the optimization algorithm in this paper can accelerate the convergence and improve the network performance.

Evaluation model training and application

The size of the population and the number of iterations have a great impact on the optimization algorithm’s optimization search. When the population size is too small, the algorithm converges faster, however, the possibility of local extremes is high, resulting in difficulty in iterating to the optimal fitness; and when the population size is too large, the complexity of the search for optimization is greatly increased, and the search time of the algorithm is also extended. In addition, too many iterations will reduce the diversity of the population. Therefore, the appropriate population size and iteration number help the algorithm to find the optimal solution. Therefore, this paper compares the performance of BP model, GA-BP model, SSA-BP model and PSO-BP model proposed in this paper in terms of the same and optimal population size and number of iterations, respectively.

Comparison based on same population size and number of iterations

In the experiments, in order to reduce the influence of the initial parameters on the performance of the algorithms, the settings for the same parameters are guaranteed to be the same. Because several algorithmic models for comparison are based on BP neural network, the parameters for BP neural network are set according to the previous section.

After setting the parameters of each model, the preprocessed sample data are input into GA-BP model, SSA-BP model, PSO-BP model and BP model for training and testing respectively. First, the training samples were input into the four models for training, and after the training was completed, the testing stage was entered, and after several iterations, the results of comparing the evaluation values of the samples in the test set of each model with the real values were obtained, as shown in Fig. 4.

Figure 4.

The evaluation results of the test sample were compared with the real value

Considering the characteristics of neural network algorithms and intelligent optimization algorithms that the results of each run are different, the four algorithm models are run 30 times, and the experimental results of the performance evaluation indexes of the different models are averaged to compare the data, as shown in Table 4.

Evaluation index data based on the same population size and iteration number

Index BP SSA-BP FASSA-BP PSO-BP
MRE 0.445% 0.358% 0.158% 0.107%
MSE 8.617% 5.913% 2.398% 1.819%
MAE 25.872% 19.301% 11.793% 10.060%
RMSE 0.2951 0.2462 0.1655 0.1288
R2 0.9834 0.9875 0.9987 0.9991

Comprehensive Figure 4 and Table 4 show that the evaluation accuracy of the PSO-BP algorithm proposed in this paper is significantly higher than the basic SSA-BP algorithm, GA-BP algorithm and traditional BP algorithm, and the value of the correlation coefficient R2 is closer to 1 than that of the other three algorithms, which verifies that the model can be better applied to the problem of evaluating the scientific research literacy of rural teachers in education.

Comparison based on optimal population size and number of iterations

By selecting the Rastrigrin test function to set different population sizes and iteration times, the optimal population size and iteration times that enable the PSO algorithm, SSA algorithm and GA-BP algorithm to reach the optimal fitness value are finally obtained.

In order to further test the superiority of PSO-BP algorithm, the best parameters under the optimal fitness of GA-BP algorithm, SSA-BP algorithm and PSO-BP algorithm, respectively, are selected, and their algorithmic performance is analyzed through experimental comparison. Similarly, the preprocessed sample data are input into GA-BP model, SSA-BP model, PSO-BP model and BP model respectively, and after many iterations, the evaluation value of each model is obtained and compared with the real value, and the results are shown in Figure 5.

Figure 5.

Test results based on optimal population size and iteration number

The GA-BP model, SSA-BP model and PSO-BP model were set to the optimal population size and the number of iterations, respectively, and since the three algorithmic models run for too long in the case of the optimal population size and the number of iterations, the GA-BP model, the SSA-BP model and the PSO-BP model were run for 20 times, and the experimental performance evaluation indexes of each model were calculated as the The average of the result data is used to analyze and compare the performance of these models, as shown in Table 5.

Evaluation index based on optimal population size and iteration number

Index BP SSA-BP GA-BP PSO-BP
MRE 0.4% 0.2765% 0.1565% 0.1386%
MSE 8.4866% 3.9755% 1.9027% 1.0203%
MAE 25.317% 14.638% 12.505% 7.64%
RMSE 0.2964 0.2043 0.1353 0.1015
R2 0.9819 0.9812 0.9906 0.9989

As can be seen from the comparison between the evaluation values of the test set samples and the true values in Figure 5, the PSO-BP evaluation model has a better evaluation effect. Combined with Table 5, it can be seen that although the PSO-BP algorithm model is under the optimal population size and iteration number, the PSO-BP model designed in this paper is significantly higher than the SSA-BP model, the GA-BP model and the traditional BP model from the point of view of evaluation accuracy, which further proves the superiority of the PSO-BP algorithm proposed in this paper.

In order to analyze the PSO-BP model more accurately, the regression analysis and relative error of the model evaluation value and the expected output value of the sample data of the test set of the PSO-BP model are selected and analyzed, as shown in Figure 6 and Table 6, respectively. Among them, Fig. 6 (a) to (d) are the results in different sample cases, respectively.

Figure 6.

Regression analysis of model evaluation value and expected output value

The actual value of the sample is relative to the model evaluation

Test set sample True value Evaluation grade Model evaluation Corresponding grade Relative error Grade accurate
1 56 Fifth level 55.9551 Fifth level -0.0449 Yes
2 54 Fifth level 54.301 Fifth level 0.301 Yes
3 58 Fifth level 58.387 Fifth level 0.387 Yes
4 74 Seventh level 73.8721 Seventh level -0.1279 Yes
5 51 Fifth level 51.079 Fifth level 0.079 Yes
6 52 Fifth level 52.3933 Fifth level 0.3933 Yes
7 71 Seventh level 71.154 Seventh level 0.154 Yes
8 63 Six level 63.4285 Six level 0.4285 Yes
9 56 Fifth level 56.1951 Fifth level 0.1951 Yes
10 68 Six level 67.5647 Six level -0.4353 Yes
11 48 Four level 47.7384 Four level -0.2616 Yes
12 51 Fifth level 51.125 Fifth level 0.125 Yes
13 74 Seventh level 73.9559 Seventh level -0.0441 Yes
14 72 Seventh level 72.0579 Seventh level 0.0579 Yes

From the regression analysis of the actual output value of the network and the expected output value in Fig. 6, it can be seen that the regression coefficient of the PSO-BP model is very close to 1, which indicates that the evaluation of the PSO-BP model is very good.

From the results of the relative error comparison between the model output value and the real value of the test set samples in Table 6, it can be concluded that the PSO-BP evaluation model has a small error, and the results of the grade division also match completely, which also shows that the model’s comprehensive evaluation score of the rural teachers’ education and research literacy is almost the same as that of the actual teacher’s literacy, and it has a certain degree of accuracy and feasibility, and it further proves that this model can be used in the evaluation study of the rural teachers’ education and research literacy. It further proves that this model can be used in rural teachers’ education and research literacy evaluation research.

Evaluation of Rural Teachers’ Educational Research Literacy

Five rural teachers from different districts and subjects were selected and the model was applied to evaluate their educational research literacy of rural teachers. The results are shown in Table 7. The evaluation results of the five teachers outputted by the model are 52.8981, 50.2208, 72.8602, 66.2254 and 73.3937, and the evaluation grades are basically consistent with the actual situation of the teachers. It can be considered that the model in this paper has met the design expectations.

Evaluation result

Teacher number Output Grade
1 52.8981 Fifth level
2 50.2208 Fifth level
3 72.8602 Seventh level
4 66.2254 Six level
5 73.3937 Seventh level
Conclusion

In this paper, in order to study the educational research literacy of rural teachers, an evaluation model of rural teachers’ educational research literacy based on PSO-BP combination algorithm is designed. After checking the performance of the algorithm, it is applied to practical evaluation. The final conclusions are as follows:

In the designed index system, the index with the highest weight value is the ability to analyze information, with a weight of 0.092, followed by the willingness to become a research teacher, with a weight of 0.052.

When the number of iterations reaches 21 times and more than 6 times respectively, the error of the model also reaches the minimum and gradually tends to be stable, which is much lower than the traditional BP algorithm, proving the excellent performance of the model.

The evaluation accuracy of the PSO-BP algorithm proposed in this paper is significantly higher than that of the basic SSA-BP algorithm, FASSA-BP algorithm and traditional BP algorithm, and the value of the correlation coefficient R2 is closer to 1 than that of the other three algorithms.

Language:
English