Research and design of auxiliary teaching system for college students’ Civics and Political Science courses in the context of deep learning
Publicado en línea: 17 mar 2025
Recibido: 10 oct 2024
Aceptado: 31 ene 2025
DOI: https://doi.org/10.2478/amns-2025-0236
Palabras clave
© 2025 Yongchao Yin, published by Sciendo
This work is licensed under the Creative Commons Attribution 4.0 International License.
As the main channel for the cultivation of college students’ "three views" and socialist core values, ideological and political education class is an important part of ideological and political education in colleges and universities [1-2]. The high-speed development of network technology has a profound impact on the learning life, way of thinking and values of college students [3], and their thoughts show a strong trend of independence, equality, openness and personalization [4]. The traditional Civics class in colleges and universities is a package teaching, which is centered on teaching materials and focuses on knowledge instillation by teachers’ lectures, so that students are in the role of passive reception in the teaching process [5]. The communication between teachers of Civics and students is limited to the classroom, and students can only rely on their own efforts to understand and consolidate their knowledge, which seems to be a bit confusing and does not achieve the purpose of ideological and political education. In addition, students’ learning is more utilitarian, resulting in students only focusing on the examination results of the Civics and Political Science course, and can not really achieve knowledge understanding and action transformation [6-8]. Therefore, how to break the limitations of time and space of Civics class lectures, improve the utilization rate of extracurricular time, stimulate students’ interest in learning Civics courses, improve the teaching effect of Civics classes, and cultivate students’ logical thinking ability has become a hot issue in the reform of Civics classes [9-11].
With the development of China’s higher education from elitist education to mass education, teachers of Civics and Political Science courses have less and less time to contact with students outside the classroom, and students rely solely on their own efforts to consolidate and review their knowledge [12-14]. And such a way is more likely to result in insufficient understanding of the Civics and Political Science course, coupled with the fact that students’ learning carries a strong utilitarian mentality, which leads to students learning for the sake of exams in the Civics and Political Science course [15]. Therefore, how to make efficient use of extracurricular time, strengthen students’ tutoring of the Civics course, at the same time, deepen college students’ understanding of the knowledge of the Civics course, and cultivate students’ logical thinking ability has the necessity of [16], and strengthen the effect of the Civics course lectures has become another focus of thinking about the innovation of the Civics course teaching method in colleges and universities [17], and with the help of the current popular course network idea, we should build a network that can be used specifically for the study and communication of the Civics course in universities. With the current popular idea of course networkization, a network platform should be constructed which can be specially used for learning and communication of university Civic and Political courses [18-19].
In this paper, a visual recognition system for students’ evaluation materials and answer results is designed using deep learning computer vision technology.Based on the CAT-SOOP system, a set of auxiliary teaching systems for Civics has been designed using augmented learning.Secondly, by building a network diagram of Civics exercises and combining the concept of cut-set in complex networks, we search for the learning behaviors of learners during the process of answering questions. By organically combining the perception ability of deep learning with the decision-making ability of reinforcement learning, the reinforcement learning recommendation method is integrated into the learning process of students to realize the improvement of students’ persistent and long-term knowledge level and learning ability. Finally, the effectiveness of the assisted teaching system in Civics learning is explored by choosing the dataset of exercises and learners’ response records from five university Civics courses as the data source.
This chapter introduces the overall design of the auxiliary teaching system for college students’ Civics courses in the context of deep learning, starting with the requirement analysis, which briefly describes the requirements of the auxiliary teaching system for Civics courses in colleges and universities. Second is the overall system architecture, discussing the overall design ideas of the system. Finally, it is the design of each module, briefly describing the new functions added in the auxiliary teaching system for college civic and political courses, and the design of the interaction between modules.
The auxiliary teaching system for the Civics course in colleges and universities to be designed in this paper needs to meet the following requirements:
1) The basic functions of the auxiliary teaching system, including user login, identity identification, course guide, exercises practice, and page design of the guide system. 2) It is time-consuming and laborious to manually enter students’ information and question results, so the OCR recognition module is needed to recognize the student handbook and students’ handwritten characters to import students’ information and question results. 3) In the design of Civics courses and exercises, students are provided with in-class exercises to consolidate their knowledge after learning the courses, as well as a general exercise mode for students to practice after class, and finally, a recommended exercise mode, where the system recommends topics to students according to their knowledge status, so that students can do the least amount of questions and get the greatest improvement in their abilities. 4) In the Civics Exercise Recommendation Module, firstly, we pre-train the reinforcement learning recommendation system offline, and then through the recommendation model interface, the system outputs the recommended topics for students according to their knowledge status. 5) The question types currently supported by this system are the Civics courses studied at the university level, including five courses: Overview of the Basic Principles of Marxism, Introduction to Mao Zedong Thought and Socialism with Chinese Characteristics, Essentials of Modern Chinese History, Foundation of Ideological and Moral Cultivation and Law, and Forms and Policies, and the main forms of the exercise questions are single-choice, multiple-choice, short-answer, and expository questions.
The architecture of the Civics Aided Teaching System designed in this paper is shown in Fig. 1, the teaching aid system includes three parts, the front-end, the database, and the exercise push part, and they are introduced one by one below.

Architecture of ideological and political assisted teaching system
This segment is broken up into two parts, the first being the current browser-based student login and exercise page in CAT-SOOP.The other part of this paper is a newly designed and implemented function, the computer visual recognition module, which is used to evaluate student evaluation materials and answer results offline.
This part is provided by CAT-SOOP, which is expanded in this paper to support the newly added features. Mainly, the user section has been enhanced to add support for student evaluations in the student handbook and enhancements to the student question history to support enhanced learning recommendation features.
This part is a newly added function in this paper, which functions to recommend suitable Civics exercises according to students’ learning status, learning ability and learning characteristics. In the implementation of specific algorithms, heuristic and enhancement learning-based recommendation algorithms are designed and implemented.
In this paper, based on CAT-SOOP, an open-source online auxiliary teaching system of College X, we modify the existing modules and add new modules to realize the functions of computer visual recognition of in-depth offline student evaluation materials and adaptive exercise recommendation. The module diagram of the Civics Aided Teaching System and their interaction is shown in Figure 2. The system consists of 8 modules. Among them, the newly added modules are OCR recognition module and recommendation module, modified course module, page module, exercise set module, scoring module, and user module, and the following mainly introduces the functions of the new modules, the way of realizing them, and the interrelationship between them.

Overall module design of ideological and political auxiliary teaching system
The OCR Recognition module is a new module added to enrich the information in the student variables with information from the student handbook and handwritten characters.A new catalog handbook has been added, and its specific functions are:
1) Pre-process the recorded images of student handbook and student handwritten characters by denoising, binarization, cutting, classification and other pre-processing processes to get the image style suitable for recognition. 2) The trained model is put on the server, and the pre-processed images are put into the corresponding API interface to recognize, and the server returns the results of the students’ evaluations and questions. 3) OCR recognition module is an API interface, users can input the whole picture and then get the result, which has better generality.
The OCR recognition module has the following relationships with other modules:
1) The recognition result of the student’s handbook will add the evaluation result given by the teacher to the student to the user variable, which makes the user’s parameter richer and helps us to analyze the next step. 2) The recognition result of the student’s handwritten characters will be fed into the recommendation model, which will be given to initialize the student’s knowledge state vector based on the student’s history of doing problems.
To provide a more efficient learning environment for students, the Civics Assisted Teaching System will recommend Civics exercises for students that are suitable for him according to his knowledge status, and we have added a new catalog, recom. Its specific functions are:
1) Recommending Civics exercises to students according to their knowledge status, which can dynamically monitor students’ Civics knowledge status, so that students can maximize the benefits of each topic they do, and reduce the time it takes for students to fully master the knowledge points. 2) The offline trained model is put into the Civics support system, which accepts students’ question data from the front-end page, updates students’ knowledge status, and finds the questions that can make students’ ability grow the fastest.
The relationship between the recommendation module and other modules is mainly to accept the data from the front-end page of the students’ questions, get the serial number of the students, as well as the time of the questions, the serial number of the questions, as well as the right and wrong of the questions. The recommendation algorithm based on reinforcement learning is utilized to recommend Civics topics to the students, and then the data is transmitted to the front-end page through the web service.
The construction of the network diagram of Civic and Political Studies is divided into two parts, one of which is the construction of the connecting edges of the topics under the same knowledge point, and the construction of the connecting edges under the knowledge points where there is a sequential relationship.
The network diagram of Civics exercises under the same knowledge point is shown in Figure 3, which can be represented as follows:
1) Under the same knowledge point 2) The predecessor and successor topics of a topic, and the weights of the edges between the topic and the predecessor and successor topics. For example, topic 3) To answer a subsequent question, you need to answer the previous question correctly, and to answer a subsequent question incorrectly, you need to answer the previous question. For example, if you answer Question

Network diagram of ideological and political exercises under the same knowledge point
A network diagram of the topics under the knowledge points where a sequential relationship exists is shown in Figure 4 and can represent the following:
1) Knowledge point 2) The predecessor and successor topics of a topic, and the weights of the edges between the topic and the predecessor and successor topics. For example, topic 3) To answer a subsequent question, you need to answer the previous question correctly, and to answer a subsequent question incorrectly, you need to answer the previous question. For example, if you answer Question

Exercise network under knowledge point with sequence relation
The cut set is a collection of branches in the connectivity graph. If these branches are removed, the connectivity graph will be separated into two parts that are not connected to each other. There is a difference between a point cut set and an edge cut set, which is used here. The historical behavior of a particular learner with respect to the solution of an exercise is reflected in Figures
Where,

Learner subgraph
The Q-Learning (Q-Learning) algorithm is an offline policy algorithm and a value-based reinforcement learning algorithm. Where Q is denoted as the
The input of the q network is made by the user of the length of h, which is a short-term browsing information group, which includes two points: the intellectual id and the thought of the political category in the sequence of the user’s browsing time. After the user’s short-term browsing information into the q network, the probability of each of the output data concentration is recommended in the next step of the second step, and the q value of the action is obtained by taking different actions in the current state.
For input data, the id and category of the thought of political knowledge are represented by the numerical index, and then, after the input network, the first of each number is embedded into the 1×2V vector with the Embeded ding layer. After embedding, the Concatanate layer connects the id vector and category vectors embedded in each group of thought government knowledge, each of which is 1×2V, and thus obtains the view information of H×2V using the dimension of each thought policy. Then, use the LSTM layer to handle these vectors. Finally, the results of the LSTM layer output are processed by the full connecting layer, and the probability of the recommended time is given by the softmax function.
The Value update formula for the Q-Learning algorithm is:
where
The elements in the reinforcement learning model are categorized into intelligences, states, actions, rewards, etc., respectively. In this paper, the recommendation scenario for Civics and Politics questions is based on reinforcement learning, so it is necessary to define the relevant reinforcement learning elements. In the next section, the elements of reinforcement learning are defined for the recommendation scenario in the personalized Civics and Politics exercise recommendation module. The flow of the Civics and Politics exercise recommendation model based on reinforcement learning is shown in Figure 6. Specifically, each element of reinforcement learning will be defined separately.

Flowchart of a recommendation model based on reinforcement learning
In this paper, a reinforcement learning-based model for recommending Civics and Politics exercises is constructed to act as an intelligent body in a reinforcement learning recommendation scenario. The role of the intelligent body is to generate a new recommendation set as the user continues to explore the environment while learning with the Civics Aided Teaching System, taking actions to obtain rewards and updating the learner’s knowledge state.
The state of an intelligent body in the context of a recommendation scenario is represented as the discretized processed user knowledge state. The DKT deep knowledge tracking framework is utilized to train the student knowledge model based on the student’s history of doing problems, and the student knowledge state
The action of an intelligent body in the context of a recommendation scenario is represented as a set of to-be-recommended exercises
The ultimate goal of the recommender intelligence is to take the optimal action so that it reaches a given state, similarly, the ultimate goal of the recommendation function is to recommend the most suitable exercise resources for the user to learn.
As the feedback given by the environment to the intelligent body for the current action, the reward for the intelligent body’s exploration in the recommended scenario environment is expressed as the integrated probability of answering the to-be-recommended exercise set in the current state of the user’s knowledge level. That is, the comprehensive mastery level of the current knowledge, denoted as {
In this paper, we consider the user’s knowledge state level when defining rewards, and abstract the reward expression as the degree of comprehensive mastery of the current learner-recommended Civics exercise set. The reward function is represented as:
Where
Meanwhile, the cumulative reward of the intelligences exploration is expressed as the sum of the discounted rewards as:
where
When the user is learning, learning the current knowledge point, browsing and clicking on the exercise recommendation module, this environment is modeled as an exploratory operating environment for the recommending intelligent body, and each time the intelligent body generates a recommended exercise set, i.e., a question. The structure of the overall reinforcement learning-based thought exercise recommendation model is schematically shown in Fig. 7, in which the recommending intelligent body makes strategy decisions based on the learner’s knowledge state and recent answer behavior, and the current action and observed sequence can be observed as:

Structure of reinforcement learning recommendation model
Where
The environment state, i.e., the intelligent body state, is denoted as the current knowledge state of the current learner, denoted as
where
In this study, five university Civics courses, including Overview of the Basic Principles of Marxism, Introduction to Mao Zedong Thought and Socialism with Chinese Characteristics, Essentials of Modern Chinese History, Foundation of Ideological and Moral Cultivation and Law, and Forms and Policies, were selected as the data sources for the exercises and the data sets of the learners’ answer records. Since the original learners’ answer record data obtained were somewhat different from the experimental data needed for this study, the original answer record data needed to be preprocessed before the experiment.
The two features in this study have different value ranges for the exercise difficulty value and the learner proficiency level value, so the two feature data need to be normalized first, and their value ranges are both mapped in the range of [0,1].
The Civics answer records of some learners were selected as the item information data, and the estimated learner competency level values were obtained after finishing the calculation and were counted into the learner competency level distribution as shown in Figure 8. The horizontal coordinate in the figure is the division interval of the learner competence level parameter

Distribution of learners’ ability level
The purpose of the intelligent body in reinforcement learning is to find the optimal path to obtain the maximum reward through learning, the reward is the feedback of the environment on the selection and execution of the action during the learning process, the execution of a poor action to obtain a small reward or even punishment, then the probability of the next time that the action is selected to be executed will become smaller. If an optimal action receives the maximum reward, the probability that the next optimal action will be selected for execution will increase.A reasonable reward strategy can improve the efficiency of model training, so it is very important to design an instant reward strategy accordingly.
The purpose of this Civics course assisted teaching system is to recommend to the learners, can improve the five university Civics course exercises, to the learner’s ability level and the difficulty of the topic to build the learner’s optimal recommendation model as shown in Figure 9. The upper and lower lines in the figure represent the upper and lower boundaries of the "Civics and Political Science Exercise Recommendation", and the upper boundary is the "confusion area" of the learners, which indicates that the learners’ ability level is insufficient, but the difficulty of the exercise questions is too high, resulting in the incorrect rate of answering the questions, and it is easy to be puzzled by the learning content. This area indicates that the learner’s ability level is insufficient, but the difficulty of the practice questions is too high, resulting in a low rate of correct answers and easy confusion about the content. Below the lower boundary is the "slack zone", which indicates that the learner’s ability level is high, but the difficulty of the practice questions is too low, resulting in too high a correct answer rate, and the learner is easily bored with the study. The area between the two boundaries is the "optimal recommendation zone". For example, when the learner’s ability level is 0.9, the optimal recommended difficulty interval for Civics exercises is [0.5,1.0]. Therefore, the optimal path for the reinforcement learning objective is to find two boundaries, and the corresponding difficulty of the topics between the boundaries is recommended to the learners with the corresponding ability level, which can complete the Civics and Politics exercise recommendation task.

Learner optimal recommendation model
As can be seen from the network diagram of Civics exercises constructed in the previous section, the knowledge point relevance is the weight of the paths in the directed relationship graph, and the direction of the paths is the knowledge points that are easier to learn. Learning directedness aims to calculate the degree of association between knowledge points, and use the forward and backward relationship to generate recommended actions, and the directed generation of learning paths can significantly improve the efficiency of Civics learning and enhance the overall Civics learning effect. Its effectiveness in the recommendation process can be verified by analyzing the distribution of relevance of each knowledge point in the original dataset and the distribution of relevance of knowledge points in the recommendation process.
A comparison of the distribution of the relevance of knowledge points in the Civics course is shown in Figure 10, which shows the distribution of the relevance of knowledge points in the source data (Original) and the recommended sequence of Civics exercises based on reinforcement learning (Q-Learning). The first is the static matrix in the domain knowledge relationship graph, while the second refers to the correlation between neighboring knowledge points in all recommended sequences during the recommendation process.The horizontal coordinate is determined by the correlation degree, while the vertical coordinate is determined by the ratio. In the source data, the correlation is most distributed around 0.00 and 0.033, with proportions of 0.382 and 0.132, respectively.During the recommendation process, Q-Learning is most distributed at 0.033 and 0.035, with proportions of 0.186 and 0.238, and the two distributions diverge the most at 0.00 and 0.035, which is generated by the algorithm’s optimal decision-making.

Comparison of distribution of knowledge point correlation degree
The larger the correlation degree, the higher the reward value, and the more the Civics exercise recommendation tends to the learning goal. So the correlation degree of 0.035 is less than 0.03 in the source data, but the proportion in the Q-Learning recommendation sequence reaches 0.238.Therefore, it can be shown that the Civics course assisted teaching system based on reinforcement learning is effective for the directedness of Civics learning in colleges and universities.
In order to verify the effect of Civics learning progressivity, the trained model is tested with 100 consecutive exercise recommendations, and the changes in the comprehensive difficulty of all recommended Civics knowledge points are recorded.The changes in the comprehensive difficulty in the recommended sequence of Q-Learning are shown in Fig. 11, where the horizontal coordinate is the recommended step, and the vertical coordinate is the comprehensive difficulty of the Civics knowledge points recommended in each step. As can be seen from the figure, the overall difficulty of the recommendation process has increased, but the change is small in the local area. For example, during STEP 0~40, the comprehensive difficulty stays between 0.683-0.805, and during STEP 40~60, the learning progressivity is effective for the overall learning effect. The comprehensive difficulty remained around 0.65, and there were successive recommendations for exercises on the same knowledge points. In summary, the analysis shows that the Civics Course Assisted Teaching System based on reinforcement learning is effective for students’ progress in learning Civics knowledge points.

Comprehensive difficulty change in recommendation sequence
If we want to verify the effectiveness of the auxiliary teaching system of the Civics course in colleges and universities for learning engagement, we can compare the change of the average reward value when the learning factor is different values under the same learning state. Comparison of learning factor

Changes of the average reward value of different learning factors
From the analysis, it can be seen that
Aiming at the current problem that the learning effect of Civics and Politics in colleges and universities is not obvious, this paper is based on the reinforcement learning algorithm, improves the wise-adaptive learning model, and designs and constructs a highly efficient and stable auxiliary teaching system for Civics and Politics courses for college students. The main research results are as follows:
1) The detailed design, construction and testing of the auxiliary teaching system for college students’ Civics and Politics courses are completed, and the system realizes the functions of knowledge state assessment and personalized exercise recommendation. When the learning factor is 0.5, the average reward value is between 0.723 and 1.262, and the coverage rate of the recommended exercises is more than 0.65, and the mastery rate is more than 0.88, which is practical to a certain extent. 2) The auxiliary teaching system for college students’ Civics courses designed in this paper has a certain degree of exploration and innovativeness, which provides a useful reference for the progress of the whole Civics education, and has an important practical significance for saving students’ learning time, enhancing students’ learning ability and improving the teaching quality of Civics education. 3) The limitation of this article is that the quality of the data selection is determined by the difficulty of the user’s cooperation and the quality of the data selection, and it may be necessary to make a certain adjustment in other subjects. In the future, the work can be expanded in many ways, and the honest inspection side can be extended to other disciplines and grades, and further verify the applicability of the algorithm. In addition, this paper is based on the classic DQN algorithm, and in fact, the higher effect of the learning algorithm in the research process is typical of a3c, ddpg, ape-x, rainbow, and the future of the algorithm.
